aboutsummaryrefslogtreecommitdiff
path: root/vendor/k8s.io
diff options
context:
space:
mode:
authorbaude <bbaude@redhat.com>2018-03-23 09:00:42 -0500
committerAtomic Bot <atomic-devel@projectatomic.io>2018-03-27 21:55:33 +0000
commit9aba605ddecc84e070a55019bb34109c5d5fd9b6 (patch)
tree5e7f39a3c293602e13871ce5ebfd4e2c4f146ca1 /vendor/k8s.io
parentaf64e10400f8533a0c48ecdf5ab9b7fbf329e14e (diff)
downloadpodman-9aba605ddecc84e070a55019bb34109c5d5fd9b6.tar.gz
podman-9aba605ddecc84e070a55019bb34109c5d5fd9b6.tar.bz2
podman-9aba605ddecc84e070a55019bb34109c5d5fd9b6.zip
Remove dependency on kubernetes
podman parse and attach were using a very small portion of the kubernetes code but using it caused a signficant increase in binary size. Signed-off-by: baude <bbaude@redhat.com> Closes: #559 Approved by: rhatdan
Diffstat (limited to 'vendor/k8s.io')
-rw-r--r--vendor/k8s.io/kubernetes/LICENSE202
-rw-r--r--vendor/k8s.io/kubernetes/README.md86
-rw-r--r--vendor/k8s.io/kubernetes/pkg/api/legacyscheme/scheme.go46
-rw-r--r--vendor/k8s.io/kubernetes/pkg/api/service/util.go85
-rw-r--r--vendor/k8s.io/kubernetes/pkg/api/v1/pod/util.go296
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/autoscaling/annotations.go34
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/autoscaling/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/autoscaling/register.go53
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/autoscaling/types.go422
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/autoscaling/zz_generated.deepcopy.go572
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/annotation_key_constants.go92
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/doc.go24
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/field_constants.go38
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/helper/helpers.go597
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/install/install.go67
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/json.go28
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/objectreference.go34
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/pods/helpers.go63
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/register.go99
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/resource.go62
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/taint.go36
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/toleration.go30
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/types.go4656
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/conversion.go619
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/defaults.go413
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/doc.go23
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/helper/helpers.go450
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/register.go46
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.conversion.go5686
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.defaults.go638
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/validation/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/validation/events.go129
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/validation/validation.go5066
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/core/zz_generated.deepcopy.go5952
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/extensions/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/extensions/helpers.go37
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/extensions/register.go68
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/extensions/types.go1049
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/extensions/zz_generated.deepcopy.go1231
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/networking/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/networking/register.go51
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/networking/types.go191
-rw-r--r--vendor/k8s.io/kubernetes/pkg/apis/networking/zz_generated.deepcopy.go282
-rw-r--r--vendor/k8s.io/kubernetes/pkg/capabilities/capabilities.go95
-rw-r--r--vendor/k8s.io/kubernetes/pkg/capabilities/doc.go18
-rw-r--r--vendor/k8s.io/kubernetes/pkg/cloudprovider/README.md16
-rw-r--r--vendor/k8s.io/kubernetes/pkg/cloudprovider/cloud.go215
-rw-r--r--vendor/k8s.io/kubernetes/pkg/cloudprovider/doc.go18
-rw-r--r--vendor/k8s.io/kubernetes/pkg/cloudprovider/plugins.go122
-rw-r--r--vendor/k8s.io/kubernetes/pkg/controller/client_builder.go292
-rw-r--r--vendor/k8s.io/kubernetes/pkg/controller/controller_ref_manager.go502
-rw-r--r--vendor/k8s.io/kubernetes/pkg/controller/controller_utils.go1043
-rw-r--r--vendor/k8s.io/kubernetes/pkg/controller/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/controller/lookup_cache.go92
-rw-r--r--vendor/k8s.io/kubernetes/pkg/features/kube_features.go326
-rw-r--r--vendor/k8s.io/kubernetes/pkg/fieldpath/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/fieldpath/fieldpath.go103
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.pb.go26627
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.proto1217
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/constants.go55
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/cache.go199
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/container_gc.go87
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/container_reference_manager.go60
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/helpers.go359
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/os.go107
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_linux.go30
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_unsupported.go28
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/ref.go73
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/resize.go46
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime.go664
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache.go96
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache_fake.go45
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/container/sync_result.go128
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/types/constants.go33
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/types/doc.go18
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/types/labels.go41
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/types/pod_update.go175
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/types/types.go100
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/util/format/pod.go72
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/util/format/resources.go36
-rw-r--r--vendor/k8s.io/kubernetes/pkg/kubelet/util/ioutils/ioutils.go37
-rw-r--r--vendor/k8s.io/kubernetes/pkg/master/ports/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/master/ports/ports.go44
-rw-r--r--vendor/k8s.io/kubernetes/pkg/scheduler/api/doc.go20
-rw-r--r--vendor/k8s.io/kubernetes/pkg/scheduler/api/register.go55
-rw-r--r--vendor/k8s.io/kubernetes/pkg/scheduler/api/types.go281
-rw-r--r--vendor/k8s.io/kubernetes/pkg/scheduler/api/zz_generated.deepcopy.go485
-rw-r--r--vendor/k8s.io/kubernetes/pkg/security/apparmor/helpers.go80
-rw-r--r--vendor/k8s.io/kubernetes/pkg/security/apparmor/validate.go229
-rw-r--r--vendor/k8s.io/kubernetes/pkg/security/apparmor/validate_disabled.go24
-rw-r--r--vendor/k8s.io/kubernetes/pkg/serviceaccount/claims.go186
-rw-r--r--vendor/k8s.io/kubernetes/pkg/serviceaccount/jwt.go212
-rw-r--r--vendor/k8s.io/kubernetes/pkg/serviceaccount/legacy.go135
-rw-r--r--vendor/k8s.io/kubernetes/pkg/serviceaccount/util.go74
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/file/file.go57
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/hash/hash.go37
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/io/consistentread.go45
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/io/writer.go78
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/doc.go18
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/exec.go50
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount.go140
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount_unsupported.go87
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/fake.go199
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/mount.go276
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/mount_linux.go700
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/mount_unsupported.go110
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/mount_windows.go346
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount.go272
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount_unsupported.go87
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/net/sets/doc.go28
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/net/sets/ipnet.go121
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter.go124
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter_unsupported.go50
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/parsers/parsers.go58
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/pointer/pointer.go68
-rw-r--r--vendor/k8s.io/kubernetes/pkg/util/taints/taints.go342
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/doc.go19
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/metrics_cached.go74
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/metrics_du.go99
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/metrics_errors.go77
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/metrics_nil.go30
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/metrics_statfs.go69
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/plugins.go807
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs.go97
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs_unsupported.go38
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/util/recyclerclient/recycler_client.go252
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/volume.go273
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/volume_linux.go105
-rw-r--r--vendor/k8s.io/kubernetes/pkg/volume/volume_unsupported.go27
-rw-r--r--vendor/k8s.io/kubernetes/third_party/forked/golang/LICENSE27
-rw-r--r--vendor/k8s.io/kubernetes/third_party/forked/golang/PATENTS22
-rw-r--r--vendor/k8s.io/kubernetes/third_party/forked/golang/expansion/expand.go102
-rw-r--r--vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/compiler/plugin.proto150
-rw-r--r--vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/descriptor.proto779
134 files changed, 0 insertions, 71026 deletions
diff --git a/vendor/k8s.io/kubernetes/LICENSE b/vendor/k8s.io/kubernetes/LICENSE
deleted file mode 100644
index d64569567..000000000
--- a/vendor/k8s.io/kubernetes/LICENSE
+++ /dev/null
@@ -1,202 +0,0 @@
-
- Apache License
- Version 2.0, January 2004
- http://www.apache.org/licenses/
-
- TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
-
- 1. Definitions.
-
- "License" shall mean the terms and conditions for use, reproduction,
- and distribution as defined by Sections 1 through 9 of this document.
-
- "Licensor" shall mean the copyright owner or entity authorized by
- the copyright owner that is granting the License.
-
- "Legal Entity" shall mean the union of the acting entity and all
- other entities that control, are controlled by, or are under common
- control with that entity. For the purposes of this definition,
- "control" means (i) the power, direct or indirect, to cause the
- direction or management of such entity, whether by contract or
- otherwise, or (ii) ownership of fifty percent (50%) or more of the
- outstanding shares, or (iii) beneficial ownership of such entity.
-
- "You" (or "Your") shall mean an individual or Legal Entity
- exercising permissions granted by this License.
-
- "Source" form shall mean the preferred form for making modifications,
- including but not limited to software source code, documentation
- source, and configuration files.
-
- "Object" form shall mean any form resulting from mechanical
- transformation or translation of a Source form, including but
- not limited to compiled object code, generated documentation,
- and conversions to other media types.
-
- "Work" shall mean the work of authorship, whether in Source or
- Object form, made available under the License, as indicated by a
- copyright notice that is included in or attached to the work
- (an example is provided in the Appendix below).
-
- "Derivative Works" shall mean any work, whether in Source or Object
- form, that is based on (or derived from) the Work and for which the
- editorial revisions, annotations, elaborations, or other modifications
- represent, as a whole, an original work of authorship. For the purposes
- of this License, Derivative Works shall not include works that remain
- separable from, or merely link (or bind by name) to the interfaces of,
- the Work and Derivative Works thereof.
-
- "Contribution" shall mean any work of authorship, including
- the original version of the Work and any modifications or additions
- to that Work or Derivative Works thereof, that is intentionally
- submitted to Licensor for inclusion in the Work by the copyright owner
- or by an individual or Legal Entity authorized to submit on behalf of
- the copyright owner. For the purposes of this definition, "submitted"
- means any form of electronic, verbal, or written communication sent
- to the Licensor or its representatives, including but not limited to
- communication on electronic mailing lists, source code control systems,
- and issue tracking systems that are managed by, or on behalf of, the
- Licensor for the purpose of discussing and improving the Work, but
- excluding communication that is conspicuously marked or otherwise
- designated in writing by the copyright owner as "Not a Contribution."
-
- "Contributor" shall mean Licensor and any individual or Legal Entity
- on behalf of whom a Contribution has been received by Licensor and
- subsequently incorporated within the Work.
-
- 2. Grant of Copyright License. Subject to the terms and conditions of
- this License, each Contributor hereby grants to You a perpetual,
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
- copyright license to reproduce, prepare Derivative Works of,
- publicly display, publicly perform, sublicense, and distribute the
- Work and such Derivative Works in Source or Object form.
-
- 3. Grant of Patent License. Subject to the terms and conditions of
- this License, each Contributor hereby grants to You a perpetual,
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
- (except as stated in this section) patent license to make, have made,
- use, offer to sell, sell, import, and otherwise transfer the Work,
- where such license applies only to those patent claims licensable
- by such Contributor that are necessarily infringed by their
- Contribution(s) alone or by combination of their Contribution(s)
- with the Work to which such Contribution(s) was submitted. If You
- institute patent litigation against any entity (including a
- cross-claim or counterclaim in a lawsuit) alleging that the Work
- or a Contribution incorporated within the Work constitutes direct
- or contributory patent infringement, then any patent licenses
- granted to You under this License for that Work shall terminate
- as of the date such litigation is filed.
-
- 4. Redistribution. You may reproduce and distribute copies of the
- Work or Derivative Works thereof in any medium, with or without
- modifications, and in Source or Object form, provided that You
- meet the following conditions:
-
- (a) You must give any other recipients of the Work or
- Derivative Works a copy of this License; and
-
- (b) You must cause any modified files to carry prominent notices
- stating that You changed the files; and
-
- (c) You must retain, in the Source form of any Derivative Works
- that You distribute, all copyright, patent, trademark, and
- attribution notices from the Source form of the Work,
- excluding those notices that do not pertain to any part of
- the Derivative Works; and
-
- (d) If the Work includes a "NOTICE" text file as part of its
- distribution, then any Derivative Works that You distribute must
- include a readable copy of the attribution notices contained
- within such NOTICE file, excluding those notices that do not
- pertain to any part of the Derivative Works, in at least one
- of the following places: within a NOTICE text file distributed
- as part of the Derivative Works; within the Source form or
- documentation, if provided along with the Derivative Works; or,
- within a display generated by the Derivative Works, if and
- wherever such third-party notices normally appear. The contents
- of the NOTICE file are for informational purposes only and
- do not modify the License. You may add Your own attribution
- notices within Derivative Works that You distribute, alongside
- or as an addendum to the NOTICE text from the Work, provided
- that such additional attribution notices cannot be construed
- as modifying the License.
-
- You may add Your own copyright statement to Your modifications and
- may provide additional or different license terms and conditions
- for use, reproduction, or distribution of Your modifications, or
- for any such Derivative Works as a whole, provided Your use,
- reproduction, and distribution of the Work otherwise complies with
- the conditions stated in this License.
-
- 5. Submission of Contributions. Unless You explicitly state otherwise,
- any Contribution intentionally submitted for inclusion in the Work
- by You to the Licensor shall be under the terms and conditions of
- this License, without any additional terms or conditions.
- Notwithstanding the above, nothing herein shall supersede or modify
- the terms of any separate license agreement you may have executed
- with Licensor regarding such Contributions.
-
- 6. Trademarks. This License does not grant permission to use the trade
- names, trademarks, service marks, or product names of the Licensor,
- except as required for reasonable and customary use in describing the
- origin of the Work and reproducing the content of the NOTICE file.
-
- 7. Disclaimer of Warranty. Unless required by applicable law or
- agreed to in writing, Licensor provides the Work (and each
- Contributor provides its Contributions) on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
- implied, including, without limitation, any warranties or conditions
- of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
- PARTICULAR PURPOSE. You are solely responsible for determining the
- appropriateness of using or redistributing the Work and assume any
- risks associated with Your exercise of permissions under this License.
-
- 8. Limitation of Liability. In no event and under no legal theory,
- whether in tort (including negligence), contract, or otherwise,
- unless required by applicable law (such as deliberate and grossly
- negligent acts) or agreed to in writing, shall any Contributor be
- liable to You for damages, including any direct, indirect, special,
- incidental, or consequential damages of any character arising as a
- result of this License or out of the use or inability to use the
- Work (including but not limited to damages for loss of goodwill,
- work stoppage, computer failure or malfunction, or any and all
- other commercial damages or losses), even if such Contributor
- has been advised of the possibility of such damages.
-
- 9. Accepting Warranty or Additional Liability. While redistributing
- the Work or Derivative Works thereof, You may choose to offer,
- and charge a fee for, acceptance of support, warranty, indemnity,
- or other liability obligations and/or rights consistent with this
- License. However, in accepting such obligations, You may act only
- on Your own behalf and on Your sole responsibility, not on behalf
- of any other Contributor, and only if You agree to indemnify,
- defend, and hold each Contributor harmless for any liability
- incurred by, or claims asserted against, such Contributor by reason
- of your accepting any such warranty or additional liability.
-
- END OF TERMS AND CONDITIONS
-
- APPENDIX: How to apply the Apache License to your work.
-
- To apply the Apache License to your work, attach the following
- boilerplate notice, with the fields enclosed by brackets "[]"
- replaced with your own identifying information. (Don't include
- the brackets!) The text should be enclosed in the appropriate
- comment syntax for the file format. We also recommend that a
- file or class name and description of purpose be included on the
- same "printed page" as the copyright notice for easier
- identification within third-party archives.
-
- Copyright [yyyy] [name of copyright owner]
-
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
diff --git a/vendor/k8s.io/kubernetes/README.md b/vendor/k8s.io/kubernetes/README.md
deleted file mode 100644
index 83e5deda7..000000000
--- a/vendor/k8s.io/kubernetes/README.md
+++ /dev/null
@@ -1,86 +0,0 @@
-# Kubernetes
-
-[![Submit Queue Widget]][Submit Queue] [![GoDoc Widget]][GoDoc] [![CII Best Practices](https://bestpractices.coreinfrastructure.org/projects/569/badge)](https://bestpractices.coreinfrastructure.org/projects/569)
-
-<img src="https://github.com/kubernetes/kubernetes/raw/master/logo/logo.png" width="100">
-
-----
-
-Kubernetes is an open source system for managing [containerized applications]
-across multiple hosts; providing basic mechanisms for deployment, maintenance,
-and scaling of applications.
-
-Kubernetes builds upon a decade and a half of experience at Google running
-production workloads at scale using a system called [Borg],
-combined with best-of-breed ideas and practices from the community.
-
-Kubernetes is hosted by the Cloud Native Computing Foundation ([CNCF]).
-If you are a company that wants to help shape the evolution of
-technologies that are container-packaged, dynamically-scheduled
-and microservices-oriented, consider joining the CNCF.
-For details about who's involved and how Kubernetes plays a role,
-read the CNCF [announcement].
-
-----
-
-## To start using Kubernetes
-
-See our documentation on [kubernetes.io].
-
-Try our [interactive tutorial].
-
-Take a free course on [Scalable Microservices with Kubernetes].
-
-## To start developing Kubernetes
-
-The [community repository] hosts all information about
-building Kubernetes from source, how to contribute code
-and documentation, who to contact about what, etc.
-
-If you want to build Kubernetes right away there are two options:
-
-##### You have a working [Go environment].
-
-```
-$ go get -d k8s.io/kubernetes
-$ cd $GOPATH/src/k8s.io/kubernetes
-$ make
-```
-
-##### You have a working [Docker environment].
-
-```
-$ git clone https://github.com/kubernetes/kubernetes
-$ cd kubernetes
-$ make quick-release
-```
-
-For the full story, head over to the [developer's documentation].
-
-## Support
-
-If you need support, start with the [troubleshooting guide],
-and work your way through the process that we've outlined.
-
-That said, if you have questions, reach out to us
-[one way or another][communication].
-
-[announcement]: https://cncf.io/news/announcement/2015/07/new-cloud-native-computing-foundation-drive-alignment-among-container
-[Borg]: https://research.google.com/pubs/pub43438.html
-[CNCF]: https://www.cncf.io/about
-[communication]: https://git.k8s.io/community/communication
-[community repository]: https://git.k8s.io/community
-[containerized applications]: https://kubernetes.io/docs/concepts/overview/what-is-kubernetes/
-[developer's documentation]: https://git.k8s.io/community/contributors/devel#readme
-[Docker environment]: https://docs.docker.com/engine
-[Go environment]: https://golang.org/doc/install
-[GoDoc]: https://godoc.org/k8s.io/kubernetes
-[GoDoc Widget]: https://godoc.org/k8s.io/kubernetes?status.svg
-[interactive tutorial]: http://kubernetes.io/docs/tutorials/kubernetes-basics
-[kubernetes.io]: http://kubernetes.io
-[Scalable Microservices with Kubernetes]: https://www.udacity.com/course/scalable-microservices-with-kubernetes--ud615
-[Submit Queue]: http://submit-queue.k8s.io/#/ci
-[Submit Queue Widget]: http://submit-queue.k8s.io/health.svg?v=1
-[troubleshooting guide]: https://kubernetes.io/docs/tasks/debug-application-cluster/troubleshooting/
-
-[![Analytics](https://kubernetes-site.appspot.com/UA-36037335-10/GitHub/README.md?pixel)]()
diff --git a/vendor/k8s.io/kubernetes/pkg/api/legacyscheme/scheme.go b/vendor/k8s.io/kubernetes/pkg/api/legacyscheme/scheme.go
deleted file mode 100644
index 64c63a360..000000000
--- a/vendor/k8s.io/kubernetes/pkg/api/legacyscheme/scheme.go
+++ /dev/null
@@ -1,46 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package legacyscheme
-
-import (
- "os"
-
- "k8s.io/apimachinery/pkg/apimachinery/announced"
- "k8s.io/apimachinery/pkg/apimachinery/registered"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/serializer"
-)
-
-// GroupFactoryRegistry is the APIGroupFactoryRegistry (overlaps a bit with Registry, see comments in package for details)
-var GroupFactoryRegistry = make(announced.APIGroupFactoryRegistry)
-
-// Registry is an instance of an API registry. This is an interim step to start removing the idea of a global
-// API registry.
-var Registry = registered.NewOrDie(os.Getenv("KUBE_API_VERSIONS"))
-
-// Scheme is the default instance of runtime.Scheme to which types in the Kubernetes API are already registered.
-// NOTE: If you are copying this file to start a new api group, STOP! Copy the
-// extensions group instead. This Scheme is special and should appear ONLY in
-// the api group, unless you really know what you're doing.
-// TODO(lavalamp): make the above error impossible.
-var Scheme = runtime.NewScheme()
-
-// Codecs provides access to encoding and decoding for the scheme
-var Codecs = serializer.NewCodecFactory(Scheme)
-
-// ParameterCodec handles versioning of objects that are converted to query parameters.
-var ParameterCodec = runtime.NewParameterCodec(Scheme)
diff --git a/vendor/k8s.io/kubernetes/pkg/api/service/util.go b/vendor/k8s.io/kubernetes/pkg/api/service/util.go
deleted file mode 100644
index 5de5f2765..000000000
--- a/vendor/k8s.io/kubernetes/pkg/api/service/util.go
+++ /dev/null
@@ -1,85 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package service
-
-import (
- "fmt"
- api "k8s.io/kubernetes/pkg/apis/core"
- netsets "k8s.io/kubernetes/pkg/util/net/sets"
- "strings"
-)
-
-const (
- defaultLoadBalancerSourceRanges = "0.0.0.0/0"
-)
-
-// IsAllowAll checks whether the netsets.IPNet allows traffic from 0.0.0.0/0
-func IsAllowAll(ipnets netsets.IPNet) bool {
- for _, s := range ipnets.StringSlice() {
- if s == "0.0.0.0/0" {
- return true
- }
- }
- return false
-}
-
-// GetLoadBalancerSourceRanges first try to parse and verify LoadBalancerSourceRanges field from a service.
-// If the field is not specified, turn to parse and verify the AnnotationLoadBalancerSourceRangesKey annotation from a service,
-// extracting the source ranges to allow, and if not present returns a default (allow-all) value.
-func GetLoadBalancerSourceRanges(service *api.Service) (netsets.IPNet, error) {
- var ipnets netsets.IPNet
- var err error
- // if SourceRange field is specified, ignore sourceRange annotation
- if len(service.Spec.LoadBalancerSourceRanges) > 0 {
- specs := service.Spec.LoadBalancerSourceRanges
- ipnets, err = netsets.ParseIPNets(specs...)
-
- if err != nil {
- return nil, fmt.Errorf("service.Spec.LoadBalancerSourceRanges: %v is not valid. Expecting a list of IP ranges. For example, 10.0.0.0/24. Error msg: %v", specs, err)
- }
- } else {
- val := service.Annotations[api.AnnotationLoadBalancerSourceRangesKey]
- val = strings.TrimSpace(val)
- if val == "" {
- val = defaultLoadBalancerSourceRanges
- }
- specs := strings.Split(val, ",")
- ipnets, err = netsets.ParseIPNets(specs...)
- if err != nil {
- return nil, fmt.Errorf("%s: %s is not valid. Expecting a comma-separated list of source IP ranges. For example, 10.0.0.0/24,192.168.2.0/24", api.AnnotationLoadBalancerSourceRangesKey, val)
- }
- }
- return ipnets, nil
-}
-
-// RequestsOnlyLocalTraffic checks if service requests OnlyLocal traffic.
-func RequestsOnlyLocalTraffic(service *api.Service) bool {
- if service.Spec.Type != api.ServiceTypeLoadBalancer &&
- service.Spec.Type != api.ServiceTypeNodePort {
- return false
- }
-
- return service.Spec.ExternalTrafficPolicy == api.ServiceExternalTrafficPolicyTypeLocal
-}
-
-// NeedsHealthCheck checks if service needs health check.
-func NeedsHealthCheck(service *api.Service) bool {
- if service.Spec.Type != api.ServiceTypeLoadBalancer {
- return false
- }
- return RequestsOnlyLocalTraffic(service)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/api/v1/pod/util.go b/vendor/k8s.io/kubernetes/pkg/api/v1/pod/util.go
deleted file mode 100644
index aa79c7250..000000000
--- a/vendor/k8s.io/kubernetes/pkg/api/v1/pod/util.go
+++ /dev/null
@@ -1,296 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package pod
-
-import (
- "fmt"
- "time"
-
- "k8s.io/api/core/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/util/intstr"
-)
-
-// FindPort locates the container port for the given pod and portName. If the
-// targetPort is a number, use that. If the targetPort is a string, look that
-// string up in all named ports in all containers in the target pod. If no
-// match is found, fail.
-func FindPort(pod *v1.Pod, svcPort *v1.ServicePort) (int, error) {
- portName := svcPort.TargetPort
- switch portName.Type {
- case intstr.String:
- name := portName.StrVal
- for _, container := range pod.Spec.Containers {
- for _, port := range container.Ports {
- if port.Name == name && port.Protocol == svcPort.Protocol {
- return int(port.ContainerPort), nil
- }
- }
- }
- case intstr.Int:
- return portName.IntValue(), nil
- }
-
- return 0, fmt.Errorf("no suitable port for manifest: %s", pod.UID)
-}
-
-// Visitor is called with each object name, and returns true if visiting should continue
-type Visitor func(name string) (shouldContinue bool)
-
-// VisitPodSecretNames invokes the visitor function with the name of every secret
-// referenced by the pod spec. If visitor returns false, visiting is short-circuited.
-// Transitive references (e.g. pod -> pvc -> pv -> secret) are not visited.
-// Returns true if visiting completed, false if visiting was short-circuited.
-func VisitPodSecretNames(pod *v1.Pod, visitor Visitor) bool {
- for _, reference := range pod.Spec.ImagePullSecrets {
- if !visitor(reference.Name) {
- return false
- }
- }
- for i := range pod.Spec.InitContainers {
- if !visitContainerSecretNames(&pod.Spec.InitContainers[i], visitor) {
- return false
- }
- }
- for i := range pod.Spec.Containers {
- if !visitContainerSecretNames(&pod.Spec.Containers[i], visitor) {
- return false
- }
- }
- var source *v1.VolumeSource
-
- for i := range pod.Spec.Volumes {
- source = &pod.Spec.Volumes[i].VolumeSource
- switch {
- case source.AzureFile != nil:
- if len(source.AzureFile.SecretName) > 0 && !visitor(source.AzureFile.SecretName) {
- return false
- }
- case source.CephFS != nil:
- if source.CephFS.SecretRef != nil && !visitor(source.CephFS.SecretRef.Name) {
- return false
- }
- case source.FlexVolume != nil:
- if source.FlexVolume.SecretRef != nil && !visitor(source.FlexVolume.SecretRef.Name) {
- return false
- }
- case source.Projected != nil:
- for j := range source.Projected.Sources {
- if source.Projected.Sources[j].Secret != nil {
- if !visitor(source.Projected.Sources[j].Secret.Name) {
- return false
- }
- }
- }
- case source.RBD != nil:
- if source.RBD.SecretRef != nil && !visitor(source.RBD.SecretRef.Name) {
- return false
- }
- case source.Secret != nil:
- if !visitor(source.Secret.SecretName) {
- return false
- }
- case source.ScaleIO != nil:
- if source.ScaleIO.SecretRef != nil && !visitor(source.ScaleIO.SecretRef.Name) {
- return false
- }
- case source.ISCSI != nil:
- if source.ISCSI.SecretRef != nil && !visitor(source.ISCSI.SecretRef.Name) {
- return false
- }
- case source.StorageOS != nil:
- if source.StorageOS.SecretRef != nil && !visitor(source.StorageOS.SecretRef.Name) {
- return false
- }
- }
- }
- return true
-}
-
-func visitContainerSecretNames(container *v1.Container, visitor Visitor) bool {
- for _, env := range container.EnvFrom {
- if env.SecretRef != nil {
- if !visitor(env.SecretRef.Name) {
- return false
- }
- }
- }
- for _, envVar := range container.Env {
- if envVar.ValueFrom != nil && envVar.ValueFrom.SecretKeyRef != nil {
- if !visitor(envVar.ValueFrom.SecretKeyRef.Name) {
- return false
- }
- }
- }
- return true
-}
-
-// VisitPodConfigmapNames invokes the visitor function with the name of every configmap
-// referenced by the pod spec. If visitor returns false, visiting is short-circuited.
-// Transitive references (e.g. pod -> pvc -> pv -> secret) are not visited.
-// Returns true if visiting completed, false if visiting was short-circuited.
-func VisitPodConfigmapNames(pod *v1.Pod, visitor Visitor) bool {
- for i := range pod.Spec.InitContainers {
- if !visitContainerConfigmapNames(&pod.Spec.InitContainers[i], visitor) {
- return false
- }
- }
- for i := range pod.Spec.Containers {
- if !visitContainerConfigmapNames(&pod.Spec.Containers[i], visitor) {
- return false
- }
- }
- var source *v1.VolumeSource
- for i := range pod.Spec.Volumes {
- source = &pod.Spec.Volumes[i].VolumeSource
- switch {
- case source.Projected != nil:
- for j := range source.Projected.Sources {
- if source.Projected.Sources[j].ConfigMap != nil {
- if !visitor(source.Projected.Sources[j].ConfigMap.Name) {
- return false
- }
- }
- }
- case source.ConfigMap != nil:
- if !visitor(source.ConfigMap.Name) {
- return false
- }
- }
- }
- return true
-}
-
-func visitContainerConfigmapNames(container *v1.Container, visitor Visitor) bool {
- for _, env := range container.EnvFrom {
- if env.ConfigMapRef != nil {
- if !visitor(env.ConfigMapRef.Name) {
- return false
- }
- }
- }
- for _, envVar := range container.Env {
- if envVar.ValueFrom != nil && envVar.ValueFrom.ConfigMapKeyRef != nil {
- if !visitor(envVar.ValueFrom.ConfigMapKeyRef.Name) {
- return false
- }
- }
- }
- return true
-}
-
-// GetContainerStatus extracts the status of container "name" from "statuses".
-// It also returns if "name" exists.
-func GetContainerStatus(statuses []v1.ContainerStatus, name string) (v1.ContainerStatus, bool) {
- for i := range statuses {
- if statuses[i].Name == name {
- return statuses[i], true
- }
- }
- return v1.ContainerStatus{}, false
-}
-
-// GetExistingContainerStatus extracts the status of container "name" from "statuses",
-// and returns empty status if "name" does not exist.
-func GetExistingContainerStatus(statuses []v1.ContainerStatus, name string) v1.ContainerStatus {
- for i := range statuses {
- if statuses[i].Name == name {
- return statuses[i]
- }
- }
- return v1.ContainerStatus{}
-}
-
-// IsPodAvailable returns true if a pod is available; false otherwise.
-// Precondition for an available pod is that it must be ready. On top
-// of that, there are two cases when a pod can be considered available:
-// 1. minReadySeconds == 0, or
-// 2. LastTransitionTime (is set) + minReadySeconds < current time
-func IsPodAvailable(pod *v1.Pod, minReadySeconds int32, now metav1.Time) bool {
- if !IsPodReady(pod) {
- return false
- }
-
- c := GetPodReadyCondition(pod.Status)
- minReadySecondsDuration := time.Duration(minReadySeconds) * time.Second
- if minReadySeconds == 0 || !c.LastTransitionTime.IsZero() && c.LastTransitionTime.Add(minReadySecondsDuration).Before(now.Time) {
- return true
- }
- return false
-}
-
-// IsPodReady returns true if a pod is ready; false otherwise.
-func IsPodReady(pod *v1.Pod) bool {
- return IsPodReadyConditionTrue(pod.Status)
-}
-
-// IsPodReady returns true if a pod is ready; false otherwise.
-func IsPodReadyConditionTrue(status v1.PodStatus) bool {
- condition := GetPodReadyCondition(status)
- return condition != nil && condition.Status == v1.ConditionTrue
-}
-
-// Extracts the pod ready condition from the given status and returns that.
-// Returns nil if the condition is not present.
-func GetPodReadyCondition(status v1.PodStatus) *v1.PodCondition {
- _, condition := GetPodCondition(&status, v1.PodReady)
- return condition
-}
-
-// GetPodCondition extracts the provided condition from the given status and returns that.
-// Returns nil and -1 if the condition is not present, and the index of the located condition.
-func GetPodCondition(status *v1.PodStatus, conditionType v1.PodConditionType) (int, *v1.PodCondition) {
- if status == nil {
- return -1, nil
- }
- for i := range status.Conditions {
- if status.Conditions[i].Type == conditionType {
- return i, &status.Conditions[i]
- }
- }
- return -1, nil
-}
-
-// Updates existing pod condition or creates a new one. Sets LastTransitionTime to now if the
-// status has changed.
-// Returns true if pod condition has changed or has been added.
-func UpdatePodCondition(status *v1.PodStatus, condition *v1.PodCondition) bool {
- condition.LastTransitionTime = metav1.Now()
- // Try to find this pod condition.
- conditionIndex, oldCondition := GetPodCondition(status, condition.Type)
-
- if oldCondition == nil {
- // We are adding new pod condition.
- status.Conditions = append(status.Conditions, *condition)
- return true
- } else {
- // We are updating an existing condition, so we need to check if it has changed.
- if condition.Status == oldCondition.Status {
- condition.LastTransitionTime = oldCondition.LastTransitionTime
- }
-
- isEqual := condition.Status == oldCondition.Status &&
- condition.Reason == oldCondition.Reason &&
- condition.Message == oldCondition.Message &&
- condition.LastProbeTime.Equal(&oldCondition.LastProbeTime) &&
- condition.LastTransitionTime.Equal(&oldCondition.LastTransitionTime)
-
- status.Conditions[conditionIndex] = *condition
- // Return true if one of the fields have changed.
- return !isEqual
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/annotations.go b/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/annotations.go
deleted file mode 100644
index ccf034535..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/annotations.go
+++ /dev/null
@@ -1,34 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package autoscaling
-
-// MetricSpecsAnnotation is the annotation which holds non-CPU-utilization HPA metric
-// specs when converting the `Metrics` field from autoscaling/v2beta1
-const MetricSpecsAnnotation = "autoscaling.alpha.kubernetes.io/metrics"
-
-// MetricStatusesAnnotation is the annotation which holds non-CPU-utilization HPA metric
-// statuses when converting the `CurrentMetrics` field from autoscaling/v2beta1
-const MetricStatusesAnnotation = "autoscaling.alpha.kubernetes.io/current-metrics"
-
-// HorizontalPodAutoscalerConditionsAnnotation is the annotation which holds the conditions
-// of an HPA when converting the `Conditions` field from autoscaling/v2beta1
-const HorizontalPodAutoscalerConditionsAnnotation = "autoscaling.alpha.kubernetes.io/conditions"
-
-// DefaultCPUUtilization is the default value for CPU utilization, provided no other
-// metrics are present. This is here because it's used by both the v2beta1 defaulting
-// logic, and the pseudo-defaulting done in v1 conversion.
-const DefaultCPUUtilization = 80
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/doc.go
deleted file mode 100644
index 7c91aac8b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:deepcopy-gen=package
-
-package autoscaling // import "k8s.io/kubernetes/pkg/apis/autoscaling"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/register.go b/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/register.go
deleted file mode 100644
index 6c321a3ab..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/register.go
+++ /dev/null
@@ -1,53 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package autoscaling
-
-import (
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-// GroupName is the group name use in this package
-const GroupName = "autoscaling"
-
-// SchemeGroupVersion is group version used to register these objects
-var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: runtime.APIVersionInternal}
-
-// Kind takes an unqualified kind and returns a Group qualified GroupKind
-func Kind(kind string) schema.GroupKind {
- return SchemeGroupVersion.WithKind(kind).GroupKind()
-}
-
-// Resource takes an unqualified resource and returns a Group qualified GroupResource
-func Resource(resource string) schema.GroupResource {
- return SchemeGroupVersion.WithResource(resource).GroupResource()
-}
-
-var (
- SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes)
- AddToScheme = SchemeBuilder.AddToScheme
-)
-
-// Adds the list of known types to the given scheme.
-func addKnownTypes(scheme *runtime.Scheme) error {
- scheme.AddKnownTypes(SchemeGroupVersion,
- &Scale{},
- &HorizontalPodAutoscaler{},
- &HorizontalPodAutoscalerList{},
- )
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/types.go b/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/types.go
deleted file mode 100644
index 5489742db..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/types.go
+++ /dev/null
@@ -1,422 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package autoscaling
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- api "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Scale represents a scaling request for a resource.
-type Scale struct {
- metav1.TypeMeta
- // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata.
- // +optional
- metav1.ObjectMeta
-
- // defines the behavior of the scale. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status.
- // +optional
- Spec ScaleSpec
-
- // current status of the scale. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. Read-only.
- // +optional
- Status ScaleStatus
-}
-
-// ScaleSpec describes the attributes of a scale subresource.
-type ScaleSpec struct {
- // desired number of instances for the scaled object.
- // +optional
- Replicas int32
-}
-
-// ScaleStatus represents the current status of a scale subresource.
-type ScaleStatus struct {
- // actual number of observed instances of the scaled object.
- Replicas int32
-
- // label query over pods that should match the replicas count. This is same
- // as the label selector but in the string format to avoid introspection
- // by clients. The string will be in the same format as the query-param syntax.
- // More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors
- // +optional
- Selector string
-}
-
-// CrossVersionObjectReference contains enough information to let you identify the referred resource.
-type CrossVersionObjectReference struct {
- // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds"
- Kind string
- // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names
- Name string
- // API version of the referent
- // +optional
- APIVersion string
-}
-
-// HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler.
-type HorizontalPodAutoscalerSpec struct {
- // ScaleTargetRef points to the target resource to scale, and is used to the pods for which metrics
- // should be collected, as well as to actually change the replica count.
- ScaleTargetRef CrossVersionObjectReference
- // MinReplicas is the lower limit for the number of replicas to which the autoscaler can scale down.
- // It defaults to 1 pod.
- // +optional
- MinReplicas *int32
- // MaxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up.
- // It cannot be less that minReplicas.
- MaxReplicas int32
- // Metrics contains the specifications for which to use to calculate the
- // desired replica count (the maximum replica count across all metrics will
- // be used). The desired replica count is calculated multiplying the
- // ratio between the target value and the current value by the current
- // number of pods. Ergo, metrics used must decrease as the pod count is
- // increased, and vice-versa. See the individual metric source types for
- // more information about how each type of metric must respond.
- // +optional
- Metrics []MetricSpec
-}
-
-// MetricSourceType indicates the type of metric.
-type MetricSourceType string
-
-var (
- // ObjectMetricSourceType is a metric describing a kubernetes object
- // (for example, hits-per-second on an Ingress object).
- ObjectMetricSourceType MetricSourceType = "Object"
- // PodsMetricSourceType is a metric describing each pod in the current scale
- // target (for example, transactions-processed-per-second). The values
- // will be averaged together before being compared to the target value.
- PodsMetricSourceType MetricSourceType = "Pods"
- // ResourceMetricSourceType is a resource metric known to Kubernetes, as
- // specified in requests and limits, describing each pod in the current
- // scale target (e.g. CPU or memory). Such metrics are built in to
- // Kubernetes, and have special scaling options on top of those available
- // to normal per-pod metrics (the "pods" source).
- ResourceMetricSourceType MetricSourceType = "Resource"
- // ExternalMetricSourceType is a global metric that is not associated
- // with any Kubernetes object. It allows autoscaling based on information
- // coming from components running outside of cluster
- // (for example length of queue in cloud messaging service, or
- // QPS from loadbalancer running outside of cluster).
- ExternalMetricSourceType MetricSourceType = "External"
-)
-
-// MetricSpec specifies how to scale based on a single metric
-// (only `type` and one other matching field should be set at once).
-type MetricSpec struct {
- // Type is the type of metric source. It should be one of "Object",
- // "Pods" or "Resource", each mapping to a matching field in the object.
- Type MetricSourceType
-
- // Object refers to a metric describing a single kubernetes object
- // (for example, hits-per-second on an Ingress object).
- // +optional
- Object *ObjectMetricSource
- // Pods refers to a metric describing each pod in the current scale target
- // (for example, transactions-processed-per-second). The values will be
- // averaged together before being compared to the target value.
- // +optional
- Pods *PodsMetricSource
- // Resource refers to a resource metric (such as those specified in
- // requests and limits) known to Kubernetes describing each pod in the
- // current scale target (e.g. CPU or memory). Such metrics are built in to
- // Kubernetes, and have special scaling options on top of those available
- // to normal per-pod metrics using the "pods" source.
- // +optional
- Resource *ResourceMetricSource
- // External refers to a global metric that is not associated
- // with any Kubernetes object. It allows autoscaling based on information
- // coming from components running outside of cluster
- // (for example length of queue in cloud messaging service, or
- // QPS from loadbalancer running outside of cluster).
- // +optional
- External *ExternalMetricSource
-}
-
-// ObjectMetricSource indicates how to scale on a metric describing a
-// kubernetes object (for example, hits-per-second on an Ingress object).
-type ObjectMetricSource struct {
- // Target is the described Kubernetes object.
- Target CrossVersionObjectReference
-
- // MetricName is the name of the metric in question.
- MetricName string
- // TargetValue is the target value of the metric (as a quantity).
- TargetValue resource.Quantity
-}
-
-// PodsMetricSource indicates how to scale on a metric describing each pod in
-// the current scale target (for example, transactions-processed-per-second).
-// The values will be averaged together before being compared to the target
-// value.
-type PodsMetricSource struct {
- // MetricName is the name of the metric in question
- MetricName string
- // TargetAverageValue is the target value of the average of the
- // metric across all relevant pods (as a quantity)
- TargetAverageValue resource.Quantity
-}
-
-// ResourceMetricSource indicates how to scale on a resource metric known to
-// Kubernetes, as specified in requests and limits, describing each pod in the
-// current scale target (e.g. CPU or memory). The values will be averaged
-// together before being compared to the target. Such metrics are built in to
-// Kubernetes, and have special scaling options on top of those available to
-// normal per-pod metrics using the "pods" source. Only one "target" type
-// should be set.
-type ResourceMetricSource struct {
- // Name is the name of the resource in question.
- Name api.ResourceName
- // TargetAverageUtilization is the target value of the average of the
- // resource metric across all relevant pods, represented as a percentage of
- // the requested value of the resource for the pods.
- // +optional
- TargetAverageUtilization *int32
- // TargetAverageValue is the target value of the average of the
- // resource metric across all relevant pods, as a raw value (instead of as
- // a percentage of the request), similar to the "pods" metric source type.
- // +optional
- TargetAverageValue *resource.Quantity
-}
-
-// ExternalMetricSource indicates how to scale on a metric not associated with
-// any Kubernetes object (for example length of queue in cloud
-// messaging service, or QPS from loadbalancer running outside of cluster).
-type ExternalMetricSource struct {
- // metricName is the name of the metric in question.
- MetricName string
- // MetricSelector is used to identify a specific time series
- // within a given metric.
- // +optional
- MetricSelector *metav1.LabelSelector
- // TargetValue is the target value of the metric (as a quantity).
- // Mutually exclusive with TargetAverageValue.
- // +optional
- TargetValue *resource.Quantity
- // TargetAverageValue is the target per-pod value of global metric (as a quantity).
- // Mutually exclusive with TargetValue.
- // +optional
- TargetAverageValue *resource.Quantity
-}
-
-// HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler.
-type HorizontalPodAutoscalerStatus struct {
- // ObservedGeneration is the most recent generation observed by this autoscaler.
- // +optional
- ObservedGeneration *int64
-
- // LastScaleTime is the last time the HorizontalPodAutoscaler scaled the number of pods,
- // used by the autoscaler to control how often the number of pods is changed.
- // +optional
- LastScaleTime *metav1.Time
-
- // CurrentReplicas is current number of replicas of pods managed by this autoscaler,
- // as last seen by the autoscaler.
- CurrentReplicas int32
-
- // DesiredReplicas is the desired number of replicas of pods managed by this autoscaler,
- // as last calculated by the autoscaler.
- DesiredReplicas int32
-
- // CurrentMetrics is the last read state of the metrics used by this autoscaler.
- CurrentMetrics []MetricStatus
-
- // Conditions is the set of conditions required for this autoscaler to scale its target,
- // and indicates whether or not those conditions are met.
- Conditions []HorizontalPodAutoscalerCondition
-}
-
-// ConditionStatus indicates the status of a condition (true, false, or unknown).
-type ConditionStatus string
-
-// These are valid condition statuses. "ConditionTrue" means a resource is in the condition;
-// "ConditionFalse" means a resource is not in the condition; "ConditionUnknown" means kubernetes
-// can't decide if a resource is in the condition or not. In the future, we could add other
-// intermediate conditions, e.g. ConditionDegraded.
-const (
- ConditionTrue ConditionStatus = "True"
- ConditionFalse ConditionStatus = "False"
- ConditionUnknown ConditionStatus = "Unknown"
-)
-
-// HorizontalPodAutoscalerConditionType are the valid conditions of
-// a HorizontalPodAutoscaler.
-type HorizontalPodAutoscalerConditionType string
-
-var (
- // ScalingActive indicates that the HPA controller is able to scale if necessary:
- // it's correctly configured, can fetch the desired metrics, and isn't disabled.
- ScalingActive HorizontalPodAutoscalerConditionType = "ScalingActive"
- // AbleToScale indicates a lack of transient issues which prevent scaling from occurring,
- // such as being in a backoff window, or being unable to access/update the target scale.
- AbleToScale HorizontalPodAutoscalerConditionType = "AbleToScale"
- // ScalingLimited indicates that the calculated scale based on metrics would be above or
- // below the range for the HPA, and has thus been capped.
- ScalingLimited HorizontalPodAutoscalerConditionType = "ScalingLimited"
-)
-
-// HorizontalPodAutoscalerCondition describes the state of
-// a HorizontalPodAutoscaler at a certain point.
-type HorizontalPodAutoscalerCondition struct {
- // Type describes the current condition
- Type HorizontalPodAutoscalerConditionType
- // Status is the status of the condition (True, False, Unknown)
- Status ConditionStatus
- // LastTransitionTime is the last time the condition transitioned from
- // one status to another
- // +optional
- LastTransitionTime metav1.Time
- // Reason is the reason for the condition's last transition.
- // +optional
- Reason string
- // Message is a human-readable explanation containing details about
- // the transition
- // +optional
- Message string
-}
-
-// MetricStatus describes the last-read state of a single metric.
-type MetricStatus struct {
- // Type is the type of metric source. It will be one of "Object",
- // "Pods" or "Resource", each corresponds to a matching field in the object.
- Type MetricSourceType
-
- // Object refers to a metric describing a single kubernetes object
- // (for example, hits-per-second on an Ingress object).
- // +optional
- Object *ObjectMetricStatus
- // Pods refers to a metric describing each pod in the current scale target
- // (for example, transactions-processed-per-second). The values will be
- // averaged together before being compared to the target value.
- // +optional
- Pods *PodsMetricStatus
- // Resource refers to a resource metric (such as those specified in
- // requests and limits) known to Kubernetes describing each pod in the
- // current scale target (e.g. CPU or memory). Such metrics are built in to
- // Kubernetes, and have special scaling options on top of those available
- // to normal per-pod metrics using the "pods" source.
- // +optional
- Resource *ResourceMetricStatus
- // External refers to a global metric that is not associated
- // with any Kubernetes object. It allows autoscaling based on information
- // coming from components running outside of cluster
- // (for example length of queue in cloud messaging service, or
- // QPS from loadbalancer running outside of cluster).
- // +optional
- External *ExternalMetricStatus
-}
-
-// ObjectMetricStatus indicates the current value of a metric describing a
-// kubernetes object (for example, hits-per-second on an Ingress object).
-type ObjectMetricStatus struct {
- // Target is the described Kubernetes object.
- Target CrossVersionObjectReference
-
- // MetricName is the name of the metric in question.
- MetricName string
- // CurrentValue is the current value of the metric (as a quantity).
- CurrentValue resource.Quantity
-}
-
-// PodsMetricStatus indicates the current value of a metric describing each pod in
-// the current scale target (for example, transactions-processed-per-second).
-type PodsMetricStatus struct {
- // MetricName is the name of the metric in question
- MetricName string
- // CurrentAverageValue is the current value of the average of the
- // metric across all relevant pods (as a quantity)
- CurrentAverageValue resource.Quantity
-}
-
-// ResourceMetricStatus indicates the current value of a resource metric known to
-// Kubernetes, as specified in requests and limits, describing each pod in the
-// current scale target (e.g. CPU or memory). Such metrics are built in to
-// Kubernetes, and have special scaling options on top of those available to
-// normal per-pod metrics using the "pods" source.
-type ResourceMetricStatus struct {
- // Name is the name of the resource in question.
- Name api.ResourceName
- // CurrentAverageUtilization is the current value of the average of the
- // resource metric across all relevant pods, represented as a percentage of
- // the requested value of the resource for the pods. It will only be
- // present if `targetAverageValue` was set in the corresponding metric
- // specification.
- // +optional
- CurrentAverageUtilization *int32
- // CurrentAverageValue is the current value of the average of the
- // resource metric across all relevant pods, as a raw value (instead of as
- // a percentage of the request), similar to the "pods" metric source type.
- // It will always be set, regardless of the corresponding metric specification.
- CurrentAverageValue resource.Quantity
-}
-
-// ExternalMetricStatus indicates the current value of a global metric
-// not associated with any Kubernetes object.
-type ExternalMetricStatus struct {
- // MetricName is the name of a metric used for autoscaling in
- // metric system.
- MetricName string
- // MetricSelector is used to identify a specific time series
- // within a given metric.
- // +optional
- MetricSelector *metav1.LabelSelector
- // CurrentValue is the current value of the metric (as a quantity)
- CurrentValue resource.Quantity
- // CurrentAverageValue is the current value of metric averaged over autoscaled pods.
- // +optional
- CurrentAverageValue *resource.Quantity
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// HorizontalPodAutoscaler is the configuration for a horizontal pod
-// autoscaler, which automatically manages the replica count of any resource
-// implementing the scale subresource based on the metrics specified.
-type HorizontalPodAutoscaler struct {
- metav1.TypeMeta
- // Metadata is the standard object metadata.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- metav1.ObjectMeta
-
- // Spec is the specification for the behaviour of the autoscaler.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status.
- // +optional
- Spec HorizontalPodAutoscalerSpec
-
- // Status is the current information about the autoscaler.
- // +optional
- Status HorizontalPodAutoscalerStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// HorizontalPodAutoscalerList is a list of horizontal pod autoscaler objects.
-type HorizontalPodAutoscalerList struct {
- metav1.TypeMeta
- // Metadata is the standard list metadata.
- // +optional
- metav1.ListMeta
-
- // Items is the list of horizontal pod autoscaler objects.
- Items []HorizontalPodAutoscaler
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/zz_generated.deepcopy.go b/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/zz_generated.deepcopy.go
deleted file mode 100644
index c01971384..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/autoscaling/zz_generated.deepcopy.go
+++ /dev/null
@@ -1,572 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by deepcopy-gen. DO NOT EDIT.
-
-package autoscaling
-
-import (
- v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
-)
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CrossVersionObjectReference) DeepCopyInto(out *CrossVersionObjectReference) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CrossVersionObjectReference.
-func (in *CrossVersionObjectReference) DeepCopy() *CrossVersionObjectReference {
- if in == nil {
- return nil
- }
- out := new(CrossVersionObjectReference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExternalMetricSource) DeepCopyInto(out *ExternalMetricSource) {
- *out = *in
- if in.MetricSelector != nil {
- in, out := &in.MetricSelector, &out.MetricSelector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.TargetValue != nil {
- in, out := &in.TargetValue, &out.TargetValue
- if *in == nil {
- *out = nil
- } else {
- x := (*in).DeepCopy()
- *out = &x
- }
- }
- if in.TargetAverageValue != nil {
- in, out := &in.TargetAverageValue, &out.TargetAverageValue
- if *in == nil {
- *out = nil
- } else {
- x := (*in).DeepCopy()
- *out = &x
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExternalMetricSource.
-func (in *ExternalMetricSource) DeepCopy() *ExternalMetricSource {
- if in == nil {
- return nil
- }
- out := new(ExternalMetricSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExternalMetricStatus) DeepCopyInto(out *ExternalMetricStatus) {
- *out = *in
- if in.MetricSelector != nil {
- in, out := &in.MetricSelector, &out.MetricSelector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- out.CurrentValue = in.CurrentValue.DeepCopy()
- if in.CurrentAverageValue != nil {
- in, out := &in.CurrentAverageValue, &out.CurrentAverageValue
- if *in == nil {
- *out = nil
- } else {
- x := (*in).DeepCopy()
- *out = &x
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExternalMetricStatus.
-func (in *ExternalMetricStatus) DeepCopy() *ExternalMetricStatus {
- if in == nil {
- return nil
- }
- out := new(ExternalMetricStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HorizontalPodAutoscaler) DeepCopyInto(out *HorizontalPodAutoscaler) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscaler.
-func (in *HorizontalPodAutoscaler) DeepCopy() *HorizontalPodAutoscaler {
- if in == nil {
- return nil
- }
- out := new(HorizontalPodAutoscaler)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *HorizontalPodAutoscaler) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HorizontalPodAutoscalerCondition) DeepCopyInto(out *HorizontalPodAutoscalerCondition) {
- *out = *in
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscalerCondition.
-func (in *HorizontalPodAutoscalerCondition) DeepCopy() *HorizontalPodAutoscalerCondition {
- if in == nil {
- return nil
- }
- out := new(HorizontalPodAutoscalerCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HorizontalPodAutoscalerList) DeepCopyInto(out *HorizontalPodAutoscalerList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]HorizontalPodAutoscaler, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscalerList.
-func (in *HorizontalPodAutoscalerList) DeepCopy() *HorizontalPodAutoscalerList {
- if in == nil {
- return nil
- }
- out := new(HorizontalPodAutoscalerList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *HorizontalPodAutoscalerList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HorizontalPodAutoscalerSpec) DeepCopyInto(out *HorizontalPodAutoscalerSpec) {
- *out = *in
- out.ScaleTargetRef = in.ScaleTargetRef
- if in.MinReplicas != nil {
- in, out := &in.MinReplicas, &out.MinReplicas
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.Metrics != nil {
- in, out := &in.Metrics, &out.Metrics
- *out = make([]MetricSpec, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscalerSpec.
-func (in *HorizontalPodAutoscalerSpec) DeepCopy() *HorizontalPodAutoscalerSpec {
- if in == nil {
- return nil
- }
- out := new(HorizontalPodAutoscalerSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HorizontalPodAutoscalerStatus) DeepCopyInto(out *HorizontalPodAutoscalerStatus) {
- *out = *in
- if in.ObservedGeneration != nil {
- in, out := &in.ObservedGeneration, &out.ObservedGeneration
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.LastScaleTime != nil {
- in, out := &in.LastScaleTime, &out.LastScaleTime
- if *in == nil {
- *out = nil
- } else {
- *out = (*in).DeepCopy()
- }
- }
- if in.CurrentMetrics != nil {
- in, out := &in.CurrentMetrics, &out.CurrentMetrics
- *out = make([]MetricStatus, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]HorizontalPodAutoscalerCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscalerStatus.
-func (in *HorizontalPodAutoscalerStatus) DeepCopy() *HorizontalPodAutoscalerStatus {
- if in == nil {
- return nil
- }
- out := new(HorizontalPodAutoscalerStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *MetricSpec) DeepCopyInto(out *MetricSpec) {
- *out = *in
- if in.Object != nil {
- in, out := &in.Object, &out.Object
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectMetricSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Pods != nil {
- in, out := &in.Pods, &out.Pods
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodsMetricSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Resource != nil {
- in, out := &in.Resource, &out.Resource
- if *in == nil {
- *out = nil
- } else {
- *out = new(ResourceMetricSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.External != nil {
- in, out := &in.External, &out.External
- if *in == nil {
- *out = nil
- } else {
- *out = new(ExternalMetricSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MetricSpec.
-func (in *MetricSpec) DeepCopy() *MetricSpec {
- if in == nil {
- return nil
- }
- out := new(MetricSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *MetricStatus) DeepCopyInto(out *MetricStatus) {
- *out = *in
- if in.Object != nil {
- in, out := &in.Object, &out.Object
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectMetricStatus)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Pods != nil {
- in, out := &in.Pods, &out.Pods
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodsMetricStatus)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Resource != nil {
- in, out := &in.Resource, &out.Resource
- if *in == nil {
- *out = nil
- } else {
- *out = new(ResourceMetricStatus)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.External != nil {
- in, out := &in.External, &out.External
- if *in == nil {
- *out = nil
- } else {
- *out = new(ExternalMetricStatus)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new MetricStatus.
-func (in *MetricStatus) DeepCopy() *MetricStatus {
- if in == nil {
- return nil
- }
- out := new(MetricStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ObjectMetricSource) DeepCopyInto(out *ObjectMetricSource) {
- *out = *in
- out.Target = in.Target
- out.TargetValue = in.TargetValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ObjectMetricSource.
-func (in *ObjectMetricSource) DeepCopy() *ObjectMetricSource {
- if in == nil {
- return nil
- }
- out := new(ObjectMetricSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ObjectMetricStatus) DeepCopyInto(out *ObjectMetricStatus) {
- *out = *in
- out.Target = in.Target
- out.CurrentValue = in.CurrentValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ObjectMetricStatus.
-func (in *ObjectMetricStatus) DeepCopy() *ObjectMetricStatus {
- if in == nil {
- return nil
- }
- out := new(ObjectMetricStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodsMetricSource) DeepCopyInto(out *PodsMetricSource) {
- *out = *in
- out.TargetAverageValue = in.TargetAverageValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodsMetricSource.
-func (in *PodsMetricSource) DeepCopy() *PodsMetricSource {
- if in == nil {
- return nil
- }
- out := new(PodsMetricSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodsMetricStatus) DeepCopyInto(out *PodsMetricStatus) {
- *out = *in
- out.CurrentAverageValue = in.CurrentAverageValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodsMetricStatus.
-func (in *PodsMetricStatus) DeepCopy() *PodsMetricStatus {
- if in == nil {
- return nil
- }
- out := new(PodsMetricStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceMetricSource) DeepCopyInto(out *ResourceMetricSource) {
- *out = *in
- if in.TargetAverageUtilization != nil {
- in, out := &in.TargetAverageUtilization, &out.TargetAverageUtilization
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.TargetAverageValue != nil {
- in, out := &in.TargetAverageValue, &out.TargetAverageValue
- if *in == nil {
- *out = nil
- } else {
- x := (*in).DeepCopy()
- *out = &x
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceMetricSource.
-func (in *ResourceMetricSource) DeepCopy() *ResourceMetricSource {
- if in == nil {
- return nil
- }
- out := new(ResourceMetricSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceMetricStatus) DeepCopyInto(out *ResourceMetricStatus) {
- *out = *in
- if in.CurrentAverageUtilization != nil {
- in, out := &in.CurrentAverageUtilization, &out.CurrentAverageUtilization
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- out.CurrentAverageValue = in.CurrentAverageValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceMetricStatus.
-func (in *ResourceMetricStatus) DeepCopy() *ResourceMetricStatus {
- if in == nil {
- return nil
- }
- out := new(ResourceMetricStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Scale) DeepCopyInto(out *Scale) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- out.Spec = in.Spec
- out.Status = in.Status
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Scale.
-func (in *Scale) DeepCopy() *Scale {
- if in == nil {
- return nil
- }
- out := new(Scale)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Scale) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ScaleSpec) DeepCopyInto(out *ScaleSpec) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ScaleSpec.
-func (in *ScaleSpec) DeepCopy() *ScaleSpec {
- if in == nil {
- return nil
- }
- out := new(ScaleSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ScaleStatus) DeepCopyInto(out *ScaleStatus) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ScaleStatus.
-func (in *ScaleStatus) DeepCopy() *ScaleStatus {
- if in == nil {
- return nil
- }
- out := new(ScaleStatus)
- in.DeepCopyInto(out)
- return out
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/annotation_key_constants.go b/vendor/k8s.io/kubernetes/pkg/apis/core/annotation_key_constants.go
deleted file mode 100644
index 131fdd990..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/annotation_key_constants.go
+++ /dev/null
@@ -1,92 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// This file should be consistent with pkg/api/v1/annotation_key_constants.go.
-
-package core
-
-const (
- // ImagePolicyFailedOpenKey is added to pods created by failing open when the image policy
- // webhook backend fails.
- ImagePolicyFailedOpenKey string = "alpha.image-policy.k8s.io/failed-open"
-
- // PodPresetOptOutAnnotationKey represents the annotation key for a pod to exempt itself from pod preset manipulation
- PodPresetOptOutAnnotationKey string = "podpreset.admission.kubernetes.io/exclude"
-
- // MirrorAnnotationKey represents the annotation key set by kubelets when creating mirror pods
- MirrorPodAnnotationKey string = "kubernetes.io/config.mirror"
-
- // TolerationsAnnotationKey represents the key of tolerations data (json serialized)
- // in the Annotations of a Pod.
- TolerationsAnnotationKey string = "scheduler.alpha.kubernetes.io/tolerations"
-
- // TaintsAnnotationKey represents the key of taints data (json serialized)
- // in the Annotations of a Node.
- TaintsAnnotationKey string = "scheduler.alpha.kubernetes.io/taints"
-
- // SeccompPodAnnotationKey represents the key of a seccomp profile applied
- // to all containers of a pod.
- SeccompPodAnnotationKey string = "seccomp.security.alpha.kubernetes.io/pod"
-
- // SeccompContainerAnnotationKeyPrefix represents the key of a seccomp profile applied
- // to one container of a pod.
- SeccompContainerAnnotationKeyPrefix string = "container.seccomp.security.alpha.kubernetes.io/"
-
- // PreferAvoidPodsAnnotationKey represents the key of preferAvoidPods data (json serialized)
- // in the Annotations of a Node.
- PreferAvoidPodsAnnotationKey string = "scheduler.alpha.kubernetes.io/preferAvoidPods"
-
- // SysctlsPodAnnotationKey represents the key of sysctls which are set for the infrastructure
- // container of a pod. The annotation value is a comma separated list of sysctl_name=value
- // key-value pairs. Only a limited set of whitelisted and isolated sysctls is supported by
- // the kubelet. Pods with other sysctls will fail to launch.
- SysctlsPodAnnotationKey string = "security.alpha.kubernetes.io/sysctls"
-
- // UnsafeSysctlsPodAnnotationKey represents the key of sysctls which are set for the infrastructure
- // container of a pod. The annotation value is a comma separated list of sysctl_name=value
- // key-value pairs. Unsafe sysctls must be explicitly enabled for a kubelet. They are properly
- // namespaced to a pod or a container, but their isolation is usually unclear or weak. Their use
- // is at-your-own-risk. Pods that attempt to set an unsafe sysctl that is not enabled for a kubelet
- // will fail to launch.
- UnsafeSysctlsPodAnnotationKey string = "security.alpha.kubernetes.io/unsafe-sysctls"
-
- // ObjectTTLAnnotations represents a suggestion for kubelet for how long it can cache
- // an object (e.g. secret, config map) before fetching it again from apiserver.
- // This annotation can be attached to node.
- ObjectTTLAnnotationKey string = "node.alpha.kubernetes.io/ttl"
-
- // BootstrapCheckpointAnnotationKey represents a Resource (Pod) that should be checkpointed by
- // the kubelet prior to running
- BootstrapCheckpointAnnotationKey string = "node.kubernetes.io/bootstrap-checkpoint"
-
- // annotation key prefix used to identify non-convertible json paths.
- NonConvertibleAnnotationPrefix = "non-convertible.kubernetes.io"
-
- kubectlPrefix = "kubectl.kubernetes.io/"
-
- // LastAppliedConfigAnnotation is the annotation used to store the previous
- // configuration of a resource for use in a three way diff by UpdateApplyAnnotation.
- LastAppliedConfigAnnotation = kubectlPrefix + "last-applied-configuration"
-
- // AnnotationLoadBalancerSourceRangesKey is the key of the annotation on a service to set allowed ingress ranges on their LoadBalancers
- //
- // It should be a comma-separated list of CIDRs, e.g. `0.0.0.0/0` to
- // allow full access (the default) or `18.0.0.0/8,56.0.0.0/8` to allow
- // access only from the CIDRs currently allocated to MIT & the USPS.
- //
- // Not all cloud providers support this annotation, though AWS & GCE do.
- AnnotationLoadBalancerSourceRangesKey = "service.beta.kubernetes.io/load-balancer-source-ranges"
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/core/doc.go
deleted file mode 100644
index 6017bfdab..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/doc.go
+++ /dev/null
@@ -1,24 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:deepcopy-gen=package
-
-// Package api contains the latest (or "internal") version of the
-// Kubernetes API objects. This is the API objects as represented in memory.
-// The contract presented to clients is located in the versioned packages,
-// which are sub-directories. The first one is "v1". Those packages
-// describe how a particular version is serialized to storage/network.
-package core // import "k8s.io/kubernetes/pkg/apis/core"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/field_constants.go b/vendor/k8s.io/kubernetes/pkg/apis/core/field_constants.go
deleted file mode 100644
index a26f80568..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/field_constants.go
+++ /dev/null
@@ -1,38 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package core
-
-// Field path constants that are specific to the internal API
-// representation.
-const (
- NodeUnschedulableField = "spec.unschedulable"
- ObjectNameField = "metadata.name"
- PodHostField = "spec.nodeName"
- PodStatusField = "status.phase"
- SecretTypeField = "type"
-
- EventReasonField = "action"
- EventSourceField = "reportingComponent"
- EventTypeField = "type"
- EventInvolvedKindField = "involvedObject.kind"
- EventInvolvedNamespaceField = "involvedObject.namespace"
- EventInvolvedNameField = "involvedObject.name"
- EventInvolvedUIDField = "involvedObject.uid"
- EventInvolvedAPIVersionField = "involvedObject.apiVersion"
- EventInvolvedResourceVersionField = "involvedObject.resourceVersion"
- EventInvolvedFieldPathField = "involvedObject.fieldPath"
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/helper/helpers.go b/vendor/k8s.io/kubernetes/pkg/apis/core/helper/helpers.go
deleted file mode 100644
index 96899f184..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/helper/helpers.go
+++ /dev/null
@@ -1,597 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package helper
-
-import (
- "encoding/json"
- "fmt"
- "strings"
-
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/conversion"
- "k8s.io/apimachinery/pkg/fields"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/selection"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/validation"
- "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// IsHugePageResourceName returns true if the resource name has the huge page
-// resource prefix.
-func IsHugePageResourceName(name core.ResourceName) bool {
- return strings.HasPrefix(string(name), core.ResourceHugePagesPrefix)
-}
-
-// IsQuotaHugePageResourceName returns true if the resource name has the quota
-// related huge page resource prefix.
-func IsQuotaHugePageResourceName(name core.ResourceName) bool {
- return strings.HasPrefix(string(name), core.ResourceHugePagesPrefix) || strings.HasPrefix(string(name), core.ResourceRequestsHugePagesPrefix)
-}
-
-// HugePageResourceName returns a ResourceName with the canonical hugepage
-// prefix prepended for the specified page size. The page size is converted
-// to its canonical representation.
-func HugePageResourceName(pageSize resource.Quantity) core.ResourceName {
- return core.ResourceName(fmt.Sprintf("%s%s", core.ResourceHugePagesPrefix, pageSize.String()))
-}
-
-// HugePageSizeFromResourceName returns the page size for the specified huge page
-// resource name. If the specified input is not a valid huge page resource name
-// an error is returned.
-func HugePageSizeFromResourceName(name core.ResourceName) (resource.Quantity, error) {
- if !IsHugePageResourceName(name) {
- return resource.Quantity{}, fmt.Errorf("resource name: %s is not valid hugepage name", name)
- }
- pageSize := strings.TrimPrefix(string(name), core.ResourceHugePagesPrefix)
- return resource.ParseQuantity(pageSize)
-}
-
-// NonConvertibleFields iterates over the provided map and filters out all but
-// any keys with the "non-convertible.kubernetes.io" prefix.
-func NonConvertibleFields(annotations map[string]string) map[string]string {
- nonConvertibleKeys := map[string]string{}
- for key, value := range annotations {
- if strings.HasPrefix(key, core.NonConvertibleAnnotationPrefix) {
- nonConvertibleKeys[key] = value
- }
- }
- return nonConvertibleKeys
-}
-
-// Semantic can do semantic deep equality checks for core objects.
-// Example: apiequality.Semantic.DeepEqual(aPod, aPodWithNonNilButEmptyMaps) == true
-var Semantic = conversion.EqualitiesOrDie(
- func(a, b resource.Quantity) bool {
- // Ignore formatting, only care that numeric value stayed the same.
- // TODO: if we decide it's important, it should be safe to start comparing the format.
- //
- // Uninitialized quantities are equivalent to 0 quantities.
- return a.Cmp(b) == 0
- },
- func(a, b metav1.MicroTime) bool {
- return a.UTC() == b.UTC()
- },
- func(a, b metav1.Time) bool {
- return a.UTC() == b.UTC()
- },
- func(a, b labels.Selector) bool {
- return a.String() == b.String()
- },
- func(a, b fields.Selector) bool {
- return a.String() == b.String()
- },
-)
-
-var standardResourceQuotaScopes = sets.NewString(
- string(core.ResourceQuotaScopeTerminating),
- string(core.ResourceQuotaScopeNotTerminating),
- string(core.ResourceQuotaScopeBestEffort),
- string(core.ResourceQuotaScopeNotBestEffort),
-)
-
-// IsStandardResourceQuotaScope returns true if the scope is a standard value
-func IsStandardResourceQuotaScope(str string) bool {
- return standardResourceQuotaScopes.Has(str)
-}
-
-var podObjectCountQuotaResources = sets.NewString(
- string(core.ResourcePods),
-)
-
-var podComputeQuotaResources = sets.NewString(
- string(core.ResourceCPU),
- string(core.ResourceMemory),
- string(core.ResourceLimitsCPU),
- string(core.ResourceLimitsMemory),
- string(core.ResourceRequestsCPU),
- string(core.ResourceRequestsMemory),
-)
-
-// IsResourceQuotaScopeValidForResource returns true if the resource applies to the specified scope
-func IsResourceQuotaScopeValidForResource(scope core.ResourceQuotaScope, resource string) bool {
- switch scope {
- case core.ResourceQuotaScopeTerminating, core.ResourceQuotaScopeNotTerminating, core.ResourceQuotaScopeNotBestEffort:
- return podObjectCountQuotaResources.Has(resource) || podComputeQuotaResources.Has(resource)
- case core.ResourceQuotaScopeBestEffort:
- return podObjectCountQuotaResources.Has(resource)
- default:
- return true
- }
-}
-
-var standardContainerResources = sets.NewString(
- string(core.ResourceCPU),
- string(core.ResourceMemory),
- string(core.ResourceEphemeralStorage),
-)
-
-// IsStandardContainerResourceName returns true if the container can make a resource request
-// for the specified resource
-func IsStandardContainerResourceName(str string) bool {
- return standardContainerResources.Has(str) || IsHugePageResourceName(core.ResourceName(str))
-}
-
-// IsExtendedResourceName returns true if:
-// 1. the resource name is not in the default namespace;
-// 2. resource name does not have "requests." prefix,
-// to avoid confusion with the convention in quota
-// 3. it satisfies the rules in IsQualifiedName() after converted into quota resource name
-func IsExtendedResourceName(name core.ResourceName) bool {
- if IsDefaultNamespaceResource(name) || strings.HasPrefix(string(name), core.DefaultResourceRequestsPrefix) {
- return false
- }
- // Ensure it satisfies the rules in IsQualifiedName() after converted into quota resource name
- nameForQuota := fmt.Sprintf("%s%s", core.DefaultResourceRequestsPrefix, string(name))
- if errs := validation.IsQualifiedName(string(nameForQuota)); len(errs) != 0 {
- return false
- }
- return true
-}
-
-// IsDefaultNamespaceResource returns true if the resource name is in the
-// *kubernetes.io/ namespace. Partially-qualified (unprefixed) names are
-// implicitly in the kubernetes.io/ namespace.
-func IsDefaultNamespaceResource(name core.ResourceName) bool {
- return !strings.Contains(string(name), "/") ||
- strings.Contains(string(name), core.ResourceDefaultNamespacePrefix)
-}
-
-var overcommitBlacklist = sets.NewString(string(core.ResourceNvidiaGPU))
-
-// IsOvercommitAllowed returns true if the resource is in the default
-// namespace and not blacklisted.
-func IsOvercommitAllowed(name core.ResourceName) bool {
- return IsDefaultNamespaceResource(name) &&
- !IsHugePageResourceName(name) &&
- !overcommitBlacklist.Has(string(name))
-}
-
-var standardLimitRangeTypes = sets.NewString(
- string(core.LimitTypePod),
- string(core.LimitTypeContainer),
- string(core.LimitTypePersistentVolumeClaim),
-)
-
-// IsStandardLimitRangeType returns true if the type is Pod or Container
-func IsStandardLimitRangeType(str string) bool {
- return standardLimitRangeTypes.Has(str)
-}
-
-var standardQuotaResources = sets.NewString(
- string(core.ResourceCPU),
- string(core.ResourceMemory),
- string(core.ResourceEphemeralStorage),
- string(core.ResourceRequestsCPU),
- string(core.ResourceRequestsMemory),
- string(core.ResourceRequestsStorage),
- string(core.ResourceRequestsEphemeralStorage),
- string(core.ResourceLimitsCPU),
- string(core.ResourceLimitsMemory),
- string(core.ResourceLimitsEphemeralStorage),
- string(core.ResourcePods),
- string(core.ResourceQuotas),
- string(core.ResourceServices),
- string(core.ResourceReplicationControllers),
- string(core.ResourceSecrets),
- string(core.ResourcePersistentVolumeClaims),
- string(core.ResourceConfigMaps),
- string(core.ResourceServicesNodePorts),
- string(core.ResourceServicesLoadBalancers),
-)
-
-// IsStandardQuotaResourceName returns true if the resource is known to
-// the quota tracking system
-func IsStandardQuotaResourceName(str string) bool {
- return standardQuotaResources.Has(str) || IsQuotaHugePageResourceName(core.ResourceName(str))
-}
-
-var standardResources = sets.NewString(
- string(core.ResourceCPU),
- string(core.ResourceMemory),
- string(core.ResourceEphemeralStorage),
- string(core.ResourceRequestsCPU),
- string(core.ResourceRequestsMemory),
- string(core.ResourceRequestsEphemeralStorage),
- string(core.ResourceLimitsCPU),
- string(core.ResourceLimitsMemory),
- string(core.ResourceLimitsEphemeralStorage),
- string(core.ResourcePods),
- string(core.ResourceQuotas),
- string(core.ResourceServices),
- string(core.ResourceReplicationControllers),
- string(core.ResourceSecrets),
- string(core.ResourceConfigMaps),
- string(core.ResourcePersistentVolumeClaims),
- string(core.ResourceStorage),
- string(core.ResourceRequestsStorage),
- string(core.ResourceServicesNodePorts),
- string(core.ResourceServicesLoadBalancers),
-)
-
-// IsStandardResourceName returns true if the resource is known to the system
-func IsStandardResourceName(str string) bool {
- return standardResources.Has(str) || IsQuotaHugePageResourceName(core.ResourceName(str))
-}
-
-var integerResources = sets.NewString(
- string(core.ResourcePods),
- string(core.ResourceQuotas),
- string(core.ResourceServices),
- string(core.ResourceReplicationControllers),
- string(core.ResourceSecrets),
- string(core.ResourceConfigMaps),
- string(core.ResourcePersistentVolumeClaims),
- string(core.ResourceServicesNodePorts),
- string(core.ResourceServicesLoadBalancers),
-)
-
-// IsIntegerResourceName returns true if the resource is measured in integer values
-func IsIntegerResourceName(str string) bool {
- return integerResources.Has(str) || IsExtendedResourceName(core.ResourceName(str))
-}
-
-// this function aims to check if the service's ClusterIP is set or not
-// the objective is not to perform validation here
-func IsServiceIPSet(service *core.Service) bool {
- return service.Spec.ClusterIP != core.ClusterIPNone && service.Spec.ClusterIP != ""
-}
-
-var standardFinalizers = sets.NewString(
- string(core.FinalizerKubernetes),
- metav1.FinalizerOrphanDependents,
- metav1.FinalizerDeleteDependents,
-)
-
-func IsStandardFinalizerName(str string) bool {
- return standardFinalizers.Has(str)
-}
-
-// AddToNodeAddresses appends the NodeAddresses to the passed-by-pointer slice,
-// only if they do not already exist
-func AddToNodeAddresses(addresses *[]core.NodeAddress, addAddresses ...core.NodeAddress) {
- for _, add := range addAddresses {
- exists := false
- for _, existing := range *addresses {
- if existing.Address == add.Address && existing.Type == add.Type {
- exists = true
- break
- }
- }
- if !exists {
- *addresses = append(*addresses, add)
- }
- }
-}
-
-// TODO: make method on LoadBalancerStatus?
-func LoadBalancerStatusEqual(l, r *core.LoadBalancerStatus) bool {
- return ingressSliceEqual(l.Ingress, r.Ingress)
-}
-
-func ingressSliceEqual(lhs, rhs []core.LoadBalancerIngress) bool {
- if len(lhs) != len(rhs) {
- return false
- }
- for i := range lhs {
- if !ingressEqual(&lhs[i], &rhs[i]) {
- return false
- }
- }
- return true
-}
-
-func ingressEqual(lhs, rhs *core.LoadBalancerIngress) bool {
- if lhs.IP != rhs.IP {
- return false
- }
- if lhs.Hostname != rhs.Hostname {
- return false
- }
- return true
-}
-
-// TODO: make method on LoadBalancerStatus?
-func LoadBalancerStatusDeepCopy(lb *core.LoadBalancerStatus) *core.LoadBalancerStatus {
- c := &core.LoadBalancerStatus{}
- c.Ingress = make([]core.LoadBalancerIngress, len(lb.Ingress))
- for i := range lb.Ingress {
- c.Ingress[i] = lb.Ingress[i]
- }
- return c
-}
-
-// GetAccessModesAsString returns a string representation of an array of access modes.
-// modes, when present, are always in the same order: RWO,ROX,RWX.
-func GetAccessModesAsString(modes []core.PersistentVolumeAccessMode) string {
- modes = removeDuplicateAccessModes(modes)
- modesStr := []string{}
- if containsAccessMode(modes, core.ReadWriteOnce) {
- modesStr = append(modesStr, "RWO")
- }
- if containsAccessMode(modes, core.ReadOnlyMany) {
- modesStr = append(modesStr, "ROX")
- }
- if containsAccessMode(modes, core.ReadWriteMany) {
- modesStr = append(modesStr, "RWX")
- }
- return strings.Join(modesStr, ",")
-}
-
-// GetAccessModesAsString returns an array of AccessModes from a string created by GetAccessModesAsString
-func GetAccessModesFromString(modes string) []core.PersistentVolumeAccessMode {
- strmodes := strings.Split(modes, ",")
- accessModes := []core.PersistentVolumeAccessMode{}
- for _, s := range strmodes {
- s = strings.Trim(s, " ")
- switch {
- case s == "RWO":
- accessModes = append(accessModes, core.ReadWriteOnce)
- case s == "ROX":
- accessModes = append(accessModes, core.ReadOnlyMany)
- case s == "RWX":
- accessModes = append(accessModes, core.ReadWriteMany)
- }
- }
- return accessModes
-}
-
-// removeDuplicateAccessModes returns an array of access modes without any duplicates
-func removeDuplicateAccessModes(modes []core.PersistentVolumeAccessMode) []core.PersistentVolumeAccessMode {
- accessModes := []core.PersistentVolumeAccessMode{}
- for _, m := range modes {
- if !containsAccessMode(accessModes, m) {
- accessModes = append(accessModes, m)
- }
- }
- return accessModes
-}
-
-func containsAccessMode(modes []core.PersistentVolumeAccessMode, mode core.PersistentVolumeAccessMode) bool {
- for _, m := range modes {
- if m == mode {
- return true
- }
- }
- return false
-}
-
-// NodeSelectorRequirementsAsSelector converts the []NodeSelectorRequirement core type into a struct that implements
-// labels.Selector.
-func NodeSelectorRequirementsAsSelector(nsm []core.NodeSelectorRequirement) (labels.Selector, error) {
- if len(nsm) == 0 {
- return labels.Nothing(), nil
- }
- selector := labels.NewSelector()
- for _, expr := range nsm {
- var op selection.Operator
- switch expr.Operator {
- case core.NodeSelectorOpIn:
- op = selection.In
- case core.NodeSelectorOpNotIn:
- op = selection.NotIn
- case core.NodeSelectorOpExists:
- op = selection.Exists
- case core.NodeSelectorOpDoesNotExist:
- op = selection.DoesNotExist
- case core.NodeSelectorOpGt:
- op = selection.GreaterThan
- case core.NodeSelectorOpLt:
- op = selection.LessThan
- default:
- return nil, fmt.Errorf("%q is not a valid node selector operator", expr.Operator)
- }
- r, err := labels.NewRequirement(expr.Key, op, expr.Values)
- if err != nil {
- return nil, err
- }
- selector = selector.Add(*r)
- }
- return selector, nil
-}
-
-// GetTolerationsFromPodAnnotations gets the json serialized tolerations data from Pod.Annotations
-// and converts it to the []Toleration type in core.
-func GetTolerationsFromPodAnnotations(annotations map[string]string) ([]core.Toleration, error) {
- var tolerations []core.Toleration
- if len(annotations) > 0 && annotations[core.TolerationsAnnotationKey] != "" {
- err := json.Unmarshal([]byte(annotations[core.TolerationsAnnotationKey]), &tolerations)
- if err != nil {
- return tolerations, err
- }
- }
- return tolerations, nil
-}
-
-// AddOrUpdateTolerationInPod tries to add a toleration to the pod's toleration list.
-// Returns true if something was updated, false otherwise.
-func AddOrUpdateTolerationInPod(pod *core.Pod, toleration *core.Toleration) bool {
- podTolerations := pod.Spec.Tolerations
-
- var newTolerations []core.Toleration
- updated := false
- for i := range podTolerations {
- if toleration.MatchToleration(&podTolerations[i]) {
- if Semantic.DeepEqual(toleration, podTolerations[i]) {
- return false
- }
- newTolerations = append(newTolerations, *toleration)
- updated = true
- continue
- }
-
- newTolerations = append(newTolerations, podTolerations[i])
- }
-
- if !updated {
- newTolerations = append(newTolerations, *toleration)
- }
-
- pod.Spec.Tolerations = newTolerations
- return true
-}
-
-// GetTaintsFromNodeAnnotations gets the json serialized taints data from Pod.Annotations
-// and converts it to the []Taint type in core.
-func GetTaintsFromNodeAnnotations(annotations map[string]string) ([]core.Taint, error) {
- var taints []core.Taint
- if len(annotations) > 0 && annotations[core.TaintsAnnotationKey] != "" {
- err := json.Unmarshal([]byte(annotations[core.TaintsAnnotationKey]), &taints)
- if err != nil {
- return []core.Taint{}, err
- }
- }
- return taints, nil
-}
-
-// SysctlsFromPodAnnotations parses the sysctl annotations into a slice of safe Sysctls
-// and a slice of unsafe Sysctls. This is only a convenience wrapper around
-// SysctlsFromPodAnnotation.
-func SysctlsFromPodAnnotations(a map[string]string) ([]core.Sysctl, []core.Sysctl, error) {
- safe, err := SysctlsFromPodAnnotation(a[core.SysctlsPodAnnotationKey])
- if err != nil {
- return nil, nil, err
- }
- unsafe, err := SysctlsFromPodAnnotation(a[core.UnsafeSysctlsPodAnnotationKey])
- if err != nil {
- return nil, nil, err
- }
-
- return safe, unsafe, nil
-}
-
-// SysctlsFromPodAnnotation parses an annotation value into a slice of Sysctls.
-func SysctlsFromPodAnnotation(annotation string) ([]core.Sysctl, error) {
- if len(annotation) == 0 {
- return nil, nil
- }
-
- kvs := strings.Split(annotation, ",")
- sysctls := make([]core.Sysctl, len(kvs))
- for i, kv := range kvs {
- cs := strings.Split(kv, "=")
- if len(cs) != 2 || len(cs[0]) == 0 {
- return nil, fmt.Errorf("sysctl %q not of the format sysctl_name=value", kv)
- }
- sysctls[i].Name = cs[0]
- sysctls[i].Value = cs[1]
- }
- return sysctls, nil
-}
-
-// PodAnnotationsFromSysctls creates an annotation value for a slice of Sysctls.
-func PodAnnotationsFromSysctls(sysctls []core.Sysctl) string {
- if len(sysctls) == 0 {
- return ""
- }
-
- kvs := make([]string, len(sysctls))
- for i := range sysctls {
- kvs[i] = fmt.Sprintf("%s=%s", sysctls[i].Name, sysctls[i].Value)
- }
- return strings.Join(kvs, ",")
-}
-
-// GetPersistentVolumeClass returns StorageClassName.
-func GetPersistentVolumeClass(volume *core.PersistentVolume) string {
- // Use beta annotation first
- if class, found := volume.Annotations[core.BetaStorageClassAnnotation]; found {
- return class
- }
-
- return volume.Spec.StorageClassName
-}
-
-// GetPersistentVolumeClaimClass returns StorageClassName. If no storage class was
-// requested, it returns "".
-func GetPersistentVolumeClaimClass(claim *core.PersistentVolumeClaim) string {
- // Use beta annotation first
- if class, found := claim.Annotations[core.BetaStorageClassAnnotation]; found {
- return class
- }
-
- if claim.Spec.StorageClassName != nil {
- return *claim.Spec.StorageClassName
- }
-
- return ""
-}
-
-// PersistentVolumeClaimHasClass returns true if given claim has set StorageClassName field.
-func PersistentVolumeClaimHasClass(claim *core.PersistentVolumeClaim) bool {
- // Use beta annotation first
- if _, found := claim.Annotations[core.BetaStorageClassAnnotation]; found {
- return true
- }
-
- if claim.Spec.StorageClassName != nil {
- return true
- }
-
- return false
-}
-
-// GetStorageNodeAffinityFromAnnotation gets the json serialized data from PersistentVolume.Annotations
-// and converts it to the NodeAffinity type in core.
-// TODO: update when storage node affinity graduates to beta
-func GetStorageNodeAffinityFromAnnotation(annotations map[string]string) (*core.NodeAffinity, error) {
- if len(annotations) > 0 && annotations[core.AlphaStorageNodeAffinityAnnotation] != "" {
- var affinity core.NodeAffinity
- err := json.Unmarshal([]byte(annotations[core.AlphaStorageNodeAffinityAnnotation]), &affinity)
- if err != nil {
- return nil, err
- }
- return &affinity, nil
- }
- return nil, nil
-}
-
-// Converts NodeAffinity type to Alpha annotation for use in PersistentVolumes
-// TODO: update when storage node affinity graduates to beta
-func StorageNodeAffinityToAlphaAnnotation(annotations map[string]string, affinity *core.NodeAffinity) error {
- if affinity == nil {
- return nil
- }
-
- json, err := json.Marshal(*affinity)
- if err != nil {
- return err
- }
- annotations[core.AlphaStorageNodeAffinityAnnotation] = string(json)
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/install/install.go b/vendor/k8s.io/kubernetes/pkg/apis/core/install/install.go
deleted file mode 100644
index cae514ec7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/install/install.go
+++ /dev/null
@@ -1,67 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package install installs the v1 monolithic api, making it available as an
-// option to all of the API encoding/decoding machinery.
-package install
-
-import (
- "k8s.io/apimachinery/pkg/apimachinery/announced"
- "k8s.io/apimachinery/pkg/apimachinery/registered"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/kubernetes/pkg/api/legacyscheme"
- "k8s.io/kubernetes/pkg/apis/core"
- "k8s.io/kubernetes/pkg/apis/core/v1"
-)
-
-func init() {
- Install(legacyscheme.GroupFactoryRegistry, legacyscheme.Registry, legacyscheme.Scheme)
-}
-
-// Install registers the API group and adds types to a scheme
-func Install(groupFactoryRegistry announced.APIGroupFactoryRegistry, registry *registered.APIRegistrationManager, scheme *runtime.Scheme) {
- if err := announced.NewGroupMetaFactory(
- &announced.GroupMetaFactoryArgs{
- GroupName: core.GroupName,
- VersionPreferenceOrder: []string{v1.SchemeGroupVersion.Version},
- AddInternalObjectsToScheme: core.AddToScheme,
- RootScopedKinds: sets.NewString(
- "Node",
- "Namespace",
- "PersistentVolume",
- "ComponentStatus",
- ),
- IgnoredKinds: sets.NewString(
- "ListOptions",
- "DeleteOptions",
- "Status",
- "PodLogOptions",
- "PodExecOptions",
- "PodAttachOptions",
- "PodPortForwardOptions",
- "PodProxyOptions",
- "NodeProxyOptions",
- "ServiceProxyOptions",
- ),
- },
- announced.VersionToSchemeFunc{
- v1.SchemeGroupVersion.Version: v1.AddToScheme,
- },
- ).Announce(groupFactoryRegistry).RegisterAndEnable(registry, scheme); err != nil {
- panic(err)
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/json.go b/vendor/k8s.io/kubernetes/pkg/apis/core/json.go
deleted file mode 100644
index 937cd056c..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/json.go
+++ /dev/null
@@ -1,28 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package core
-
-import "encoding/json"
-
-// This file implements json marshaling/unmarshaling interfaces on objects that are currently marshaled into annotations
-// to prevent anyone from marshaling these internal structs.
-
-var _ = json.Marshaler(&AvoidPods{})
-var _ = json.Unmarshaler(&AvoidPods{})
-
-func (AvoidPods) MarshalJSON() ([]byte, error) { panic("do not marshal internal struct") }
-func (*AvoidPods) UnmarshalJSON([]byte) error { panic("do not unmarshal to internal struct") }
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/objectreference.go b/vendor/k8s.io/kubernetes/pkg/apis/core/objectreference.go
deleted file mode 100644
index 55b27f30b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/objectreference.go
+++ /dev/null
@@ -1,34 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-//TODO: consider making these methods functions, because we don't want helper
-//functions in the k8s.io/api repo.
-
-package core
-
-import (
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-func (obj *ObjectReference) SetGroupVersionKind(gvk schema.GroupVersionKind) {
- obj.APIVersion, obj.Kind = gvk.ToAPIVersionAndKind()
-}
-
-func (obj *ObjectReference) GroupVersionKind() schema.GroupVersionKind {
- return schema.FromAPIVersionAndKind(obj.APIVersion, obj.Kind)
-}
-
-func (obj *ObjectReference) GetObjectKind() schema.ObjectKind { return obj }
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/pods/helpers.go b/vendor/k8s.io/kubernetes/pkg/apis/core/pods/helpers.go
deleted file mode 100644
index cf199cee7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/pods/helpers.go
+++ /dev/null
@@ -1,63 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package pods
-
-import (
- "fmt"
-
- "k8s.io/kubernetes/pkg/fieldpath"
-)
-
-// ConvertDownwardAPIFieldLabel converts the specified downward API field label
-// and its value in the pod of the specified version to the internal version,
-// and returns the converted label and value. This function returns an error if
-// the conversion fails.
-func ConvertDownwardAPIFieldLabel(version, label, value string) (string, string, error) {
- if version != "v1" {
- return "", "", fmt.Errorf("unsupported pod version: %s", version)
- }
-
- if path, _, ok := fieldpath.SplitMaybeSubscriptedPath(label); ok {
- switch path {
- case "metadata.annotations", "metadata.labels":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label does not support subscript: %s", label)
- }
- }
-
- switch label {
- case "metadata.annotations",
- "metadata.labels",
- "metadata.name",
- "metadata.namespace",
- "metadata.uid",
- "spec.nodeName",
- "spec.restartPolicy",
- "spec.serviceAccountName",
- "spec.schedulerName",
- "status.phase",
- "status.hostIP",
- "status.podIP":
- return label, value, nil
- // This is for backwards compatibility with old v1 clients which send spec.host
- case "spec.host":
- return "spec.nodeName", value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/register.go b/vendor/k8s.io/kubernetes/pkg/apis/core/register.go
deleted file mode 100644
index 2784cbe15..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/register.go
+++ /dev/null
@@ -1,99 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package core
-
-import (
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-// GroupName is the group name use in this package
-const GroupName = ""
-
-// SchemeGroupVersion is group version used to register these objects
-var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: runtime.APIVersionInternal}
-
-// Kind takes an unqualified kind and returns a Group qualified GroupKind
-func Kind(kind string) schema.GroupKind {
- return SchemeGroupVersion.WithKind(kind).GroupKind()
-}
-
-// Resource takes an unqualified resource and returns a Group qualified GroupResource
-func Resource(resource string) schema.GroupResource {
- return SchemeGroupVersion.WithResource(resource).GroupResource()
-}
-
-var (
- SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes)
- AddToScheme = SchemeBuilder.AddToScheme
-)
-
-func addKnownTypes(scheme *runtime.Scheme) error {
- if err := scheme.AddIgnoredConversionType(&metav1.TypeMeta{}, &metav1.TypeMeta{}); err != nil {
- return err
- }
- scheme.AddKnownTypes(SchemeGroupVersion,
- &Pod{},
- &PodList{},
- &PodStatusResult{},
- &PodTemplate{},
- &PodTemplateList{},
- &ReplicationControllerList{},
- &ReplicationController{},
- &ServiceList{},
- &Service{},
- &ServiceProxyOptions{},
- &NodeList{},
- &Node{},
- &NodeConfigSource{},
- &NodeProxyOptions{},
- &Endpoints{},
- &EndpointsList{},
- &Binding{},
- &Event{},
- &EventList{},
- &List{},
- &LimitRange{},
- &LimitRangeList{},
- &ResourceQuota{},
- &ResourceQuotaList{},
- &Namespace{},
- &NamespaceList{},
- &ServiceAccount{},
- &ServiceAccountList{},
- &Secret{},
- &SecretList{},
- &PersistentVolume{},
- &PersistentVolumeList{},
- &PersistentVolumeClaim{},
- &PersistentVolumeClaimList{},
- &PodAttachOptions{},
- &PodLogOptions{},
- &PodExecOptions{},
- &PodPortForwardOptions{},
- &PodProxyOptions{},
- &ComponentStatus{},
- &ComponentStatusList{},
- &SerializedReference{},
- &RangeAllocation{},
- &ConfigMap{},
- &ConfigMapList{},
- )
-
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/resource.go b/vendor/k8s.io/kubernetes/pkg/apis/core/resource.go
deleted file mode 100644
index 1910cd921..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/resource.go
+++ /dev/null
@@ -1,62 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package core
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
-)
-
-func (self ResourceName) String() string {
- return string(self)
-}
-
-// Returns the CPU limit if specified.
-func (self *ResourceList) Cpu() *resource.Quantity {
- if val, ok := (*self)[ResourceCPU]; ok {
- return &val
- }
- return &resource.Quantity{Format: resource.DecimalSI}
-}
-
-// Returns the Memory limit if specified.
-func (self *ResourceList) Memory() *resource.Quantity {
- if val, ok := (*self)[ResourceMemory]; ok {
- return &val
- }
- return &resource.Quantity{Format: resource.BinarySI}
-}
-
-func (self *ResourceList) Pods() *resource.Quantity {
- if val, ok := (*self)[ResourcePods]; ok {
- return &val
- }
- return &resource.Quantity{}
-}
-
-func (self *ResourceList) NvidiaGPU() *resource.Quantity {
- if val, ok := (*self)[ResourceNvidiaGPU]; ok {
- return &val
- }
- return &resource.Quantity{}
-}
-
-func (self *ResourceList) StorageEphemeral() *resource.Quantity {
- if val, ok := (*self)[ResourceEphemeralStorage]; ok {
- return &val
- }
- return &resource.Quantity{}
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/taint.go b/vendor/k8s.io/kubernetes/pkg/apis/core/taint.go
deleted file mode 100644
index ae1feb74d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/taint.go
+++ /dev/null
@@ -1,36 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-//TODO: consider making these methods functions, because we don't want helper
-//functions in the k8s.io/api repo.
-
-package core
-
-import "fmt"
-
-// MatchTaint checks if the taint matches taintToMatch. Taints are unique by key:effect,
-// if the two taints have same key:effect, regard as they match.
-func (t *Taint) MatchTaint(taintToMatch Taint) bool {
- return t.Key == taintToMatch.Key && t.Effect == taintToMatch.Effect
-}
-
-// taint.ToString() converts taint struct to string in format key=value:effect or key:effect.
-func (t *Taint) ToString() string {
- if len(t.Value) == 0 {
- return fmt.Sprintf("%v:%v", t.Key, t.Effect)
- }
- return fmt.Sprintf("%v=%v:%v", t.Key, t.Value, t.Effect)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/toleration.go b/vendor/k8s.io/kubernetes/pkg/apis/core/toleration.go
deleted file mode 100644
index 1dfbc9f1b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/toleration.go
+++ /dev/null
@@ -1,30 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-//TODO: consider making these methods functions, because we don't want helper
-//functions in the k8s.io/api repo.
-
-package core
-
-// MatchToleration checks if the toleration matches tolerationToMatch. Tolerations are unique by <key,effect,operator,value>,
-// if the two tolerations have same <key,effect,operator,value> combination, regard as they match.
-// TODO: uniqueness check for tolerations in api validations.
-func (t *Toleration) MatchToleration(tolerationToMatch *Toleration) bool {
- return t.Key == tolerationToMatch.Key &&
- t.Effect == tolerationToMatch.Effect &&
- t.Operator == tolerationToMatch.Operator &&
- t.Value == tolerationToMatch.Value
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/types.go b/vendor/k8s.io/kubernetes/pkg/apis/core/types.go
deleted file mode 100644
index a112791b1..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/types.go
+++ /dev/null
@@ -1,4656 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package core
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
- metainternalversion "k8s.io/apimachinery/pkg/apis/meta/internalversion"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/fields"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/types"
- "k8s.io/apimachinery/pkg/util/intstr"
-)
-
-// Common string formats
-// ---------------------
-// Many fields in this API have formatting requirements. The commonly used
-// formats are defined here.
-//
-// C_IDENTIFIER: This is a string that conforms to the definition of an "identifier"
-// in the C language. This is captured by the following regex:
-// [A-Za-z_][A-Za-z0-9_]*
-// This defines the format, but not the length restriction, which should be
-// specified at the definition of any field of this type.
-//
-// DNS_LABEL: This is a string, no more than 63 characters long, that conforms
-// to the definition of a "label" in RFCs 1035 and 1123. This is captured
-// by the following regex:
-// [a-z0-9]([-a-z0-9]*[a-z0-9])?
-//
-// DNS_SUBDOMAIN: This is a string, no more than 253 characters long, that conforms
-// to the definition of a "subdomain" in RFCs 1035 and 1123. This is captured
-// by the following regex:
-// [a-z0-9]([-a-z0-9]*[a-z0-9])?(\.[a-z0-9]([-a-z0-9]*[a-z0-9])?)*
-// or more simply:
-// DNS_LABEL(\.DNS_LABEL)*
-//
-// IANA_SVC_NAME: This is a string, no more than 15 characters long, that
-// conforms to the definition of IANA service name in RFC 6335.
-// It must contains at least one letter [a-z] and it must contains only [a-z0-9-].
-// Hypens ('-') cannot be leading or trailing character of the string
-// and cannot be adjacent to other hyphens.
-
-// ObjectMeta is metadata that all persisted resources must have, which includes all objects
-// users must create.
-// DEPRECATED: Use k8s.io/apimachinery/pkg/apis/meta/v1.ObjectMeta instead - this type will be removed soon.
-type ObjectMeta struct {
- // Name is unique within a namespace. Name is required when creating resources, although
- // some resources may allow a client to request the generation of an appropriate name
- // automatically. Name is primarily intended for creation idempotence and configuration
- // definition.
- // +optional
- Name string
-
- // GenerateName indicates that the name should be made unique by the server prior to persisting
- // it. A non-empty value for the field indicates the name will be made unique (and the name
- // returned to the client will be different than the name passed). The value of this field will
- // be combined with a unique suffix on the server if the Name field has not been provided.
- // The provided value must be valid within the rules for Name, and may be truncated by the length
- // of the suffix required to make the value unique on the server.
- //
- // If this field is specified, and Name is not present, the server will NOT return a 409 if the
- // generated name exists - instead, it will either return 201 Created or 500 with Reason
- // ServerTimeout indicating a unique name could not be found in the time allotted, and the client
- // should retry (optionally after the time indicated in the Retry-After header).
- // +optional
- GenerateName string
-
- // Namespace defines the space within which name must be unique. An empty namespace is
- // equivalent to the "default" namespace, but "default" is the canonical representation.
- // Not all objects are required to be scoped to a namespace - the value of this field for
- // those objects will be empty.
- // +optional
- Namespace string
-
- // SelfLink is a URL representing this object.
- // +optional
- SelfLink string
-
- // UID is the unique in time and space value for this object. It is typically generated by
- // the server on successful creation of a resource and is not allowed to change on PUT
- // operations.
- // +optional
- UID types.UID
-
- // An opaque value that represents the version of this resource. May be used for optimistic
- // concurrency, change detection, and the watch operation on a resource or set of resources.
- // Clients must treat these values as opaque and values may only be valid for a particular
- // resource or set of resources. Only servers will generate resource versions.
- // +optional
- ResourceVersion string
-
- // A sequence number representing a specific generation of the desired state.
- // Populated by the system. Read-only.
- // +optional
- Generation int64
-
- // CreationTimestamp is a timestamp representing the server time when this object was
- // created. It is not guaranteed to be set in happens-before order across separate operations.
- // Clients may not set this value. It is represented in RFC3339 form and is in UTC.
- // +optional
- CreationTimestamp metav1.Time
-
- // DeletionTimestamp is RFC 3339 date and time at which this resource will be deleted. This
- // field is set by the server when a graceful deletion is requested by the user, and is not
- // directly settable by a client. The resource is expected to be deleted (no longer visible
- // from resource lists, and not reachable by name) after the time in this field. Once set,
- // this value may not be unset or be set further into the future, although it may be shortened
- // or the resource may be deleted prior to this time. For example, a user may request that
- // a pod is deleted in 30 seconds. The Kubelet will react by sending a graceful termination
- // signal to the containers in the pod. After that 30 seconds, the Kubelet will send a hard
- // termination signal (SIGKILL) to the container and after cleanup, remove the pod from the
- // API. In the presence of network partitions, this object may still exist after this
- // timestamp, until an administrator or automated process can determine the resource is
- // fully terminated.
- // If not set, graceful deletion of the object has not been requested.
- //
- // Populated by the system when a graceful deletion is requested.
- // Read-only.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- DeletionTimestamp *metav1.Time
-
- // DeletionGracePeriodSeconds records the graceful deletion value set when graceful deletion
- // was requested. Represents the most recent grace period, and may only be shortened once set.
- // +optional
- DeletionGracePeriodSeconds *int64
-
- // Labels are key value pairs that may be used to scope and select individual resources.
- // Label keys are of the form:
- // label-key ::= prefixed-name | name
- // prefixed-name ::= prefix '/' name
- // prefix ::= DNS_SUBDOMAIN
- // name ::= DNS_LABEL
- // The prefix is optional. If the prefix is not specified, the key is assumed to be private
- // to the user. Other system components that wish to use labels must specify a prefix. The
- // "kubernetes.io/" prefix is reserved for use by kubernetes components.
- // +optional
- Labels map[string]string
-
- // Annotations are unstructured key value data stored with a resource that may be set by
- // external tooling. They are not queryable and should be preserved when modifying
- // objects. Annotation keys have the same formatting restrictions as Label keys. See the
- // comments on Labels for details.
- // +optional
- Annotations map[string]string
-
- // List of objects depended by this object. If ALL objects in the list have
- // been deleted, this object will be garbage collected. If this object is managed by a controller,
- // then an entry in this list will point to this controller, with the controller field set to true.
- // There cannot be more than one managing controller.
- // +optional
- OwnerReferences []metav1.OwnerReference
-
- // An initializer is a controller which enforces some system invariant at object creation time.
- // This field is a list of initializers that have not yet acted on this object. If nil or empty,
- // this object has been completely initialized. Otherwise, the object is considered uninitialized
- // and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to
- // observe uninitialized objects.
- //
- // When an object is created, the system will populate this list with the current set of initializers.
- // Only privileged users may set or modify this list. Once it is empty, it may not be modified further
- // by any user.
- Initializers *metav1.Initializers
-
- // Must be empty before the object is deleted from the registry. Each entry
- // is an identifier for the responsible component that will remove the entry
- // from the list. If the deletionTimestamp of the object is non-nil, entries
- // in this list can only be removed.
- // +optional
- Finalizers []string
-
- // The name of the cluster which the object belongs to.
- // This is used to distinguish resources with same name and namespace in different clusters.
- // This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.
- // +optional
- ClusterName string
-}
-
-const (
- // NamespaceDefault means the object is in the default namespace which is applied when not specified by clients
- NamespaceDefault string = "default"
- // NamespaceAll is the default argument to specify on a context when you want to list or filter resources across all namespaces
- NamespaceAll string = ""
- // NamespaceNone is the argument for a context when there is no namespace.
- NamespaceNone string = ""
- // NamespaceSystem is the system namespace where we place system components.
- NamespaceSystem string = "kube-system"
- // NamespacePublic is the namespace where we place public info (ConfigMaps)
- NamespacePublic string = "kube-public"
- // TerminationMessagePathDefault means the default path to capture the application termination message running in a container
- TerminationMessagePathDefault string = "/dev/termination-log"
-)
-
-// Volume represents a named volume in a pod that may be accessed by any containers in the pod.
-type Volume struct {
- // Required: This must be a DNS_LABEL. Each volume in a pod must have
- // a unique name.
- Name string
- // The VolumeSource represents the location and type of a volume to mount.
- // This is optional for now. If not specified, the Volume is implied to be an EmptyDir.
- // This implied behavior is deprecated and will be removed in a future version.
- // +optional
- VolumeSource
-}
-
-// VolumeSource represents the source location of a volume to mount.
-// Only one of its members may be specified.
-type VolumeSource struct {
- // HostPath represents file or directory on the host machine that is
- // directly exposed to the container. This is generally used for system
- // agents or other privileged things that are allowed to see the host
- // machine. Most containers will NOT need this.
- // ---
- // TODO(jonesdl) We need to restrict who can use host directory mounts and who can/can not
- // mount host directories as read/write.
- // +optional
- HostPath *HostPathVolumeSource
- // EmptyDir represents a temporary directory that shares a pod's lifetime.
- // +optional
- EmptyDir *EmptyDirVolumeSource
- // GCEPersistentDisk represents a GCE Disk resource that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- GCEPersistentDisk *GCEPersistentDiskVolumeSource
- // AWSElasticBlockStore represents an AWS EBS disk that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- AWSElasticBlockStore *AWSElasticBlockStoreVolumeSource
- // GitRepo represents a git repository at a particular revision.
- // +optional
- GitRepo *GitRepoVolumeSource
- // Secret represents a secret that should populate this volume.
- // +optional
- Secret *SecretVolumeSource
- // NFS represents an NFS mount on the host that shares a pod's lifetime
- // +optional
- NFS *NFSVolumeSource
- // ISCSIVolumeSource represents an ISCSI Disk resource that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- ISCSI *ISCSIVolumeSource
- // Glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime
- // +optional
- Glusterfs *GlusterfsVolumeSource
- // PersistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace
- // +optional
- PersistentVolumeClaim *PersistentVolumeClaimVolumeSource
- // RBD represents a Rados Block Device mount on the host that shares a pod's lifetime
- // +optional
- RBD *RBDVolumeSource
-
- // Quobyte represents a Quobyte mount on the host that shares a pod's lifetime
- // +optional
- Quobyte *QuobyteVolumeSource
-
- // FlexVolume represents a generic volume resource that is
- // provisioned/attached using an exec based plugin.
- // +optional
- FlexVolume *FlexVolumeSource
-
- // Cinder represents a cinder volume attached and mounted on kubelets host machine
- // +optional
- Cinder *CinderVolumeSource
-
- // CephFS represents a Cephfs mount on the host that shares a pod's lifetime
- // +optional
- CephFS *CephFSVolumeSource
-
- // Flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running
- // +optional
- Flocker *FlockerVolumeSource
-
- // DownwardAPI represents metadata about the pod that should populate this volume
- // +optional
- DownwardAPI *DownwardAPIVolumeSource
- // FC represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod.
- // +optional
- FC *FCVolumeSource
- // AzureFile represents an Azure File Service mount on the host and bind mount to the pod.
- // +optional
- AzureFile *AzureFileVolumeSource
- // ConfigMap represents a configMap that should populate this volume
- // +optional
- ConfigMap *ConfigMapVolumeSource
- // VsphereVolume represents a vSphere volume attached and mounted on kubelets host machine
- // +optional
- VsphereVolume *VsphereVirtualDiskVolumeSource
- // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.
- // +optional
- AzureDisk *AzureDiskVolumeSource
- // PhotonPersistentDisk represents a Photon Controller persistent disk attached and mounted on kubelets host machine
- PhotonPersistentDisk *PhotonPersistentDiskVolumeSource
- // Items for all in one resources secrets, configmaps, and downward API
- Projected *ProjectedVolumeSource
- // PortworxVolume represents a portworx volume attached and mounted on kubelets host machine
- // +optional
- PortworxVolume *PortworxVolumeSource
- // ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes.
- // +optional
- ScaleIO *ScaleIOVolumeSource
- // StorageOS represents a StorageOS volume that is attached to the kubelet's host machine and mounted into the pod
- // +optional
- StorageOS *StorageOSVolumeSource
-}
-
-// Similar to VolumeSource but meant for the administrator who creates PVs.
-// Exactly one of its members must be set.
-type PersistentVolumeSource struct {
- // GCEPersistentDisk represents a GCE Disk resource that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- GCEPersistentDisk *GCEPersistentDiskVolumeSource
- // AWSElasticBlockStore represents an AWS EBS disk that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- AWSElasticBlockStore *AWSElasticBlockStoreVolumeSource
- // HostPath represents a directory on the host.
- // Provisioned by a developer or tester.
- // This is useful for single-node development and testing only!
- // On-host storage is not supported in any way and WILL NOT WORK in a multi-node cluster.
- // +optional
- HostPath *HostPathVolumeSource
- // Glusterfs represents a Glusterfs volume that is attached to a host and exposed to the pod
- // +optional
- Glusterfs *GlusterfsVolumeSource
- // NFS represents an NFS mount on the host that shares a pod's lifetime
- // +optional
- NFS *NFSVolumeSource
- // RBD represents a Rados Block Device mount on the host that shares a pod's lifetime
- // +optional
- RBD *RBDPersistentVolumeSource
- // Quobyte represents a Quobyte mount on the host that shares a pod's lifetime
- // +optional
- Quobyte *QuobyteVolumeSource
- // ISCSIPersistentVolumeSource represents an ISCSI resource that is attached to a
- // kubelet's host machine and then exposed to the pod.
- // +optional
- ISCSI *ISCSIPersistentVolumeSource
- // FlexVolume represents a generic volume resource that is
- // provisioned/attached using an exec based plugin.
- // +optional
- FlexVolume *FlexPersistentVolumeSource
- // Cinder represents a cinder volume attached and mounted on kubelets host machine
- // +optional
- Cinder *CinderVolumeSource
- // CephFS represents a Ceph FS mount on the host that shares a pod's lifetime
- // +optional
- CephFS *CephFSPersistentVolumeSource
- // FC represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod.
- // +optional
- FC *FCVolumeSource
- // Flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running
- // +optional
- Flocker *FlockerVolumeSource
- // AzureFile represents an Azure File Service mount on the host and bind mount to the pod.
- // +optional
- AzureFile *AzureFilePersistentVolumeSource
- // VsphereVolume represents a vSphere volume attached and mounted on kubelets host machine
- // +optional
- VsphereVolume *VsphereVirtualDiskVolumeSource
- // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.
- // +optional
- AzureDisk *AzureDiskVolumeSource
- // PhotonPersistentDisk represents a Photon Controller persistent disk attached and mounted on kubelets host machine
- PhotonPersistentDisk *PhotonPersistentDiskVolumeSource
- // PortworxVolume represents a portworx volume attached and mounted on kubelets host machine
- // +optional
- PortworxVolume *PortworxVolumeSource
- // ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes.
- // +optional
- ScaleIO *ScaleIOPersistentVolumeSource
- // Local represents directly-attached storage with node affinity
- // +optional
- Local *LocalVolumeSource
- // StorageOS represents a StorageOS volume that is attached to the kubelet's host machine and mounted into the pod
- // More info: https://releases.k8s.io/HEAD/examples/volumes/storageos/README.md
- // +optional
- StorageOS *StorageOSPersistentVolumeSource
- // CSI (Container Storage Interface) represents storage that handled by an external CSI driver (Beta feature).
- // +optional
- CSI *CSIPersistentVolumeSource
-}
-
-type PersistentVolumeClaimVolumeSource struct {
- // ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume
- ClaimName string
- // Optional: Defaults to false (read/write). ReadOnly here
- // will force the ReadOnly setting in VolumeMounts
- // +optional
- ReadOnly bool
-}
-
-const (
- // BetaStorageClassAnnotation represents the beta/previous StorageClass annotation.
- // It's deprecated and will be removed in a future release. (#51440)
- BetaStorageClassAnnotation = "volume.beta.kubernetes.io/storage-class"
-
- // MountOptionAnnotation defines mount option annotation used in PVs
- MountOptionAnnotation = "volume.beta.kubernetes.io/mount-options"
-
- // AlphaStorageNodeAffinityAnnotation defines node affinity policies for a PersistentVolume.
- // Value is a string of the json representation of type NodeAffinity
- AlphaStorageNodeAffinityAnnotation = "volume.alpha.kubernetes.io/node-affinity"
-)
-
-// +genclient
-// +genclient:nonNamespaced
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type PersistentVolume struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- //Spec defines a persistent volume owned by the cluster
- // +optional
- Spec PersistentVolumeSpec
-
- // Status represents the current information about persistent volume.
- // +optional
- Status PersistentVolumeStatus
-}
-
-type PersistentVolumeSpec struct {
- // Resources represents the actual resources of the volume
- Capacity ResourceList
- // Source represents the location and type of a volume to mount.
- PersistentVolumeSource
- // AccessModes contains all ways the volume can be mounted
- // +optional
- AccessModes []PersistentVolumeAccessMode
- // ClaimRef is part of a bi-directional binding between PersistentVolume and PersistentVolumeClaim.
- // ClaimRef is expected to be non-nil when bound.
- // claim.VolumeName is the authoritative bind between PV and PVC.
- // When set to non-nil value, PVC.Spec.Selector of the referenced PVC is
- // ignored, i.e. labels of this PV do not need to match PVC selector.
- // +optional
- ClaimRef *ObjectReference
- // Optional: what happens to a persistent volume when released from its claim.
- // +optional
- PersistentVolumeReclaimPolicy PersistentVolumeReclaimPolicy
- // Name of StorageClass to which this persistent volume belongs. Empty value
- // means that this volume does not belong to any StorageClass.
- // +optional
- StorageClassName string
- // A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will
- // simply fail if one is invalid.
- // +optional
- MountOptions []string
- // volumeMode defines if a volume is intended to be used with a formatted filesystem
- // or to remain in raw block state. Value of Filesystem is implied when not included in spec.
- // This is an alpha feature and may change in the future.
- // +optional
- VolumeMode *PersistentVolumeMode
- // NodeAffinity defines constraints that limit what nodes this volume can be accessed from.
- // This field influences the scheduling of pods that use this volume.
- // +optional
- NodeAffinity *VolumeNodeAffinity
-}
-
-// VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.
-type VolumeNodeAffinity struct {
- // Required specifies hard node constraints that must be met.
- Required *NodeSelector
-}
-
-// PersistentVolumeReclaimPolicy describes a policy for end-of-life maintenance of persistent volumes
-type PersistentVolumeReclaimPolicy string
-
-const (
- // PersistentVolumeReclaimRecycle means the volume will be recycled back into the pool of unbound persistent volumes on release from its claim.
- // The volume plugin must support Recycling.
- // DEPRECATED: The PersistentVolumeReclaimRecycle called Recycle is being deprecated. See announcement here: https://groups.google.com/forum/#!topic/kubernetes-dev/uexugCza84I
- PersistentVolumeReclaimRecycle PersistentVolumeReclaimPolicy = "Recycle"
- // PersistentVolumeReclaimDelete means the volume will be deleted from Kubernetes on release from its claim.
- // The volume plugin must support Deletion.
- PersistentVolumeReclaimDelete PersistentVolumeReclaimPolicy = "Delete"
- // PersistentVolumeReclaimRetain means the volume will be left in its current phase (Released) for manual reclamation by the administrator.
- // The default policy is Retain.
- PersistentVolumeReclaimRetain PersistentVolumeReclaimPolicy = "Retain"
-)
-
-// PersistentVolumeMode describes how a volume is intended to be consumed, either Block or Filesystem.
-type PersistentVolumeMode string
-
-const (
- // PersistentVolumeBlock means the volume will not be formatted with a filesystem and will remain a raw block device.
- PersistentVolumeBlock PersistentVolumeMode = "Block"
- // PersistentVolumeFilesystem means the volume will be or is formatted with a filesystem.
- PersistentVolumeFilesystem PersistentVolumeMode = "Filesystem"
-)
-
-type PersistentVolumeStatus struct {
- // Phase indicates if a volume is available, bound to a claim, or released by a claim
- // +optional
- Phase PersistentVolumePhase
- // A human-readable message indicating details about why the volume is in this state.
- // +optional
- Message string
- // Reason is a brief CamelCase string that describes any failure and is meant for machine parsing and tidy display in the CLI
- // +optional
- Reason string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type PersistentVolumeList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
- Items []PersistentVolume
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PersistentVolumeClaim is a user's request for and claim to a persistent volume
-type PersistentVolumeClaim struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the volume requested by a pod author
- // +optional
- Spec PersistentVolumeClaimSpec
-
- // Status represents the current information about a claim
- // +optional
- Status PersistentVolumeClaimStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type PersistentVolumeClaimList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
- Items []PersistentVolumeClaim
-}
-
-// PersistentVolumeClaimSpec describes the common attributes of storage devices
-// and allows a Source for provider-specific attributes
-type PersistentVolumeClaimSpec struct {
- // Contains the types of access modes required
- // +optional
- AccessModes []PersistentVolumeAccessMode
- // A label query over volumes to consider for binding. This selector is
- // ignored when VolumeName is set
- // +optional
- Selector *metav1.LabelSelector
- // Resources represents the minimum resources required
- // +optional
- Resources ResourceRequirements
- // VolumeName is the binding reference to the PersistentVolume backing this
- // claim. When set to non-empty value Selector is not evaluated
- // +optional
- VolumeName string
- // Name of the StorageClass required by the claim.
- // More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#class-1
- // +optional
- StorageClassName *string
- // volumeMode defines what type of volume is required by the claim.
- // Value of Filesystem is implied when not included in claim spec.
- // This is an alpha feature and may change in the future.
- // +optional
- VolumeMode *PersistentVolumeMode
-}
-
-type PersistentVolumeClaimConditionType string
-
-// These are valid conditions of Pvc
-const (
- // An user trigger resize of pvc has been started
- PersistentVolumeClaimResizing PersistentVolumeClaimConditionType = "Resizing"
- // PersistentVolumeClaimFileSystemResizePending - controller resize is finished and a file system resize is pending on node
- PersistentVolumeClaimFileSystemResizePending PersistentVolumeClaimConditionType = "FileSystemResizePending"
-)
-
-type PersistentVolumeClaimCondition struct {
- Type PersistentVolumeClaimConditionType
- Status ConditionStatus
- // +optional
- LastProbeTime metav1.Time
- // +optional
- LastTransitionTime metav1.Time
- // +optional
- Reason string
- // +optional
- Message string
-}
-
-type PersistentVolumeClaimStatus struct {
- // Phase represents the current phase of PersistentVolumeClaim
- // +optional
- Phase PersistentVolumeClaimPhase
- // AccessModes contains all ways the volume backing the PVC can be mounted
- // +optional
- AccessModes []PersistentVolumeAccessMode
- // Represents the actual resources of the underlying volume
- // +optional
- Capacity ResourceList
- // +optional
- Conditions []PersistentVolumeClaimCondition
-}
-
-type PersistentVolumeAccessMode string
-
-const (
- // can be mounted read/write mode to exactly 1 host
- ReadWriteOnce PersistentVolumeAccessMode = "ReadWriteOnce"
- // can be mounted in read-only mode to many hosts
- ReadOnlyMany PersistentVolumeAccessMode = "ReadOnlyMany"
- // can be mounted in read/write mode to many hosts
- ReadWriteMany PersistentVolumeAccessMode = "ReadWriteMany"
-)
-
-type PersistentVolumePhase string
-
-const (
- // used for PersistentVolumes that are not available
- VolumePending PersistentVolumePhase = "Pending"
- // used for PersistentVolumes that are not yet bound
- // Available volumes are held by the binder and matched to PersistentVolumeClaims
- VolumeAvailable PersistentVolumePhase = "Available"
- // used for PersistentVolumes that are bound
- VolumeBound PersistentVolumePhase = "Bound"
- // used for PersistentVolumes where the bound PersistentVolumeClaim was deleted
- // released volumes must be recycled before becoming available again
- // this phase is used by the persistent volume claim binder to signal to another process to reclaim the resource
- VolumeReleased PersistentVolumePhase = "Released"
- // used for PersistentVolumes that failed to be correctly recycled or deleted after being released from a claim
- VolumeFailed PersistentVolumePhase = "Failed"
-)
-
-type PersistentVolumeClaimPhase string
-
-const (
- // used for PersistentVolumeClaims that are not yet bound
- ClaimPending PersistentVolumeClaimPhase = "Pending"
- // used for PersistentVolumeClaims that are bound
- ClaimBound PersistentVolumeClaimPhase = "Bound"
- // used for PersistentVolumeClaims that lost their underlying
- // PersistentVolume. The claim was bound to a PersistentVolume and this
- // volume does not exist any longer and all data on it was lost.
- ClaimLost PersistentVolumeClaimPhase = "Lost"
-)
-
-type HostPathType string
-
-const (
- // For backwards compatible, leave it empty if unset
- HostPathUnset HostPathType = ""
- // If nothing exists at the given path, an empty directory will be created there
- // as needed with file mode 0755, having the same group and ownership with Kubelet.
- HostPathDirectoryOrCreate HostPathType = "DirectoryOrCreate"
- // A directory must exist at the given path
- HostPathDirectory HostPathType = "Directory"
- // If nothing exists at the given path, an empty file will be created there
- // as needed with file mode 0644, having the same group and ownership with Kubelet.
- HostPathFileOrCreate HostPathType = "FileOrCreate"
- // A file must exist at the given path
- HostPathFile HostPathType = "File"
- // A UNIX socket must exist at the given path
- HostPathSocket HostPathType = "Socket"
- // A character device must exist at the given path
- HostPathCharDev HostPathType = "CharDevice"
- // A block device must exist at the given path
- HostPathBlockDev HostPathType = "BlockDevice"
-)
-
-// Represents a host path mapped into a pod.
-// Host path volumes do not support ownership management or SELinux relabeling.
-type HostPathVolumeSource struct {
- // If the path is a symlink, it will follow the link to the real path.
- Path string
- // Defaults to ""
- Type *HostPathType
-}
-
-// Represents an empty directory for a pod.
-// Empty directory volumes support ownership management and SELinux relabeling.
-type EmptyDirVolumeSource struct {
- // TODO: Longer term we want to represent the selection of underlying
- // media more like a scheduling problem - user says what traits they
- // need, we give them a backing store that satisfies that. For now
- // this will cover the most common needs.
- // Optional: what type of storage medium should back this directory.
- // The default is "" which means to use the node's default medium.
- // +optional
- Medium StorageMedium
- // Total amount of local storage required for this EmptyDir volume.
- // The size limit is also applicable for memory medium.
- // The maximum usage on memory medium EmptyDir would be the minimum value between
- // the SizeLimit specified here and the sum of memory limits of all containers in a pod.
- // The default is nil which means that the limit is undefined.
- // More info: http://kubernetes.io/docs/user-guide/volumes#emptydir
- // +optional
- SizeLimit *resource.Quantity
-}
-
-// StorageMedium defines ways that storage can be allocated to a volume.
-type StorageMedium string
-
-const (
- StorageMediumDefault StorageMedium = "" // use whatever the default is for the node
- StorageMediumMemory StorageMedium = "Memory" // use memory (tmpfs)
- StorageMediumHugePages StorageMedium = "HugePages" // use hugepages
-)
-
-// Protocol defines network protocols supported for things like container ports.
-type Protocol string
-
-const (
- // ProtocolTCP is the TCP protocol.
- ProtocolTCP Protocol = "TCP"
- // ProtocolUDP is the UDP protocol.
- ProtocolUDP Protocol = "UDP"
-)
-
-// Represents a Persistent Disk resource in Google Compute Engine.
-//
-// A GCE PD must exist before mounting to a container. The disk must
-// also be in the same GCE project and zone as the kubelet. A GCE PD
-// can only be mounted as read/write once or read-only many times. GCE
-// PDs support ownership management and SELinux relabeling.
-type GCEPersistentDiskVolumeSource struct {
- // Unique name of the PD resource. Used to identify the disk in GCE
- PDName string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: Partition on the disk to mount.
- // If omitted, kubelet will attempt to mount the device name.
- // Ex. For /dev/sda1, this field is "1", for /dev/sda, this field is 0 or empty.
- // +optional
- Partition int32
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents an ISCSI disk.
-// ISCSI volumes can only be mounted as read/write once.
-// ISCSI volumes support ownership management and SELinux relabeling.
-type ISCSIVolumeSource struct {
- // Required: iSCSI target portal
- // the portal is either an IP or ip_addr:port if port is other than default (typically TCP ports 860 and 3260)
- // +optional
- TargetPortal string
- // Required: target iSCSI Qualified Name
- // +optional
- IQN string
- // Required: iSCSI target lun number
- // +optional
- Lun int32
- // Optional: Defaults to 'default' (tcp). iSCSI interface name that uses an iSCSI transport.
- // +optional
- ISCSIInterface string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // Optional: list of iSCSI target portal ips for high availability.
- // the portal is either an IP or ip_addr:port if port is other than default (typically TCP ports 860 and 3260)
- // +optional
- Portals []string
- // Optional: whether support iSCSI Discovery CHAP authentication
- // +optional
- DiscoveryCHAPAuth bool
- // Optional: whether support iSCSI Session CHAP authentication
- // +optional
- SessionCHAPAuth bool
- // Optional: CHAP secret for iSCSI target and initiator authentication.
- // The secret is used if either DiscoveryCHAPAuth or SessionCHAPAuth is true
- // +optional
- SecretRef *LocalObjectReference
- // Optional: Custom initiator name per volume.
- // If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface
- // <target portal>:<volume name> will be created for the connection.
- // +optional
- InitiatorName *string
-}
-
-// ISCSIPersistentVolumeSource represents an ISCSI disk.
-// ISCSI volumes can only be mounted as read/write once.
-// ISCSI volumes support ownership management and SELinux relabeling.
-type ISCSIPersistentVolumeSource struct {
- // Required: iSCSI target portal
- // the portal is either an IP or ip_addr:port if port is other than default (typically TCP ports 860 and 3260)
- // +optional
- TargetPortal string
- // Required: target iSCSI Qualified Name
- // +optional
- IQN string
- // Required: iSCSI target lun number
- // +optional
- Lun int32
- // Optional: Defaults to 'default' (tcp). iSCSI interface name that uses an iSCSI transport.
- // +optional
- ISCSIInterface string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // Optional: list of iSCSI target portal ips for high availability.
- // the portal is either an IP or ip_addr:port if port is other than default (typically TCP ports 860 and 3260)
- // +optional
- Portals []string
- // Optional: whether support iSCSI Discovery CHAP authentication
- // +optional
- DiscoveryCHAPAuth bool
- // Optional: whether support iSCSI Session CHAP authentication
- // +optional
- SessionCHAPAuth bool
- // Optional: CHAP secret for iSCSI target and initiator authentication.
- // The secret is used if either DiscoveryCHAPAuth or SessionCHAPAuth is true
- // +optional
- SecretRef *SecretReference
- // Optional: Custom initiator name per volume.
- // If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface
- // <target portal>:<volume name> will be created for the connection.
- // +optional
- InitiatorName *string
-}
-
-// Represents a Fibre Channel volume.
-// Fibre Channel volumes can only be mounted as read/write once.
-// Fibre Channel volumes support ownership management and SELinux relabeling.
-type FCVolumeSource struct {
- // Optional: FC target worldwide names (WWNs)
- // +optional
- TargetWWNs []string
- // Optional: FC target lun number
- // +optional
- Lun *int32
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // Optional: FC volume World Wide Identifiers (WWIDs)
- // Either WWIDs or TargetWWNs and Lun must be set, but not both simultaneously.
- // +optional
- WWIDs []string
-}
-
-// FlexPersistentVolumeSource represents a generic persistent volume resource that is
-// provisioned/attached using an exec based plugin.
-type FlexPersistentVolumeSource struct {
- // Driver is the name of the driver to use for this volume.
- Driver string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.
- // +optional
- FSType string
- // Optional: SecretRef is reference to the secret object containing
- // sensitive information to pass to the plugin scripts. This may be
- // empty if no secret object is specified. If the secret object
- // contains more than one secret, all secrets are passed to the plugin
- // scripts.
- // +optional
- SecretRef *SecretReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // Optional: Extra driver options if any.
- // +optional
- Options map[string]string
-}
-
-// FlexVolume represents a generic volume resource that is
-// provisioned/attached using an exec based plugin.
-type FlexVolumeSource struct {
- // Driver is the name of the driver to use for this volume.
- Driver string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.
- // +optional
- FSType string
- // Optional: SecretRef is reference to the secret object containing
- // sensitive information to pass to the plugin scripts. This may be
- // empty if no secret object is specified. If the secret object
- // contains more than one secret, all secrets are passed to the plugin
- // scripts.
- // +optional
- SecretRef *LocalObjectReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // Optional: Extra driver options if any.
- // +optional
- Options map[string]string
-}
-
-// Represents a Persistent Disk resource in AWS.
-//
-// An AWS EBS disk must exist before mounting to a container. The disk
-// must also be in the same AWS zone as the kubelet. An AWS EBS disk
-// can only be mounted as read/write once. AWS EBS volumes support
-// ownership management and SELinux relabeling.
-type AWSElasticBlockStoreVolumeSource struct {
- // Unique id of the persistent disk resource. Used to identify the disk in AWS
- VolumeID string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: Partition on the disk to mount.
- // If omitted, kubelet will attempt to mount the device name.
- // Ex. For /dev/sda1, this field is "1", for /dev/sda, this field is 0 or empty.
- // +optional
- Partition int32
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a volume that is populated with the contents of a git repository.
-// Git repo volumes do not support ownership management.
-// Git repo volumes support SELinux relabeling.
-type GitRepoVolumeSource struct {
- // Repository URL
- Repository string
- // Commit hash, this is optional
- // +optional
- Revision string
- // Clone target, this is optional
- // Must not contain or start with '..'. If '.' is supplied, the volume directory will be the
- // git repository. Otherwise, if specified, the volume will contain the git repository in
- // the subdirectory with the given name.
- // +optional
- Directory string
- // TODO: Consider credentials here.
-}
-
-// Adapts a Secret into a volume.
-//
-// The contents of the target Secret's Data field will be presented in a volume
-// as files using the keys in the Data field as the file names.
-// Secret volumes support ownership management and SELinux relabeling.
-type SecretVolumeSource struct {
- // Name of the secret in the pod's namespace to use.
- // +optional
- SecretName string
- // If unspecified, each key-value pair in the Data field of the referenced
- // Secret will be projected into the volume as a file whose name is the
- // key and content is the value. If specified, the listed keys will be
- // projected into the specified paths, and unlisted keys will not be
- // present. If a key is specified which is not present in the Secret,
- // the volume setup will error unless it is marked optional. Paths must be
- // relative and may not contain the '..' path or start with '..'.
- // +optional
- Items []KeyToPath
- // Mode bits to use on created files by default. Must be a value between
- // 0 and 0777.
- // Directories within the path are not affected by this setting.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- DefaultMode *int32
- // Specify whether the Secret or its key must be defined
- // +optional
- Optional *bool
-}
-
-// Adapts a secret into a projected volume.
-//
-// The contents of the target Secret's Data field will be presented in a
-// projected volume as files using the keys in the Data field as the file names.
-// Note that this is identical to a secret volume source without the default
-// mode.
-type SecretProjection struct {
- LocalObjectReference
- // If unspecified, each key-value pair in the Data field of the referenced
- // Secret will be projected into the volume as a file whose name is the
- // key and content is the value. If specified, the listed keys will be
- // projected into the specified paths, and unlisted keys will not be
- // present. If a key is specified which is not present in the Secret,
- // the volume setup will error unless it is marked optional. Paths must be
- // relative and may not contain the '..' path or start with '..'.
- // +optional
- Items []KeyToPath
- // Specify whether the Secret or its key must be defined
- // +optional
- Optional *bool
-}
-
-// Represents an NFS mount that lasts the lifetime of a pod.
-// NFS volumes do not support ownership management or SELinux relabeling.
-type NFSVolumeSource struct {
- // Server is the hostname or IP address of the NFS server
- Server string
-
- // Path is the exported NFS share
- Path string
-
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the NFS export to be mounted with read-only permissions
- // +optional
- ReadOnly bool
-}
-
-// Represents a Quobyte mount that lasts the lifetime of a pod.
-// Quobyte volumes do not support ownership management or SELinux relabeling.
-type QuobyteVolumeSource struct {
- // Registry represents a single or multiple Quobyte Registry services
- // specified as a string as host:port pair (multiple entries are separated with commas)
- // which acts as the central registry for volumes
- Registry string
-
- // Volume is a string that references an already created Quobyte volume by name.
- Volume string
-
- // Defaults to false (read/write). ReadOnly here will force
- // the Quobyte to be mounted with read-only permissions
- // +optional
- ReadOnly bool
-
- // User to map volume access to
- // Defaults to the root user
- // +optional
- User string
-
- // Group to map volume access to
- // Default is no group
- // +optional
- Group string
-}
-
-// Represents a Glusterfs mount that lasts the lifetime of a pod.
-// Glusterfs volumes do not support ownership management or SELinux relabeling.
-type GlusterfsVolumeSource struct {
- // Required: EndpointsName is the endpoint name that details Glusterfs topology
- EndpointsName string
-
- // Required: Path is the Glusterfs volume path
- Path string
-
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the Glusterfs to be mounted with read-only permissions
- // +optional
- ReadOnly bool
-}
-
-// Represents a Rados Block Device mount that lasts the lifetime of a pod.
-// RBD volumes support ownership management and SELinux relabeling.
-type RBDVolumeSource struct {
- // Required: CephMonitors is a collection of Ceph monitors
- CephMonitors []string
- // Required: RBDImage is the rados image name
- RBDImage string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: RadosPool is the rados pool name,default is rbd
- // +optional
- RBDPool string
- // Optional: RBDUser is the rados user name, default is admin
- // +optional
- RadosUser string
- // Optional: Keyring is the path to key ring for RBDUser, default is /etc/ceph/keyring
- // +optional
- Keyring string
- // Optional: SecretRef is name of the authentication secret for RBDUser, default is nil.
- // +optional
- SecretRef *LocalObjectReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a Rados Block Device mount that lasts the lifetime of a pod.
-// RBD volumes support ownership management and SELinux relabeling.
-type RBDPersistentVolumeSource struct {
- // Required: CephMonitors is a collection of Ceph monitors
- CephMonitors []string
- // Required: RBDImage is the rados image name
- RBDImage string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // TODO: how do we prevent errors in the filesystem from compromising the machine
- // +optional
- FSType string
- // Optional: RadosPool is the rados pool name,default is rbd
- // +optional
- RBDPool string
- // Optional: RBDUser is the rados user name, default is admin
- // +optional
- RadosUser string
- // Optional: Keyring is the path to key ring for RBDUser, default is /etc/ceph/keyring
- // +optional
- Keyring string
- // Optional: SecretRef is reference to the authentication secret for User, default is empty.
- // +optional
- SecretRef *SecretReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a cinder volume resource in Openstack. A Cinder volume
-// must exist before mounting to a container. The volume must also be
-// in the same region as the kubelet. Cinder volumes support ownership
-// management and SELinux relabeling.
-type CinderVolumeSource struct {
- // Unique id of the volume used to identify the cinder volume
- VolumeID string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a Ceph Filesystem mount that lasts the lifetime of a pod
-// Cephfs volumes do not support ownership management or SELinux relabeling.
-type CephFSVolumeSource struct {
- // Required: Monitors is a collection of Ceph monitors
- Monitors []string
- // Optional: Used as the mounted root, rather than the full Ceph tree, default is /
- // +optional
- Path string
- // Optional: User is the rados user name, default is admin
- // +optional
- User string
- // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret
- // +optional
- SecretFile string
- // Optional: SecretRef is reference to the authentication secret for User, default is empty.
- // +optional
- SecretRef *LocalObjectReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// SecretReference represents a Secret Reference. It has enough information to retrieve secret
-// in any namespace
-type SecretReference struct {
- // Name is unique within a namespace to reference a secret resource.
- // +optional
- Name string
- // Namespace defines the space within which the secret name must be unique.
- // +optional
- Namespace string
-}
-
-// Represents a Ceph Filesystem mount that lasts the lifetime of a pod
-// Cephfs volumes do not support ownership management or SELinux relabeling.
-type CephFSPersistentVolumeSource struct {
- // Required: Monitors is a collection of Ceph monitors
- Monitors []string
- // Optional: Used as the mounted root, rather than the full Ceph tree, default is /
- // +optional
- Path string
- // Optional: User is the rados user name, default is admin
- // +optional
- User string
- // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret
- // +optional
- SecretFile string
- // Optional: SecretRef is reference to the authentication secret for User, default is empty.
- // +optional
- SecretRef *SecretReference
- // Optional: Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a Flocker volume mounted by the Flocker agent.
-// One and only one of datasetName and datasetUUID should be set.
-// Flocker volumes do not support ownership management or SELinux relabeling.
-type FlockerVolumeSource struct {
- // Name of the dataset stored as metadata -> name on the dataset for Flocker
- // should be considered as deprecated
- // +optional
- DatasetName string
- // UUID of the dataset. This is unique identifier of a Flocker dataset
- // +optional
- DatasetUUID string
-}
-
-// Represents a volume containing downward API info.
-// Downward API volumes support ownership management and SELinux relabeling.
-type DownwardAPIVolumeSource struct {
- // Items is a list of DownwardAPIVolume file
- // +optional
- Items []DownwardAPIVolumeFile
- // Mode bits to use on created files by default. Must be a value between
- // 0 and 0777.
- // Directories within the path are not affected by this setting.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- DefaultMode *int32
-}
-
-// Represents a single file containing information from the downward API
-type DownwardAPIVolumeFile struct {
- // Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'
- Path string
- // Required: Selects a field of the pod: only annotations, labels, name, namespace and uid are supported.
- // +optional
- FieldRef *ObjectFieldSelector
- // Selects a resource of the container: only resources limits and requests
- // (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported.
- // +optional
- ResourceFieldRef *ResourceFieldSelector
- // Optional: mode bits to use on this file, must be a value between 0
- // and 0777. If not specified, the volume defaultMode will be used.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- Mode *int32
-}
-
-// Represents downward API info for projecting into a projected volume.
-// Note that this is identical to a downwardAPI volume source without the default
-// mode.
-type DownwardAPIProjection struct {
- // Items is a list of DownwardAPIVolume file
- // +optional
- Items []DownwardAPIVolumeFile
-}
-
-// AzureFile represents an Azure File Service mount on the host and bind mount to the pod.
-type AzureFileVolumeSource struct {
- // the name of secret that contains Azure Storage Account Name and Key
- SecretName string
- // Share Name
- ShareName string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// AzureFile represents an Azure File Service mount on the host and bind mount to the pod.
-type AzureFilePersistentVolumeSource struct {
- // the name of secret that contains Azure Storage Account Name and Key
- SecretName string
- // Share Name
- ShareName string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // the namespace of the secret that contains Azure Storage Account Name and Key
- // default is the same as the Pod
- // +optional
- SecretNamespace *string
-}
-
-// Represents a vSphere volume resource.
-type VsphereVirtualDiskVolumeSource struct {
- // Path that identifies vSphere volume vmdk
- VolumePath string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Storage Policy Based Management (SPBM) profile name.
- // +optional
- StoragePolicyName string
- // Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.
- // +optional
- StoragePolicyID string
-}
-
-// Represents a Photon Controller persistent disk resource.
-type PhotonPersistentDiskVolumeSource struct {
- // ID that identifies Photon Controller persistent disk
- PdID string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- FSType string
-}
-
-// PortworxVolumeSource represents a Portworx volume resource.
-type PortworxVolumeSource struct {
- // VolumeID uniquely identifies a Portworx volume
- VolumeID string
- // FSType represents the filesystem type to mount
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-type AzureDataDiskCachingMode string
-type AzureDataDiskKind string
-
-const (
- AzureDataDiskCachingNone AzureDataDiskCachingMode = "None"
- AzureDataDiskCachingReadOnly AzureDataDiskCachingMode = "ReadOnly"
- AzureDataDiskCachingReadWrite AzureDataDiskCachingMode = "ReadWrite"
-
- AzureSharedBlobDisk AzureDataDiskKind = "Shared"
- AzureDedicatedBlobDisk AzureDataDiskKind = "Dedicated"
- AzureManagedDisk AzureDataDiskKind = "Managed"
-)
-
-// AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.
-type AzureDiskVolumeSource struct {
- // The Name of the data disk in the blob storage
- DiskName string
- // The URI of the data disk in the blob storage
- DataDiskURI string
- // Host Caching mode: None, Read Only, Read Write.
- // +optional
- CachingMode *AzureDataDiskCachingMode
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType *string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly *bool
- // Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared
- Kind *AzureDataDiskKind
-}
-
-// ScaleIOVolumeSource represents a persistent ScaleIO volume
-type ScaleIOVolumeSource struct {
- // The host address of the ScaleIO API Gateway.
- Gateway string
- // The name of the storage system as configured in ScaleIO.
- System string
- // SecretRef references to the secret for ScaleIO user and other
- // sensitive information. If this is not provided, Login operation will fail.
- SecretRef *LocalObjectReference
- // Flag to enable/disable SSL communication with Gateway, default false
- // +optional
- SSLEnabled bool
- // The name of the ScaleIO Protection Domain for the configured storage.
- // +optional
- ProtectionDomain string
- // The ScaleIO Storage Pool associated with the protection domain.
- // +optional
- StoragePool string
- // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned.
- // +optional
- StorageMode string
- // The name of a volume already created in the ScaleIO system
- // that is associated with this volume source.
- VolumeName string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume that can be defined
-// by a an admin via a storage class, for instance.
-type ScaleIOPersistentVolumeSource struct {
- // The host address of the ScaleIO API Gateway.
- Gateway string
- // The name of the storage system as configured in ScaleIO.
- System string
- // SecretRef references to the secret for ScaleIO user and other
- // sensitive information. If this is not provided, Login operation will fail.
- SecretRef *SecretReference
- // Flag to enable/disable SSL communication with Gateway, default false
- // +optional
- SSLEnabled bool
- // The name of the ScaleIO Protection Domain for the configured storage.
- // +optional
- ProtectionDomain string
- // The ScaleIO Storage Pool associated with the protection domain.
- // +optional
- StoragePool string
- // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned.
- // +optional
- StorageMode string
- // The name of a volume created in the ScaleIO system
- // that is associated with this volume source.
- VolumeName string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
-}
-
-// Represents a StorageOS persistent volume resource.
-type StorageOSVolumeSource struct {
- // VolumeName is the human-readable name of the StorageOS volume. Volume
- // names are only unique within a namespace.
- VolumeName string
- // VolumeNamespace specifies the scope of the volume within StorageOS. If no
- // namespace is specified then the Pod's namespace will be used. This allows the
- // Kubernetes name scoping to be mirrored within StorageOS for tighter integration.
- // Set VolumeName to any name to override the default behaviour.
- // Set to "default" if you are not using namespaces within StorageOS.
- // Namespaces that do not pre-exist within StorageOS will be created.
- // +optional
- VolumeNamespace string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // SecretRef specifies the secret to use for obtaining the StorageOS API
- // credentials. If not specified, default values will be attempted.
- // +optional
- SecretRef *LocalObjectReference
-}
-
-// Represents a StorageOS persistent volume resource.
-type StorageOSPersistentVolumeSource struct {
- // VolumeName is the human-readable name of the StorageOS volume. Volume
- // names are only unique within a namespace.
- VolumeName string
- // VolumeNamespace specifies the scope of the volume within StorageOS. If no
- // namespace is specified then the Pod's namespace will be used. This allows the
- // Kubernetes name scoping to be mirrored within StorageOS for tighter integration.
- // Set VolumeName to any name to override the default behaviour.
- // Set to "default" if you are not using namespaces within StorageOS.
- // Namespaces that do not pre-exist within StorageOS will be created.
- // +optional
- VolumeNamespace string
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
- // Defaults to false (read/write). ReadOnly here will force
- // the ReadOnly setting in VolumeMounts.
- // +optional
- ReadOnly bool
- // SecretRef specifies the secret to use for obtaining the StorageOS API
- // credentials. If not specified, default values will be attempted.
- // +optional
- SecretRef *ObjectReference
-}
-
-// Adapts a ConfigMap into a volume.
-//
-// The contents of the target ConfigMap's Data field will be presented in a
-// volume as files using the keys in the Data field as the file names, unless
-// the items element is populated with specific mappings of keys to paths.
-// ConfigMap volumes support ownership management and SELinux relabeling.
-type ConfigMapVolumeSource struct {
- LocalObjectReference
- // If unspecified, each key-value pair in the Data field of the referenced
- // ConfigMap will be projected into the volume as a file whose name is the
- // key and content is the value. If specified, the listed keys will be
- // projected into the specified paths, and unlisted keys will not be
- // present. If a key is specified which is not present in the ConfigMap,
- // the volume setup will error unless it is marked optional. Paths must be
- // relative and may not contain the '..' path or start with '..'.
- // +optional
- Items []KeyToPath
- // Mode bits to use on created files by default. Must be a value between
- // 0 and 0777.
- // Directories within the path are not affected by this setting.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- DefaultMode *int32
- // Specify whether the ConfigMap or it's keys must be defined
- // +optional
- Optional *bool
-}
-
-// Adapts a ConfigMap into a projected volume.
-//
-// The contents of the target ConfigMap's Data field will be presented in a
-// projected volume as files using the keys in the Data field as the file names,
-// unless the items element is populated with specific mappings of keys to paths.
-// Note that this is identical to a configmap volume source without the default
-// mode.
-type ConfigMapProjection struct {
- LocalObjectReference
- // If unspecified, each key-value pair in the Data field of the referenced
- // ConfigMap will be projected into the volume as a file whose name is the
- // key and content is the value. If specified, the listed keys will be
- // projected into the specified paths, and unlisted keys will not be
- // present. If a key is specified which is not present in the ConfigMap,
- // the volume setup will error unless it is marked optional. Paths must be
- // relative and may not contain the '..' path or start with '..'.
- // +optional
- Items []KeyToPath
- // Specify whether the ConfigMap or it's keys must be defined
- // +optional
- Optional *bool
-}
-
-// Represents a projected volume source
-type ProjectedVolumeSource struct {
- // list of volume projections
- Sources []VolumeProjection
- // Mode bits to use on created files by default. Must be a value between
- // 0 and 0777.
- // Directories within the path are not affected by this setting.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- DefaultMode *int32
-}
-
-// Projection that may be projected along with other supported volume types
-type VolumeProjection struct {
- // all types below are the supported types for projection into the same volume
-
- // information about the secret data to project
- Secret *SecretProjection
- // information about the downwardAPI data to project
- DownwardAPI *DownwardAPIProjection
- // information about the configMap data to project
- ConfigMap *ConfigMapProjection
-}
-
-// Maps a string key to a path within a volume.
-type KeyToPath struct {
- // The key to project.
- Key string
-
- // The relative path of the file to map the key to.
- // May not be an absolute path.
- // May not contain the path element '..'.
- // May not start with the string '..'.
- Path string
- // Optional: mode bits to use on this file, should be a value between 0
- // and 0777. If not specified, the volume defaultMode will be used.
- // This might be in conflict with other options that affect the file
- // mode, like fsGroup, and the result can be other mode bits set.
- // +optional
- Mode *int32
-}
-
-// Local represents directly-attached storage with node affinity
-type LocalVolumeSource struct {
- // The full path to the volume on the node
- // For alpha, this path must be a directory
- // Once block as a source is supported, then this path can point to a block device
- Path string
-}
-
-// Represents storage that is managed by an external CSI volume driver (Beta feature)
-type CSIPersistentVolumeSource struct {
- // Driver is the name of the driver to use for this volume.
- // Required.
- Driver string
-
- // VolumeHandle is the unique volume name returned by the CSI volume
- // plugin’s CreateVolume to refer to the volume on all subsequent calls.
- // Required.
- VolumeHandle string
-
- // Optional: The value to pass to ControllerPublishVolumeRequest.
- // Defaults to false (read/write).
- // +optional
- ReadOnly bool
-
- // Filesystem type to mount.
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.
- // +optional
- FSType string
-
- // Attributes of the volume to publish.
- // +optional
- VolumeAttributes map[string]string
-
- // ControllerPublishSecretRef is a reference to the secret object containing
- // sensitive information to pass to the CSI driver to complete the CSI
- // ControllerPublishVolume and ControllerUnpublishVolume calls.
- // This field is optional, and may be empty if no secret is required. If the
- // secret object contains more than one secret, all secrets are passed.
- // +optional
- ControllerPublishSecretRef *SecretReference
-
- // NodeStageSecretRef is a reference to the secret object containing sensitive
- // information to pass to the CSI driver to complete the CSI NodeStageVolume
- // and NodeStageVolume and NodeUnstageVolume calls.
- // This field is optional, and may be empty if no secret is required. If the
- // secret object contains more than one secret, all secrets are passed.
- // +optional
- NodeStageSecretRef *SecretReference
-
- // NodePublishSecretRef is a reference to the secret object containing
- // sensitive information to pass to the CSI driver to complete the CSI
- // NodePublishVolume and NodeUnpublishVolume calls.
- // This field is optional, and may be empty if no secret is required. If the
- // secret object contains more than one secret, all secrets are passed.
- // +optional
- NodePublishSecretRef *SecretReference
-}
-
-// ContainerPort represents a network port in a single container
-type ContainerPort struct {
- // Optional: If specified, this must be an IANA_SVC_NAME Each named port
- // in a pod must have a unique name.
- // +optional
- Name string
- // Optional: If specified, this must be a valid port number, 0 < x < 65536.
- // If HostNetwork is specified, this must match ContainerPort.
- // +optional
- HostPort int32
- // Required: This must be a valid port number, 0 < x < 65536.
- ContainerPort int32
- // Required: Supports "TCP" and "UDP".
- // +optional
- Protocol Protocol
- // Optional: What host IP to bind the external port to.
- // +optional
- HostIP string
-}
-
-// VolumeMount describes a mounting of a Volume within a container.
-type VolumeMount struct {
- // Required: This must match the Name of a Volume [above].
- Name string
- // Optional: Defaults to false (read-write).
- // +optional
- ReadOnly bool
- // Required. If the path is not an absolute path (e.g. some/path) it
- // will be prepended with the appropriate root prefix for the operating
- // system. On Linux this is '/', on Windows this is 'C:\'.
- MountPath string
- // Path within the volume from which the container's volume should be mounted.
- // Defaults to "" (volume's root).
- // +optional
- SubPath string
- // mountPropagation determines how mounts are propagated from the host
- // to container and the other way around.
- // When not set, MountPropagationHostToContainer is used.
- // This field is alpha in 1.8 and can be reworked or removed in a future
- // release.
- // +optional
- MountPropagation *MountPropagationMode
-}
-
-// MountPropagationMode describes mount propagation.
-type MountPropagationMode string
-
-const (
- // MountPropagationHostToContainer means that the volume in a container will
- // receive new mounts from the host or other containers, but filesystems
- // mounted inside the container won't be propagated to the host or other
- // containers.
- // Note that this mode is recursively applied to all mounts in the volume
- // ("rslave" in Linux terminology).
- MountPropagationHostToContainer MountPropagationMode = "HostToContainer"
- // MountPropagationBidirectional means that the volume in a container will
- // receive new mounts from the host or other containers, and its own mounts
- // will be propagated from the container to the host or other containers.
- // Note that this mode is recursively applied to all mounts in the volume
- // ("rshared" in Linux terminology).
- MountPropagationBidirectional MountPropagationMode = "Bidirectional"
-)
-
-// VolumeDevice describes a mapping of a raw block device within a container.
-type VolumeDevice struct {
- // name must match the name of a persistentVolumeClaim in the pod
- Name string
- // devicePath is the path inside of the container that the device will be mapped to.
- DevicePath string
-}
-
-// EnvVar represents an environment variable present in a Container.
-type EnvVar struct {
- // Required: This must be a C_IDENTIFIER.
- Name string
- // Optional: no more than one of the following may be specified.
- // Optional: Defaults to ""; variable references $(VAR_NAME) are expanded
- // using the previous defined environment variables in the container and
- // any service environment variables. If a variable cannot be resolved,
- // the reference in the input string will be unchanged. The $(VAR_NAME)
- // syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped
- // references will never be expanded, regardless of whether the variable
- // exists or not.
- // +optional
- Value string
- // Optional: Specifies a source the value of this var should come from.
- // +optional
- ValueFrom *EnvVarSource
-}
-
-// EnvVarSource represents a source for the value of an EnvVar.
-// Only one of its fields may be set.
-type EnvVarSource struct {
- // Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations,
- // metadata.uid, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP.
- // +optional
- FieldRef *ObjectFieldSelector
- // Selects a resource of the container: only resources limits and requests
- // (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported.
- // +optional
- ResourceFieldRef *ResourceFieldSelector
- // Selects a key of a ConfigMap.
- // +optional
- ConfigMapKeyRef *ConfigMapKeySelector
- // Selects a key of a secret in the pod's namespace.
- // +optional
- SecretKeyRef *SecretKeySelector
-}
-
-// ObjectFieldSelector selects an APIVersioned field of an object.
-type ObjectFieldSelector struct {
- // Required: Version of the schema the FieldPath is written in terms of.
- // If no value is specified, it will be defaulted to the APIVersion of the
- // enclosing object.
- APIVersion string
- // Required: Path of the field to select in the specified API version
- FieldPath string
-}
-
-// ResourceFieldSelector represents container resources (cpu, memory) and their output format
-type ResourceFieldSelector struct {
- // Container name: required for volumes, optional for env vars
- // +optional
- ContainerName string
- // Required: resource to select
- Resource string
- // Specifies the output format of the exposed resources, defaults to "1"
- // +optional
- Divisor resource.Quantity
-}
-
-// Selects a key from a ConfigMap.
-type ConfigMapKeySelector struct {
- // The ConfigMap to select from.
- LocalObjectReference
- // The key to select.
- Key string
- // Specify whether the ConfigMap or it's key must be defined
- // +optional
- Optional *bool
-}
-
-// SecretKeySelector selects a key of a Secret.
-type SecretKeySelector struct {
- // The name of the secret in the pod's namespace to select from.
- LocalObjectReference
- // The key of the secret to select from. Must be a valid secret key.
- Key string
- // Specify whether the Secret or it's key must be defined
- // +optional
- Optional *bool
-}
-
-// EnvFromSource represents the source of a set of ConfigMaps
-type EnvFromSource struct {
- // An optional identifier to prepend to each key in the ConfigMap.
- // +optional
- Prefix string
- // The ConfigMap to select from.
- //+optional
- ConfigMapRef *ConfigMapEnvSource
- // The Secret to select from.
- //+optional
- SecretRef *SecretEnvSource
-}
-
-// ConfigMapEnvSource selects a ConfigMap to populate the environment
-// variables with.
-//
-// The contents of the target ConfigMap's Data field will represent the
-// key-value pairs as environment variables.
-type ConfigMapEnvSource struct {
- // The ConfigMap to select from.
- LocalObjectReference
- // Specify whether the ConfigMap must be defined
- // +optional
- Optional *bool
-}
-
-// SecretEnvSource selects a Secret to populate the environment
-// variables with.
-//
-// The contents of the target Secret's Data field will represent the
-// key-value pairs as environment variables.
-type SecretEnvSource struct {
- // The Secret to select from.
- LocalObjectReference
- // Specify whether the Secret must be defined
- // +optional
- Optional *bool
-}
-
-// HTTPHeader describes a custom header to be used in HTTP probes
-type HTTPHeader struct {
- // The header field name
- Name string
- // The header field value
- Value string
-}
-
-// HTTPGetAction describes an action based on HTTP Get requests.
-type HTTPGetAction struct {
- // Optional: Path to access on the HTTP server.
- // +optional
- Path string
- // Required: Name or number of the port to access on the container.
- // +optional
- Port intstr.IntOrString
- // Optional: Host name to connect to, defaults to the pod IP. You
- // probably want to set "Host" in httpHeaders instead.
- // +optional
- Host string
- // Optional: Scheme to use for connecting to the host, defaults to HTTP.
- // +optional
- Scheme URIScheme
- // Optional: Custom headers to set in the request. HTTP allows repeated headers.
- // +optional
- HTTPHeaders []HTTPHeader
-}
-
-// URIScheme identifies the scheme used for connection to a host for Get actions
-type URIScheme string
-
-const (
- // URISchemeHTTP means that the scheme used will be http://
- URISchemeHTTP URIScheme = "HTTP"
- // URISchemeHTTPS means that the scheme used will be https://
- URISchemeHTTPS URIScheme = "HTTPS"
-)
-
-// TCPSocketAction describes an action based on opening a socket
-type TCPSocketAction struct {
- // Required: Port to connect to.
- // +optional
- Port intstr.IntOrString
- // Optional: Host name to connect to, defaults to the pod IP.
- // +optional
- Host string
-}
-
-// ExecAction describes a "run in container" action.
-type ExecAction struct {
- // Command is the command line to execute inside the container, the working directory for the
- // command is root ('/') in the container's filesystem. The command is simply exec'd, it is
- // not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use
- // a shell, you need to explicitly call out to that shell.
- // +optional
- Command []string
-}
-
-// Probe describes a health check to be performed against a container to determine whether it is
-// alive or ready to receive traffic.
-type Probe struct {
- // The action taken to determine the health of a container
- Handler
- // Length of time before health checking is activated. In seconds.
- // +optional
- InitialDelaySeconds int32
- // Length of time before health checking times out. In seconds.
- // +optional
- TimeoutSeconds int32
- // How often (in seconds) to perform the probe.
- // +optional
- PeriodSeconds int32
- // Minimum consecutive successes for the probe to be considered successful after having failed.
- // Must be 1 for liveness.
- // +optional
- SuccessThreshold int32
- // Minimum consecutive failures for the probe to be considered failed after having succeeded.
- // +optional
- FailureThreshold int32
-}
-
-// PullPolicy describes a policy for if/when to pull a container image
-type PullPolicy string
-
-const (
- // PullAlways means that kubelet always attempts to pull the latest image. Container will fail If the pull fails.
- PullAlways PullPolicy = "Always"
- // PullNever means that kubelet never pulls an image, but only uses a local image. Container will fail if the image isn't present
- PullNever PullPolicy = "Never"
- // PullIfNotPresent means that kubelet pulls if the image isn't present on disk. Container will fail if the image isn't present and the pull fails.
- PullIfNotPresent PullPolicy = "IfNotPresent"
-)
-
-// TerminationMessagePolicy describes how termination messages are retrieved from a container.
-type TerminationMessagePolicy string
-
-const (
- // TerminationMessageReadFile is the default behavior and will set the container status message to
- // the contents of the container's terminationMessagePath when the container exits.
- TerminationMessageReadFile TerminationMessagePolicy = "File"
- // TerminationMessageFallbackToLogsOnError will read the most recent contents of the container logs
- // for the container status message when the container exits with an error and the
- // terminationMessagePath has no contents.
- TerminationMessageFallbackToLogsOnError TerminationMessagePolicy = "FallbackToLogsOnError"
-)
-
-// Capability represent POSIX capabilities type
-type Capability string
-
-// Capabilities represent POSIX capabilities that can be added or removed to a running container.
-type Capabilities struct {
- // Added capabilities
- // +optional
- Add []Capability
- // Removed capabilities
- // +optional
- Drop []Capability
-}
-
-// ResourceRequirements describes the compute resource requirements.
-type ResourceRequirements struct {
- // Limits describes the maximum amount of compute resources allowed.
- // +optional
- Limits ResourceList
- // Requests describes the minimum amount of compute resources required.
- // If Request is omitted for a container, it defaults to Limits if that is explicitly specified,
- // otherwise to an implementation-defined value
- // +optional
- Requests ResourceList
-}
-
-// Container represents a single container that is expected to be run on the host.
-type Container struct {
- // Required: This must be a DNS_LABEL. Each container in a pod must
- // have a unique name.
- Name string
- // Required.
- Image string
- // Optional: The docker image's entrypoint is used if this is not provided; cannot be updated.
- // Variable references $(VAR_NAME) are expanded using the container's environment. If a variable
- // cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax
- // can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded,
- // regardless of whether the variable exists or not.
- // +optional
- Command []string
- // Optional: The docker image's cmd is used if this is not provided; cannot be updated.
- // Variable references $(VAR_NAME) are expanded using the container's environment. If a variable
- // cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax
- // can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded,
- // regardless of whether the variable exists or not.
- // +optional
- Args []string
- // Optional: Defaults to Docker's default.
- // +optional
- WorkingDir string
- // +optional
- Ports []ContainerPort
- // List of sources to populate environment variables in the container.
- // The keys defined within a source must be a C_IDENTIFIER. All invalid keys
- // will be reported as an event when the container is starting. When a key exists in multiple
- // sources, the value associated with the last source will take precedence.
- // Values defined by an Env with a duplicate key will take precedence.
- // Cannot be updated.
- // +optional
- EnvFrom []EnvFromSource
- // +optional
- Env []EnvVar
- // Compute resource requirements.
- // +optional
- Resources ResourceRequirements
- // +optional
- VolumeMounts []VolumeMount
- // volumeDevices is the list of block devices to be used by the container.
- // This is an alpha feature and may change in the future.
- // +optional
- VolumeDevices []VolumeDevice
- // +optional
- LivenessProbe *Probe
- // +optional
- ReadinessProbe *Probe
- // +optional
- Lifecycle *Lifecycle
- // Required.
- // +optional
- TerminationMessagePath string
- // +optional
- TerminationMessagePolicy TerminationMessagePolicy
- // Required: Policy for pulling images for this container
- ImagePullPolicy PullPolicy
- // Optional: SecurityContext defines the security options the container should be run with.
- // If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext.
- // +optional
- SecurityContext *SecurityContext
-
- // Variables for interactive containers, these have very specialized use-cases (e.g. debugging)
- // and shouldn't be used for general purpose containers.
- // +optional
- Stdin bool
- // +optional
- StdinOnce bool
- // +optional
- TTY bool
-}
-
-// Handler defines a specific action that should be taken
-// TODO: pass structured data to these actions, and document that data here.
-type Handler struct {
- // One and only one of the following should be specified.
- // Exec specifies the action to take.
- // +optional
- Exec *ExecAction
- // HTTPGet specifies the http request to perform.
- // +optional
- HTTPGet *HTTPGetAction
- // TCPSocket specifies an action involving a TCP port.
- // TODO: implement a realistic TCP lifecycle hook
- // +optional
- TCPSocket *TCPSocketAction
-}
-
-// Lifecycle describes actions that the management system should take in response to container lifecycle
-// events. For the PostStart and PreStop lifecycle handlers, management of the container blocks
-// until the action is complete, unless the container process fails, in which case the handler is aborted.
-type Lifecycle struct {
- // PostStart is called immediately after a container is created. If the handler fails, the container
- // is terminated and restarted.
- // +optional
- PostStart *Handler
- // PreStop is called immediately before a container is terminated. The reason for termination is
- // passed to the handler. Regardless of the outcome of the handler, the container is eventually terminated.
- // +optional
- PreStop *Handler
-}
-
-// The below types are used by kube_client and api_server.
-
-type ConditionStatus string
-
-// These are valid condition statuses. "ConditionTrue" means a resource is in the condition;
-// "ConditionFalse" means a resource is not in the condition; "ConditionUnknown" means kubernetes
-// can't decide if a resource is in the condition or not. In the future, we could add other
-// intermediate conditions, e.g. ConditionDegraded.
-const (
- ConditionTrue ConditionStatus = "True"
- ConditionFalse ConditionStatus = "False"
- ConditionUnknown ConditionStatus = "Unknown"
-)
-
-type ContainerStateWaiting struct {
- // A brief CamelCase string indicating details about why the container is in waiting state.
- // +optional
- Reason string
- // A human-readable message indicating details about why the container is in waiting state.
- // +optional
- Message string
-}
-
-type ContainerStateRunning struct {
- // +optional
- StartedAt metav1.Time
-}
-
-type ContainerStateTerminated struct {
- ExitCode int32
- // +optional
- Signal int32
- // +optional
- Reason string
- // +optional
- Message string
- // +optional
- StartedAt metav1.Time
- // +optional
- FinishedAt metav1.Time
- // +optional
- ContainerID string
-}
-
-// ContainerState holds a possible state of container.
-// Only one of its members may be specified.
-// If none of them is specified, the default one is ContainerStateWaiting.
-type ContainerState struct {
- // +optional
- Waiting *ContainerStateWaiting
- // +optional
- Running *ContainerStateRunning
- // +optional
- Terminated *ContainerStateTerminated
-}
-
-type ContainerStatus struct {
- // Each container in a pod must have a unique name.
- Name string
- // +optional
- State ContainerState
- // +optional
- LastTerminationState ContainerState
- // Ready specifies whether the container has passed its readiness check.
- Ready bool
- // Note that this is calculated from dead containers. But those containers are subject to
- // garbage collection. This value will get capped at 5 by GC.
- RestartCount int32
- Image string
- ImageID string
- // +optional
- ContainerID string
-}
-
-// PodPhase is a label for the condition of a pod at the current time.
-type PodPhase string
-
-// These are the valid statuses of pods.
-const (
- // PodPending means the pod has been accepted by the system, but one or more of the containers
- // has not been started. This includes time before being bound to a node, as well as time spent
- // pulling images onto the host.
- PodPending PodPhase = "Pending"
- // PodRunning means the pod has been bound to a node and all of the containers have been started.
- // At least one container is still running or is in the process of being restarted.
- PodRunning PodPhase = "Running"
- // PodSucceeded means that all containers in the pod have voluntarily terminated
- // with a container exit code of 0, and the system is not going to restart any of these containers.
- PodSucceeded PodPhase = "Succeeded"
- // PodFailed means that all containers in the pod have terminated, and at least one container has
- // terminated in a failure (exited with a non-zero exit code or was stopped by the system).
- PodFailed PodPhase = "Failed"
- // PodUnknown means that for some reason the state of the pod could not be obtained, typically due
- // to an error in communicating with the host of the pod.
- PodUnknown PodPhase = "Unknown"
-)
-
-type PodConditionType string
-
-// These are valid conditions of pod.
-const (
- // PodScheduled represents status of the scheduling process for this pod.
- PodScheduled PodConditionType = "PodScheduled"
- // PodReady means the pod is able to service requests and should be added to the
- // load balancing pools of all matching services.
- PodReady PodConditionType = "Ready"
- // PodInitialized means that all init containers in the pod have started successfully.
- PodInitialized PodConditionType = "Initialized"
- // PodReasonUnschedulable reason in PodScheduled PodCondition means that the scheduler
- // can't schedule the pod right now, for example due to insufficient resources in the cluster.
- PodReasonUnschedulable = "Unschedulable"
-)
-
-type PodCondition struct {
- Type PodConditionType
- Status ConditionStatus
- // +optional
- LastProbeTime metav1.Time
- // +optional
- LastTransitionTime metav1.Time
- // +optional
- Reason string
- // +optional
- Message string
-}
-
-// RestartPolicy describes how the container should be restarted.
-// Only one of the following restart policies may be specified.
-// If none of the following policies is specified, the default one
-// is RestartPolicyAlways.
-type RestartPolicy string
-
-const (
- RestartPolicyAlways RestartPolicy = "Always"
- RestartPolicyOnFailure RestartPolicy = "OnFailure"
- RestartPolicyNever RestartPolicy = "Never"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodList is a list of Pods.
-type PodList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Pod
-}
-
-// DNSPolicy defines how a pod's DNS will be configured.
-type DNSPolicy string
-
-const (
- // DNSClusterFirstWithHostNet indicates that the pod should use cluster DNS
- // first, if it is available, then fall back on the default
- // (as determined by kubelet) DNS settings.
- DNSClusterFirstWithHostNet DNSPolicy = "ClusterFirstWithHostNet"
-
- // DNSClusterFirst indicates that the pod should use cluster DNS
- // first unless hostNetwork is true, if it is available, then
- // fall back on the default (as determined by kubelet) DNS settings.
- DNSClusterFirst DNSPolicy = "ClusterFirst"
-
- // DNSDefault indicates that the pod should use the default (as
- // determined by kubelet) DNS settings.
- DNSDefault DNSPolicy = "Default"
-
- // DNSNone indicates that the pod should use empty DNS settings. DNS
- // parameters such as nameservers and search paths should be defined via
- // DNSConfig.
- DNSNone DNSPolicy = "None"
-)
-
-// A node selector represents the union of the results of one or more label queries
-// over a set of nodes; that is, it represents the OR of the selectors represented
-// by the node selector terms.
-type NodeSelector struct {
- //Required. A list of node selector terms. The terms are ORed.
- NodeSelectorTerms []NodeSelectorTerm
-}
-
-// A null or empty node selector term matches no objects.
-type NodeSelectorTerm struct {
- //Required. A list of node selector requirements. The requirements are ANDed.
- MatchExpressions []NodeSelectorRequirement
-}
-
-// A node selector requirement is a selector that contains values, a key, and an operator
-// that relates the key and values.
-type NodeSelectorRequirement struct {
- // The label key that the selector applies to.
- Key string
- // Represents a key's relationship to a set of values.
- // Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.
- Operator NodeSelectorOperator
- // An array of string values. If the operator is In or NotIn,
- // the values array must be non-empty. If the operator is Exists or DoesNotExist,
- // the values array must be empty. If the operator is Gt or Lt, the values
- // array must have a single element, which will be interpreted as an integer.
- // This array is replaced during a strategic merge patch.
- // +optional
- Values []string
-}
-
-// A node selector operator is the set of operators that can be used in
-// a node selector requirement.
-type NodeSelectorOperator string
-
-const (
- NodeSelectorOpIn NodeSelectorOperator = "In"
- NodeSelectorOpNotIn NodeSelectorOperator = "NotIn"
- NodeSelectorOpExists NodeSelectorOperator = "Exists"
- NodeSelectorOpDoesNotExist NodeSelectorOperator = "DoesNotExist"
- NodeSelectorOpGt NodeSelectorOperator = "Gt"
- NodeSelectorOpLt NodeSelectorOperator = "Lt"
-)
-
-// Affinity is a group of affinity scheduling rules.
-type Affinity struct {
- // Describes node affinity scheduling rules for the pod.
- // +optional
- NodeAffinity *NodeAffinity
- // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)).
- // +optional
- PodAffinity *PodAffinity
- // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)).
- // +optional
- PodAntiAffinity *PodAntiAffinity
-}
-
-// Pod affinity is a group of inter pod affinity scheduling rules.
-type PodAffinity struct {
- // NOT YET IMPLEMENTED. TODO: Uncomment field once it is implemented.
- // If the affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to a pod label update), the
- // system will try to eventually evict the pod from its node.
- // When there are multiple elements, the lists of nodes corresponding to each
- // podAffinityTerm are intersected, i.e. all terms must be satisfied.
- // +optional
- // RequiredDuringSchedulingRequiredDuringExecution []PodAffinityTerm
-
- // If the affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to a pod label update), the
- // system may or may not try to eventually evict the pod from its node.
- // When there are multiple elements, the lists of nodes corresponding to each
- // podAffinityTerm are intersected, i.e. all terms must be satisfied.
- // +optional
- RequiredDuringSchedulingIgnoredDuringExecution []PodAffinityTerm
- // The scheduler will prefer to schedule pods to nodes that satisfy
- // the affinity expressions specified by this field, but it may choose
- // a node that violates one or more of the expressions. The node that is
- // most preferred is the one with the greatest sum of weights, i.e.
- // for each node that meets all of the scheduling requirements (resource
- // request, requiredDuringScheduling affinity expressions, etc.),
- // compute a sum by iterating through the elements of this field and adding
- // "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the
- // node(s) with the highest sum are the most preferred.
- // +optional
- PreferredDuringSchedulingIgnoredDuringExecution []WeightedPodAffinityTerm
-}
-
-// Pod anti affinity is a group of inter pod anti affinity scheduling rules.
-type PodAntiAffinity struct {
- // NOT YET IMPLEMENTED. TODO: Uncomment field once it is implemented.
- // If the anti-affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the anti-affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to a pod label update), the
- // system will try to eventually evict the pod from its node.
- // When there are multiple elements, the lists of nodes corresponding to each
- // podAffinityTerm are intersected, i.e. all terms must be satisfied.
- // +optional
- // RequiredDuringSchedulingRequiredDuringExecution []PodAffinityTerm
-
- // If the anti-affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the anti-affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to a pod label update), the
- // system may or may not try to eventually evict the pod from its node.
- // When there are multiple elements, the lists of nodes corresponding to each
- // podAffinityTerm are intersected, i.e. all terms must be satisfied.
- // +optional
- RequiredDuringSchedulingIgnoredDuringExecution []PodAffinityTerm
- // The scheduler will prefer to schedule pods to nodes that satisfy
- // the anti-affinity expressions specified by this field, but it may choose
- // a node that violates one or more of the expressions. The node that is
- // most preferred is the one with the greatest sum of weights, i.e.
- // for each node that meets all of the scheduling requirements (resource
- // request, requiredDuringScheduling anti-affinity expressions, etc.),
- // compute a sum by iterating through the elements of this field and adding
- // "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the
- // node(s) with the highest sum are the most preferred.
- // +optional
- PreferredDuringSchedulingIgnoredDuringExecution []WeightedPodAffinityTerm
-}
-
-// The weights of all of the matched WeightedPodAffinityTerm fields are added per-node to find the most preferred node(s)
-type WeightedPodAffinityTerm struct {
- // weight associated with matching the corresponding podAffinityTerm,
- // in the range 1-100.
- Weight int32
- // Required. A pod affinity term, associated with the corresponding weight.
- PodAffinityTerm PodAffinityTerm
-}
-
-// Defines a set of pods (namely those matching the labelSelector
-// relative to the given namespace(s)) that this pod should be
-// co-located (affinity) or not co-located (anti-affinity) with,
-// where co-located is defined as running on a node whose value of
-// the label with key <topologyKey> matches that of any node on which
-// a pod of the set of pods is running.
-type PodAffinityTerm struct {
- // A label query over a set of resources, in this case pods.
- // +optional
- LabelSelector *metav1.LabelSelector
- // namespaces specifies which namespaces the labelSelector applies to (matches against);
- // null or empty list means "this pod's namespace"
- // +optional
- Namespaces []string
- // This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching
- // the labelSelector in the specified namespaces, where co-located is defined as running on a node
- // whose value of the label with key topologyKey matches that of any node on which any of the
- // selected pods is running.
- // Empty topologyKey is not allowed.
- TopologyKey string
-}
-
-// Node affinity is a group of node affinity scheduling rules.
-type NodeAffinity struct {
- // NOT YET IMPLEMENTED. TODO: Uncomment field once it is implemented.
- // If the affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to an update), the system
- // will try to eventually evict the pod from its node.
- // +optional
- // RequiredDuringSchedulingRequiredDuringExecution *NodeSelector
-
- // If the affinity requirements specified by this field are not met at
- // scheduling time, the pod will not be scheduled onto the node.
- // If the affinity requirements specified by this field cease to be met
- // at some point during pod execution (e.g. due to an update), the system
- // may or may not try to eventually evict the pod from its node.
- // +optional
- RequiredDuringSchedulingIgnoredDuringExecution *NodeSelector
- // The scheduler will prefer to schedule pods to nodes that satisfy
- // the affinity expressions specified by this field, but it may choose
- // a node that violates one or more of the expressions. The node that is
- // most preferred is the one with the greatest sum of weights, i.e.
- // for each node that meets all of the scheduling requirements (resource
- // request, requiredDuringScheduling affinity expressions, etc.),
- // compute a sum by iterating through the elements of this field and adding
- // "weight" to the sum if the node matches the corresponding matchExpressions; the
- // node(s) with the highest sum are the most preferred.
- // +optional
- PreferredDuringSchedulingIgnoredDuringExecution []PreferredSchedulingTerm
-}
-
-// An empty preferred scheduling term matches all objects with implicit weight 0
-// (i.e. it's a no-op). A null preferred scheduling term matches no objects (i.e. is also a no-op).
-type PreferredSchedulingTerm struct {
- // Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100.
- Weight int32
- // A node selector term, associated with the corresponding weight.
- Preference NodeSelectorTerm
-}
-
-// The node this Taint is attached to has the "effect" on
-// any pod that does not tolerate the Taint.
-type Taint struct {
- // Required. The taint key to be applied to a node.
- Key string
- // Required. The taint value corresponding to the taint key.
- // +optional
- Value string
- // Required. The effect of the taint on pods
- // that do not tolerate the taint.
- // Valid effects are NoSchedule, PreferNoSchedule and NoExecute.
- Effect TaintEffect
- // TimeAdded represents the time at which the taint was added.
- // It is only written for NoExecute taints.
- // +optional
- TimeAdded *metav1.Time
-}
-
-type TaintEffect string
-
-const (
- // Do not allow new pods to schedule onto the node unless they tolerate the taint,
- // but allow all pods submitted to Kubelet without going through the scheduler
- // to start, and allow all already-running pods to continue running.
- // Enforced by the scheduler.
- TaintEffectNoSchedule TaintEffect = "NoSchedule"
- // Like TaintEffectNoSchedule, but the scheduler tries not to schedule
- // new pods onto the node, rather than prohibiting new pods from scheduling
- // onto the node entirely. Enforced by the scheduler.
- TaintEffectPreferNoSchedule TaintEffect = "PreferNoSchedule"
- // NOT YET IMPLEMENTED. TODO: Uncomment field once it is implemented.
- // Like TaintEffectNoSchedule, but additionally do not allow pods submitted to
- // Kubelet without going through the scheduler to start.
- // Enforced by Kubelet and the scheduler.
- // TaintEffectNoScheduleNoAdmit TaintEffect = "NoScheduleNoAdmit"
-
- // Evict any already-running pods that do not tolerate the taint.
- // Currently enforced by NodeController.
- TaintEffectNoExecute TaintEffect = "NoExecute"
-)
-
-// The pod this Toleration is attached to tolerates any taint that matches
-// the triple <key,value,effect> using the matching operator <operator>.
-type Toleration struct {
- // Key is the taint key that the toleration applies to. Empty means match all taint keys.
- // If the key is empty, operator must be Exists; this combination means to match all values and all keys.
- // +optional
- Key string
- // Operator represents a key's relationship to the value.
- // Valid operators are Exists and Equal. Defaults to Equal.
- // Exists is equivalent to wildcard for value, so that a pod can
- // tolerate all taints of a particular category.
- // +optional
- Operator TolerationOperator
- // Value is the taint value the toleration matches to.
- // If the operator is Exists, the value should be empty, otherwise just a regular string.
- // +optional
- Value string
- // Effect indicates the taint effect to match. Empty means match all taint effects.
- // When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute.
- // +optional
- Effect TaintEffect
- // TolerationSeconds represents the period of time the toleration (which must be
- // of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default,
- // it is not set, which means tolerate the taint forever (do not evict). Zero and
- // negative values will be treated as 0 (evict immediately) by the system.
- // +optional
- TolerationSeconds *int64
-}
-
-// A toleration operator is the set of operators that can be used in a toleration.
-type TolerationOperator string
-
-const (
- TolerationOpExists TolerationOperator = "Exists"
- TolerationOpEqual TolerationOperator = "Equal"
-)
-
-// PodSpec is a description of a pod
-type PodSpec struct {
- Volumes []Volume
- // List of initialization containers belonging to the pod.
- InitContainers []Container
- // List of containers belonging to the pod.
- Containers []Container
- // +optional
- RestartPolicy RestartPolicy
- // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request.
- // Value must be non-negative integer. The value zero indicates delete immediately.
- // If this value is nil, the default grace period will be used instead.
- // The grace period is the duration in seconds after the processes running in the pod are sent
- // a termination signal and the time when the processes are forcibly halted with a kill signal.
- // Set this value longer than the expected cleanup time for your process.
- // +optional
- TerminationGracePeriodSeconds *int64
- // Optional duration in seconds relative to the StartTime that the pod may be active on a node
- // before the system actively tries to terminate the pod; value must be positive integer
- // +optional
- ActiveDeadlineSeconds *int64
- // Set DNS policy for the pod.
- // Defaults to "ClusterFirst".
- // Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'.
- // DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy.
- // To have DNS options set along with hostNetwork, you have to specify DNS policy
- // explicitly to 'ClusterFirstWithHostNet'.
- // +optional
- DNSPolicy DNSPolicy
- // NodeSelector is a selector which must be true for the pod to fit on a node
- // +optional
- NodeSelector map[string]string
-
- // ServiceAccountName is the name of the ServiceAccount to use to run this pod
- // The pod will be allowed to use secrets referenced by the ServiceAccount
- ServiceAccountName string
- // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.
- // +optional
- AutomountServiceAccountToken *bool
-
- // NodeName is a request to schedule this pod onto a specific node. If it is non-empty,
- // the scheduler simply schedules this pod onto that node, assuming that it fits resource
- // requirements.
- // +optional
- NodeName string
- // SecurityContext holds pod-level security attributes and common container settings.
- // Optional: Defaults to empty. See type description for default values of each field.
- // +optional
- SecurityContext *PodSecurityContext
- // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec.
- // If specified, these secrets will be passed to individual puller implementations for them to use. For example,
- // in the case of docker, only DockerConfig type secrets are honored.
- // +optional
- ImagePullSecrets []LocalObjectReference
- // Specifies the hostname of the Pod.
- // If not specified, the pod's hostname will be set to a system-defined value.
- // +optional
- Hostname string
- // If specified, the fully qualified Pod hostname will be "<hostname>.<subdomain>.<pod namespace>.svc.<cluster domain>".
- // If not specified, the pod will not have a domainname at all.
- // +optional
- Subdomain string
- // If specified, the pod's scheduling constraints
- // +optional
- Affinity *Affinity
- // If specified, the pod will be dispatched by specified scheduler.
- // If not specified, the pod will be dispatched by default scheduler.
- // +optional
- SchedulerName string
- // If specified, the pod's tolerations.
- // +optional
- Tolerations []Toleration
- // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts
- // file if specified. This is only valid for non-hostNetwork pods.
- // +optional
- HostAliases []HostAlias
- // If specified, indicates the pod's priority. "system-node-critical" and
- // "system-cluster-critical" are two special keywords which indicate the
- // highest priorities with the former being the highest priority. Any other
- // name must be defined by creating a PriorityClass object with that name.
- // If not specified, the pod priority will be default or zero if there is no
- // default.
- // +optional
- PriorityClassName string
- // The priority value. Various system components use this field to find the
- // priority of the pod. When Priority Admission Controller is enabled, it
- // prevents users from setting this field. The admission controller populates
- // this field from PriorityClassName.
- // The higher the value, the higher the priority.
- // +optional
- Priority *int32
- // Specifies the DNS parameters of a pod.
- // Parameters specified here will be merged to the generated DNS
- // configuration based on DNSPolicy.
- // +optional
- DNSConfig *PodDNSConfig
-}
-
-// HostAlias holds the mapping between IP and hostnames that will be injected as an entry in the
-// pod's hosts file.
-type HostAlias struct {
- IP string
- Hostnames []string
-}
-
-// Sysctl defines a kernel parameter to be set
-type Sysctl struct {
- // Name of a property to set
- Name string
- // Value of a property to set
- Value string
-}
-
-// PodSecurityContext holds pod-level security attributes and common container settings.
-// Some fields are also present in container.securityContext. Field values of
-// container.securityContext take precedence over field values of PodSecurityContext.
-type PodSecurityContext struct {
- // Use the host's network namespace. If this option is set, the ports that will be
- // used must be specified.
- // Optional: Default to false
- // +k8s:conversion-gen=false
- // +optional
- HostNetwork bool
- // Use the host's pid namespace.
- // Optional: Default to false.
- // +k8s:conversion-gen=false
- // +optional
- HostPID bool
- // Use the host's ipc namespace.
- // Optional: Default to false.
- // +k8s:conversion-gen=false
- // +optional
- HostIPC bool
- // Share a single process namespace between all of the containers in a pod.
- // When this is set containers will be able to view and signal processes from other containers
- // in the same pod, and the first process in each container will not be assigned PID 1.
- // HostPID and ShareProcessNamespace cannot both be set.
- // Optional: Default to false.
- // This field is alpha-level and is honored only by servers that enable the PodShareProcessNamespace feature.
- // +k8s:conversion-gen=false
- // +optional
- ShareProcessNamespace *bool
- // The SELinux context to be applied to all containers.
- // If unspecified, the container runtime will allocate a random SELinux context for each
- // container. May also be set in SecurityContext. If set in
- // both SecurityContext and PodSecurityContext, the value specified in SecurityContext
- // takes precedence for that container.
- // +optional
- SELinuxOptions *SELinuxOptions
- // The UID to run the entrypoint of the container process.
- // Defaults to user specified in image metadata if unspecified.
- // May also be set in SecurityContext. If set in both SecurityContext and
- // PodSecurityContext, the value specified in SecurityContext takes precedence
- // for that container.
- // +optional
- RunAsUser *int64
- // Indicates that the container must run as a non-root user.
- // If true, the Kubelet will validate the image at runtime to ensure that it
- // does not run as UID 0 (root) and fail to start the container if it does.
- // If unset or false, no such validation will be performed.
- // May also be set in SecurityContext. If set in both SecurityContext and
- // PodSecurityContext, the value specified in SecurityContext takes precedence.
- // +optional
- RunAsNonRoot *bool
- // A list of groups applied to the first process run in each container, in addition
- // to the container's primary GID. If unspecified, no groups will be added to
- // any container.
- // +optional
- SupplementalGroups []int64
- // A special supplemental group that applies to all containers in a pod.
- // Some volume types allow the Kubelet to change the ownership of that volume
- // to be owned by the pod:
- //
- // 1. The owning GID will be the FSGroup
- // 2. The setgid bit is set (new files created in the volume will be owned by FSGroup)
- // 3. The permission bits are OR'd with rw-rw----
- //
- // If unset, the Kubelet will not modify the ownership and permissions of any volume.
- // +optional
- FSGroup *int64
-}
-
-// PodQOSClass defines the supported qos classes of Pods.
-type PodQOSClass string
-
-const (
- // PodQOSGuaranteed is the Guaranteed qos class.
- PodQOSGuaranteed PodQOSClass = "Guaranteed"
- // PodQOSBurstable is the Burstable qos class.
- PodQOSBurstable PodQOSClass = "Burstable"
- // PodQOSBestEffort is the BestEffort qos class.
- PodQOSBestEffort PodQOSClass = "BestEffort"
-)
-
-// PodDNSConfig defines the DNS parameters of a pod in addition to
-// those generated from DNSPolicy.
-type PodDNSConfig struct {
- // A list of DNS name server IP addresses.
- // This will be appended to the base nameservers generated from DNSPolicy.
- // Duplicated nameservers will be removed.
- // +optional
- Nameservers []string
- // A list of DNS search domains for host-name lookup.
- // This will be appended to the base search paths generated from DNSPolicy.
- // Duplicated search paths will be removed.
- // +optional
- Searches []string
- // A list of DNS resolver options.
- // This will be merged with the base options generated from DNSPolicy.
- // Duplicated entries will be removed. Resolution options given in Options
- // will override those that appear in the base DNSPolicy.
- // +optional
- Options []PodDNSConfigOption
-}
-
-// PodDNSConfigOption defines DNS resolver options of a pod.
-type PodDNSConfigOption struct {
- // Required.
- Name string
- // +optional
- Value *string
-}
-
-// PodStatus represents information about the status of a pod. Status may trail the actual
-// state of a system.
-type PodStatus struct {
- // +optional
- Phase PodPhase
- // +optional
- Conditions []PodCondition
- // A human readable message indicating details about why the pod is in this state.
- // +optional
- Message string
- // A brief CamelCase message indicating details about why the pod is in this state. e.g. 'Evicted'
- // +optional
- Reason string
- // nominatedNodeName is set when this pod preempts other pods on the node, but it cannot be
- // scheduled right away as preemption victims receive their graceful termination periods.
- // This field does not guarantee that the pod will be scheduled on this node. Scheduler may decide
- // to place the pod elsewhere if other nodes become available sooner. Scheduler may also decide to
- // give the resources on this node to a higher priority pod that is created after preemption.
- // +optional
- NominatedNodeName string
-
- // +optional
- HostIP string
- // +optional
- PodIP string
-
- // Date and time at which the object was acknowledged by the Kubelet.
- // This is before the Kubelet pulled the container image(s) for the pod.
- // +optional
- StartTime *metav1.Time
- // +optional
- QOSClass PodQOSClass
-
- // The list has one entry per init container in the manifest. The most recent successful
- // init container will have ready = true, the most recently started container will have
- // startTime set.
- // More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#pod-and-container-status
- InitContainerStatuses []ContainerStatus
- // The list has one entry per container in the manifest. Each entry is
- // currently the output of `docker inspect`. This output format is *not*
- // final and should not be relied upon.
- // TODO: Make real decisions about what our info should look like. Re-enable fuzz test
- // when we have done this.
- // +optional
- ContainerStatuses []ContainerStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodStatusResult is a wrapper for PodStatus returned by kubelet that can be encode/decoded
-type PodStatusResult struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
- // Status represents the current information about a pod. This data may not be up
- // to date.
- // +optional
- Status PodStatus
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Pod is a collection of containers, used as either input (create, update) or as output (list, get).
-type Pod struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the behavior of a pod.
- // +optional
- Spec PodSpec
-
- // Status represents the current information about a pod. This data may not be up
- // to date.
- // +optional
- Status PodStatus
-}
-
-// PodTemplateSpec describes the data a pod should have when created from a template
-type PodTemplateSpec struct {
- // Metadata of the pods created from this template.
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the behavior of a pod.
- // +optional
- Spec PodSpec
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodTemplate describes a template for creating copies of a predefined pod.
-type PodTemplate struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Template defines the pods that will be created from this pod template
- // +optional
- Template PodTemplateSpec
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodTemplateList is a list of PodTemplates.
-type PodTemplateList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []PodTemplate
-}
-
-// ReplicationControllerSpec is the specification of a replication controller.
-// As the internal representation of a replication controller, it may have either
-// a TemplateRef or a Template set.
-type ReplicationControllerSpec struct {
- // Replicas is the number of desired replicas.
- Replicas int32
-
- // Minimum number of seconds for which a newly created pod should be ready
- // without any of its container crashing, for it to be considered available.
- // Defaults to 0 (pod will be considered available as soon as it is ready)
- // +optional
- MinReadySeconds int32
-
- // Selector is a label query over pods that should match the Replicas count.
- Selector map[string]string
-
- // TemplateRef is a reference to an object that describes the pod that will be created if
- // insufficient replicas are detected. This reference is ignored if a Template is set.
- // Must be set before converting to a versioned API object
- // +optional
- //TemplateRef *ObjectReference
-
- // Template is the object that describes the pod that will be created if
- // insufficient replicas are detected. Internally, this takes precedence over a
- // TemplateRef.
- // +optional
- Template *PodTemplateSpec
-}
-
-// ReplicationControllerStatus represents the current status of a replication
-// controller.
-type ReplicationControllerStatus struct {
- // Replicas is the number of actual replicas.
- Replicas int32
-
- // The number of pods that have labels matching the labels of the pod template of the replication controller.
- // +optional
- FullyLabeledReplicas int32
-
- // The number of ready replicas for this replication controller.
- // +optional
- ReadyReplicas int32
-
- // The number of available replicas (ready for at least minReadySeconds) for this replication controller.
- // +optional
- AvailableReplicas int32
-
- // ObservedGeneration is the most recent generation observed by the controller.
- // +optional
- ObservedGeneration int64
-
- // Represents the latest available observations of a replication controller's current state.
- // +optional
- Conditions []ReplicationControllerCondition
-}
-
-type ReplicationControllerConditionType string
-
-// These are valid conditions of a replication controller.
-const (
- // ReplicationControllerReplicaFailure is added in a replication controller when one of its pods
- // fails to be created due to insufficient quota, limit ranges, pod security policy, node selectors,
- // etc. or deleted due to kubelet being down or finalizers are failing.
- ReplicationControllerReplicaFailure ReplicationControllerConditionType = "ReplicaFailure"
-)
-
-// ReplicationControllerCondition describes the state of a replication controller at a certain point.
-type ReplicationControllerCondition struct {
- // Type of replication controller condition.
- Type ReplicationControllerConditionType
- // Status of the condition, one of True, False, Unknown.
- Status ConditionStatus
- // The last time the condition transitioned from one status to another.
- // +optional
- LastTransitionTime metav1.Time
- // The reason for the condition's last transition.
- // +optional
- Reason string
- // A human readable message indicating details about the transition.
- // +optional
- Message string
-}
-
-// +genclient
-// +genclient:method=GetScale,verb=get,subresource=scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +genclient:method=UpdateScale,verb=update,subresource=scale,input=k8s.io/kubernetes/pkg/apis/autoscaling.Scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ReplicationController represents the configuration of a replication controller.
-type ReplicationController struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the desired behavior of this replication controller.
- // +optional
- Spec ReplicationControllerSpec
-
- // Status is the current status of this replication controller. This data may be
- // out of date by some window of time.
- // +optional
- Status ReplicationControllerStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ReplicationControllerList is a collection of replication controllers.
-type ReplicationControllerList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []ReplicationController
-}
-
-const (
- // ClusterIPNone - do not assign a cluster IP
- // no proxying required and no environment variables should be created for pods
- ClusterIPNone = "None"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ServiceList holds a list of services.
-type ServiceList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Service
-}
-
-// Session Affinity Type string
-type ServiceAffinity string
-
-const (
- // ServiceAffinityClientIP is the Client IP based.
- ServiceAffinityClientIP ServiceAffinity = "ClientIP"
-
- // ServiceAffinityNone - no session affinity.
- ServiceAffinityNone ServiceAffinity = "None"
-)
-
-const (
- // DefaultClientIPServiceAffinitySeconds is the default timeout seconds
- // of Client IP based session affinity - 3 hours.
- DefaultClientIPServiceAffinitySeconds int32 = 10800
- // MaxClientIPServiceAffinitySeconds is the max timeout seconds
- // of Client IP based session affinity - 1 day.
- MaxClientIPServiceAffinitySeconds int32 = 86400
-)
-
-// SessionAffinityConfig represents the configurations of session affinity.
-type SessionAffinityConfig struct {
- // clientIP contains the configurations of Client IP based session affinity.
- // +optional
- ClientIP *ClientIPConfig
-}
-
-// ClientIPConfig represents the configurations of Client IP based session affinity.
-type ClientIPConfig struct {
- // timeoutSeconds specifies the seconds of ClientIP type session sticky time.
- // The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP".
- // Default value is 10800(for 3 hours).
- // +optional
- TimeoutSeconds *int32
-}
-
-// Service Type string describes ingress methods for a service
-type ServiceType string
-
-const (
- // ServiceTypeClusterIP means a service will only be accessible inside the
- // cluster, via the ClusterIP.
- ServiceTypeClusterIP ServiceType = "ClusterIP"
-
- // ServiceTypeNodePort means a service will be exposed on one port of
- // every node, in addition to 'ClusterIP' type.
- ServiceTypeNodePort ServiceType = "NodePort"
-
- // ServiceTypeLoadBalancer means a service will be exposed via an
- // external load balancer (if the cloud provider supports it), in addition
- // to 'NodePort' type.
- ServiceTypeLoadBalancer ServiceType = "LoadBalancer"
-
- // ServiceTypeExternalName means a service consists of only a reference to
- // an external name that kubedns or equivalent will return as a CNAME
- // record, with no exposing or proxying of any pods involved.
- ServiceTypeExternalName ServiceType = "ExternalName"
-)
-
-// Service External Traffic Policy Type string
-type ServiceExternalTrafficPolicyType string
-
-const (
- // ServiceExternalTrafficPolicyTypeLocal specifies node-local endpoints behavior.
- ServiceExternalTrafficPolicyTypeLocal ServiceExternalTrafficPolicyType = "Local"
- // ServiceExternalTrafficPolicyTypeCluster specifies cluster-wide (legacy) behavior.
- ServiceExternalTrafficPolicyTypeCluster ServiceExternalTrafficPolicyType = "Cluster"
-)
-
-// ServiceStatus represents the current status of a service
-type ServiceStatus struct {
- // LoadBalancer contains the current status of the load-balancer,
- // if one is present.
- // +optional
- LoadBalancer LoadBalancerStatus
-}
-
-// LoadBalancerStatus represents the status of a load-balancer
-type LoadBalancerStatus struct {
- // Ingress is a list containing ingress points for the load-balancer;
- // traffic intended for the service should be sent to these ingress points.
- // +optional
- Ingress []LoadBalancerIngress
-}
-
-// LoadBalancerIngress represents the status of a load-balancer ingress point:
-// traffic intended for the service should be sent to an ingress point.
-type LoadBalancerIngress struct {
- // IP is set for load-balancer ingress points that are IP based
- // (typically GCE or OpenStack load-balancers)
- // +optional
- IP string
-
- // Hostname is set for load-balancer ingress points that are DNS based
- // (typically AWS load-balancers)
- // +optional
- Hostname string
-}
-
-// ServiceSpec describes the attributes that a user creates on a service
-type ServiceSpec struct {
- // Type determines how the Service is exposed. Defaults to ClusterIP. Valid
- // options are ExternalName, ClusterIP, NodePort, and LoadBalancer.
- // "ExternalName" maps to the specified externalName.
- // "ClusterIP" allocates a cluster-internal IP address for load-balancing to
- // endpoints. Endpoints are determined by the selector or if that is not
- // specified, by manual construction of an Endpoints object. If clusterIP is
- // "None", no virtual IP is allocated and the endpoints are published as a
- // set of endpoints rather than a stable IP.
- // "NodePort" builds on ClusterIP and allocates a port on every node which
- // routes to the clusterIP.
- // "LoadBalancer" builds on NodePort and creates an
- // external load-balancer (if supported in the current cloud) which routes
- // to the clusterIP.
- // More info: https://kubernetes.io/docs/concepts/services-networking/service/
- // +optional
- Type ServiceType
-
- // Required: The list of ports that are exposed by this service.
- Ports []ServicePort
-
- // Route service traffic to pods with label keys and values matching this
- // selector. If empty or not present, the service is assumed to have an
- // external process managing its endpoints, which Kubernetes will not
- // modify. Only applies to types ClusterIP, NodePort, and LoadBalancer.
- // Ignored if type is ExternalName.
- // More info: https://kubernetes.io/docs/concepts/services-networking/service/
- Selector map[string]string
-
- // ClusterIP is the IP address of the service and is usually assigned
- // randomly by the master. If an address is specified manually and is not in
- // use by others, it will be allocated to the service; otherwise, creation
- // of the service will fail. This field can not be changed through updates.
- // Valid values are "None", empty string (""), or a valid IP address. "None"
- // can be specified for headless services when proxying is not required.
- // Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if
- // type is ExternalName.
- // More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies
- // +optional
- ClusterIP string
-
- // ExternalName is the external reference that kubedns or equivalent will
- // return as a CNAME record for this service. No proxying will be involved.
- // Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123)
- // and requires Type to be ExternalName.
- ExternalName string
-
- // ExternalIPs are used by external load balancers, or can be set by
- // users to handle external traffic that arrives at a node.
- // +optional
- ExternalIPs []string
-
- // Only applies to Service Type: LoadBalancer
- // LoadBalancer will get created with the IP specified in this field.
- // This feature depends on whether the underlying cloud-provider supports specifying
- // the loadBalancerIP when a load balancer is created.
- // This field will be ignored if the cloud-provider does not support the feature.
- // +optional
- LoadBalancerIP string
-
- // Optional: Supports "ClientIP" and "None". Used to maintain session affinity.
- // +optional
- SessionAffinity ServiceAffinity
-
- // sessionAffinityConfig contains the configurations of session affinity.
- // +optional
- SessionAffinityConfig *SessionAffinityConfig
-
- // Optional: If specified and supported by the platform, this will restrict traffic through the cloud-provider
- // load-balancer will be restricted to the specified client IPs. This field will be ignored if the
- // cloud-provider does not support the feature."
- // +optional
- LoadBalancerSourceRanges []string
-
- // externalTrafficPolicy denotes if this Service desires to route external
- // traffic to node-local or cluster-wide endpoints. "Local" preserves the
- // client source IP and avoids a second hop for LoadBalancer and Nodeport
- // type services, but risks potentially imbalanced traffic spreading.
- // "Cluster" obscures the client source IP and may cause a second hop to
- // another node, but should have good overall load-spreading.
- // +optional
- ExternalTrafficPolicy ServiceExternalTrafficPolicyType
-
- // healthCheckNodePort specifies the healthcheck nodePort for the service.
- // If not specified, HealthCheckNodePort is created by the service api
- // backend with the allocated nodePort. Will use user-specified nodePort value
- // if specified by the client. Only effects when Type is set to LoadBalancer
- // and ExternalTrafficPolicy is set to Local.
- // +optional
- HealthCheckNodePort int32
-
- // publishNotReadyAddresses, when set to true, indicates that DNS implementations
- // must publish the notReadyAddresses of subsets for the Endpoints associated with
- // the Service. The default value is false.
- // The primary use case for setting this field is to use a StatefulSet's Headless Service
- // to propagate SRV records for its Pods without respect to their readiness for purpose
- // of peer discovery.
- // This field will replace the service.alpha.kubernetes.io/tolerate-unready-endpoints
- // when that annotation is deprecated and all clients have been converted to use this
- // field.
- // +optional
- PublishNotReadyAddresses bool
-}
-
-type ServicePort struct {
- // Optional if only one ServicePort is defined on this service: The
- // name of this port within the service. This must be a DNS_LABEL.
- // All ports within a ServiceSpec must have unique names. This maps to
- // the 'Name' field in EndpointPort objects.
- Name string
-
- // The IP protocol for this port. Supports "TCP" and "UDP".
- Protocol Protocol
-
- // The port that will be exposed on the service.
- Port int32
-
- // Optional: The target port on pods selected by this service. If this
- // is a string, it will be looked up as a named port in the target
- // Pod's container ports. If this is not specified, the value
- // of the 'port' field is used (an identity map).
- // This field is ignored for services with clusterIP=None, and should be
- // omitted or set equal to the 'port' field.
- TargetPort intstr.IntOrString
-
- // The port on each node on which this service is exposed.
- // Default is to auto-allocate a port if the ServiceType of this Service requires one.
- NodePort int32
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Service is a named abstraction of software service (for example, mysql) consisting of local port
-// (for example 3306) that the proxy listens on, and the selector that determines which pods
-// will answer requests sent through the proxy.
-type Service struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the behavior of a service.
- // +optional
- Spec ServiceSpec
-
- // Status represents the current status of a service.
- // +optional
- Status ServiceStatus
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ServiceAccount binds together:
-// * a name, understood by users, and perhaps by peripheral systems, for an identity
-// * a principal that can be authenticated and authorized
-// * a set of secrets
-type ServiceAccount struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Secrets is the list of secrets allowed to be used by pods running using this ServiceAccount
- Secrets []ObjectReference
-
- // ImagePullSecrets is a list of references to secrets in the same namespace to use for pulling any images
- // in pods that reference this ServiceAccount. ImagePullSecrets are distinct from Secrets because Secrets
- // can be mounted in the pod, but ImagePullSecrets are only accessed by the kubelet.
- // +optional
- ImagePullSecrets []LocalObjectReference
-
- // AutomountServiceAccountToken indicates whether pods running as this service account should have an API token automatically mounted.
- // Can be overridden at the pod level.
- // +optional
- AutomountServiceAccountToken *bool
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ServiceAccountList is a list of ServiceAccount objects
-type ServiceAccountList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []ServiceAccount
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Endpoints is a collection of endpoints that implement the actual service. Example:
-// Name: "mysvc",
-// Subsets: [
-// {
-// Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}],
-// Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}]
-// },
-// {
-// Addresses: [{"ip": "10.10.3.3"}],
-// Ports: [{"name": "a", "port": 93}, {"name": "b", "port": 76}]
-// },
-// ]
-type Endpoints struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // The set of all endpoints is the union of all subsets.
- Subsets []EndpointSubset
-}
-
-// EndpointSubset is a group of addresses with a common set of ports. The
-// expanded set of endpoints is the Cartesian product of Addresses x Ports.
-// For example, given:
-// {
-// Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}],
-// Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}]
-// }
-// The resulting set of endpoints can be viewed as:
-// a: [ 10.10.1.1:8675, 10.10.2.2:8675 ],
-// b: [ 10.10.1.1:309, 10.10.2.2:309 ]
-type EndpointSubset struct {
- Addresses []EndpointAddress
- NotReadyAddresses []EndpointAddress
- Ports []EndpointPort
-}
-
-// EndpointAddress is a tuple that describes single IP address.
-type EndpointAddress struct {
- // The IP of this endpoint.
- // IPv6 is also accepted but not fully supported on all platforms. Also, certain
- // kubernetes components, like kube-proxy, are not IPv6 ready.
- // TODO: This should allow hostname or IP, see #4447.
- IP string
- // Optional: Hostname of this endpoint
- // Meant to be used by DNS servers etc.
- // +optional
- Hostname string
- // Optional: Node hosting this endpoint. This can be used to determine endpoints local to a node.
- // +optional
- NodeName *string
- // Optional: The kubernetes object related to the entry point.
- TargetRef *ObjectReference
-}
-
-// EndpointPort is a tuple that describes a single port.
-type EndpointPort struct {
- // The name of this port (corresponds to ServicePort.Name). Optional
- // if only one port is defined. Must be a DNS_LABEL.
- Name string
-
- // The port number.
- Port int32
-
- // The IP protocol for this port.
- Protocol Protocol
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// EndpointsList is a list of endpoints.
-type EndpointsList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Endpoints
-}
-
-// NodeSpec describes the attributes that a node is created with.
-type NodeSpec struct {
- // PodCIDR represents the pod IP range assigned to the node
- // Note: assigning IP ranges to nodes might need to be revisited when we support migratable IPs.
- // +optional
- PodCIDR string
-
- // External ID of the node assigned by some machine database (e.g. a cloud provider)
- // +optional
- ExternalID string
-
- // ID of the node assigned by the cloud provider
- // Note: format is "<ProviderName>://<ProviderSpecificNodeID>"
- // +optional
- ProviderID string
-
- // Unschedulable controls node schedulability of new pods. By default node is schedulable.
- // +optional
- Unschedulable bool
-
- // If specified, the node's taints.
- // +optional
- Taints []Taint
-
- // If specified, the source to get node configuration from
- // The DynamicKubeletConfig feature gate must be enabled for the Kubelet to use this field
- // +optional
- ConfigSource *NodeConfigSource
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NodeConfigSource specifies a source of node configuration. Exactly one subfield must be non-nil.
-type NodeConfigSource struct {
- metav1.TypeMeta
- ConfigMapRef *ObjectReference
-}
-
-// DaemonEndpoint contains information about a single Daemon endpoint.
-type DaemonEndpoint struct {
- /*
- The port tag was not properly in quotes in earlier releases, so it must be
- uppercase for backwards compatibility (since it was falling back to var name of
- 'Port').
- */
-
- // Port number of the given endpoint.
- Port int32
-}
-
-// NodeDaemonEndpoints lists ports opened by daemons running on the Node.
-type NodeDaemonEndpoints struct {
- // Endpoint on which Kubelet is listening.
- // +optional
- KubeletEndpoint DaemonEndpoint
-}
-
-// NodeSystemInfo is a set of ids/uuids to uniquely identify the node.
-type NodeSystemInfo struct {
- // MachineID reported by the node. For unique machine identification
- // in the cluster this field is preferred. Learn more from man(5)
- // machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html
- MachineID string
- // SystemUUID reported by the node. For unique machine identification
- // MachineID is preferred. This field is specific to Red Hat hosts
- // https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html
- SystemUUID string
- // Boot ID reported by the node.
- BootID string
- // Kernel Version reported by the node.
- KernelVersion string
- // OS Image reported by the node.
- OSImage string
- // ContainerRuntime Version reported by the node.
- ContainerRuntimeVersion string
- // Kubelet Version reported by the node.
- KubeletVersion string
- // KubeProxy Version reported by the node.
- KubeProxyVersion string
- // The Operating System reported by the node
- OperatingSystem string
- // The Architecture reported by the node
- Architecture string
-}
-
-// NodeStatus is information about the current status of a node.
-type NodeStatus struct {
- // Capacity represents the total resources of a node.
- // +optional
- Capacity ResourceList
- // Allocatable represents the resources of a node that are available for scheduling.
- // +optional
- Allocatable ResourceList
- // NodePhase is the current lifecycle phase of the node.
- // +optional
- Phase NodePhase
- // Conditions is an array of current node conditions.
- // +optional
- Conditions []NodeCondition
- // Queried from cloud provider, if available.
- // +optional
- Addresses []NodeAddress
- // Endpoints of daemons running on the Node.
- // +optional
- DaemonEndpoints NodeDaemonEndpoints
- // Set of ids/uuids to uniquely identify the node.
- // +optional
- NodeInfo NodeSystemInfo
- // List of container images on this node
- // +optional
- Images []ContainerImage
- // List of attachable volumes in use (mounted) by the node.
- // +optional
- VolumesInUse []UniqueVolumeName
- // List of volumes that are attached to the node.
- // +optional
- VolumesAttached []AttachedVolume
-}
-
-type UniqueVolumeName string
-
-// AttachedVolume describes a volume attached to a node
-type AttachedVolume struct {
- // Name of the attached volume
- Name UniqueVolumeName
-
- // DevicePath represents the device path where the volume should be available
- DevicePath string
-}
-
-// AvoidPods describes pods that should avoid this node. This is the value for a
-// Node annotation with key scheduler.alpha.kubernetes.io/preferAvoidPods and
-// will eventually become a field of NodeStatus.
-type AvoidPods struct {
- // Bounded-sized list of signatures of pods that should avoid this node, sorted
- // in timestamp order from oldest to newest. Size of the slice is unspecified.
- // +optional
- PreferAvoidPods []PreferAvoidPodsEntry
-}
-
-// Describes a class of pods that should avoid this node.
-type PreferAvoidPodsEntry struct {
- // The class of pods.
- PodSignature PodSignature
- // Time at which this entry was added to the list.
- // +optional
- EvictionTime metav1.Time
- // (brief) reason why this entry was added to the list.
- // +optional
- Reason string
- // Human readable message indicating why this entry was added to the list.
- // +optional
- Message string
-}
-
-// Describes the class of pods that should avoid this node.
-// Exactly one field should be set.
-type PodSignature struct {
- // Reference to controller whose pods should avoid this node.
- // +optional
- PodController *metav1.OwnerReference
-}
-
-// Describe a container image
-type ContainerImage struct {
- // Names by which this image is known.
- Names []string
- // The size of the image in bytes.
- // +optional
- SizeBytes int64
-}
-
-type NodePhase string
-
-// These are the valid phases of node.
-const (
- // NodePending means the node has been created/added by the system, but not configured.
- NodePending NodePhase = "Pending"
- // NodeRunning means the node has been configured and has Kubernetes components running.
- NodeRunning NodePhase = "Running"
- // NodeTerminated means the node has been removed from the cluster.
- NodeTerminated NodePhase = "Terminated"
-)
-
-type NodeConditionType string
-
-// These are valid conditions of node. Currently, we don't have enough information to decide
-// node condition. In the future, we will add more. The proposed set of conditions are:
-// NodeReady, NodeReachable
-const (
- // NodeReady means kubelet is healthy and ready to accept pods.
- NodeReady NodeConditionType = "Ready"
- // NodeOutOfDisk means the kubelet will not accept new pods due to insufficient free disk
- // space on the node.
- NodeOutOfDisk NodeConditionType = "OutOfDisk"
- // NodeMemoryPressure means the kubelet is under pressure due to insufficient available memory.
- NodeMemoryPressure NodeConditionType = "MemoryPressure"
- // NodeDiskPressure means the kubelet is under pressure due to insufficient available disk.
- NodeDiskPressure NodeConditionType = "DiskPressure"
- // NodeNetworkUnavailable means that network for the node is not correctly configured.
- NodeNetworkUnavailable NodeConditionType = "NetworkUnavailable"
- // NodeKubeletConfigOk indicates whether the kubelet is correctly configured
- NodeKubeletConfigOk NodeConditionType = "KubeletConfigOk"
-)
-
-type NodeCondition struct {
- Type NodeConditionType
- Status ConditionStatus
- // +optional
- LastHeartbeatTime metav1.Time
- // +optional
- LastTransitionTime metav1.Time
- // +optional
- Reason string
- // +optional
- Message string
-}
-
-type NodeAddressType string
-
-const (
- NodeHostName NodeAddressType = "Hostname"
- NodeExternalIP NodeAddressType = "ExternalIP"
- NodeInternalIP NodeAddressType = "InternalIP"
- NodeExternalDNS NodeAddressType = "ExternalDNS"
- NodeInternalDNS NodeAddressType = "InternalDNS"
-)
-
-type NodeAddress struct {
- Type NodeAddressType
- Address string
-}
-
-// NodeResources is an object for conveying resource information about a node.
-// see http://releases.k8s.io/HEAD/docs/design/resources.md for more details.
-type NodeResources struct {
- // Capacity represents the available resources of a node
- // +optional
- Capacity ResourceList
-}
-
-// ResourceName is the name identifying various resources in a ResourceList.
-type ResourceName string
-
-// Resource names must be not more than 63 characters, consisting of upper- or lower-case alphanumeric characters,
-// with the -, _, and . characters allowed anywhere, except the first or last character.
-// The default convention, matching that for annotations, is to use lower-case names, with dashes, rather than
-// camel case, separating compound words.
-// Fully-qualified resource typenames are constructed from a DNS-style subdomain, followed by a slash `/` and a name.
-const (
- // CPU, in cores. (500m = .5 cores)
- ResourceCPU ResourceName = "cpu"
- // Memory, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- ResourceMemory ResourceName = "memory"
- // Volume size, in bytes (e,g. 5Gi = 5GiB = 5 * 1024 * 1024 * 1024)
- ResourceStorage ResourceName = "storage"
- // Local ephemeral storage, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- // The resource name for ResourceEphemeralStorage is alpha and it can change across releases.
- ResourceEphemeralStorage ResourceName = "ephemeral-storage"
- // NVIDIA GPU, in devices. Alpha, might change: although fractional and allowing values >1, only one whole device per node is assigned.
- ResourceNvidiaGPU ResourceName = "alpha.kubernetes.io/nvidia-gpu"
-)
-
-const (
- // Default namespace prefix.
- ResourceDefaultNamespacePrefix = "kubernetes.io/"
- // Name prefix for huge page resources (alpha).
- ResourceHugePagesPrefix = "hugepages-"
-)
-
-// ResourceList is a set of (resource name, quantity) pairs.
-type ResourceList map[ResourceName]resource.Quantity
-
-// +genclient
-// +genclient:nonNamespaced
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Node is a worker node in Kubernetes
-// The name of the node according to etcd is in ObjectMeta.Name.
-type Node struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the behavior of a node.
- // +optional
- Spec NodeSpec
-
- // Status describes the current status of a Node
- // +optional
- Status NodeStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NodeList is a list of nodes.
-type NodeList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Node
-}
-
-// NamespaceSpec describes the attributes on a Namespace
-type NamespaceSpec struct {
- // Finalizers is an opaque list of values that must be empty to permanently remove object from storage
- Finalizers []FinalizerName
-}
-
-// FinalizerName is the name identifying a finalizer during namespace lifecycle.
-type FinalizerName string
-
-// These are internal finalizer values to Kubernetes, must be qualified name unless defined here or
-// in metav1.
-const (
- FinalizerKubernetes FinalizerName = "kubernetes"
-)
-
-// NamespaceStatus is information about the current status of a Namespace.
-type NamespaceStatus struct {
- // Phase is the current lifecycle phase of the namespace.
- // +optional
- Phase NamespacePhase
-}
-
-type NamespacePhase string
-
-// These are the valid phases of a namespace.
-const (
- // NamespaceActive means the namespace is available for use in the system
- NamespaceActive NamespacePhase = "Active"
- // NamespaceTerminating means the namespace is undergoing graceful termination
- NamespaceTerminating NamespacePhase = "Terminating"
-)
-
-// +genclient
-// +genclient:nonNamespaced
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// A namespace provides a scope for Names.
-// Use of multiple namespaces is optional
-type Namespace struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the behavior of the Namespace.
- // +optional
- Spec NamespaceSpec
-
- // Status describes the current status of a Namespace
- // +optional
- Status NamespaceStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NamespaceList is a list of Namespaces.
-type NamespaceList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Namespace
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Binding ties one object to another; for example, a pod is bound to a node by a scheduler.
-// Deprecated in 1.7, please use the bindings subresource of pods instead.
-type Binding struct {
- metav1.TypeMeta
- // ObjectMeta describes the object that is being bound.
- // +optional
- metav1.ObjectMeta
-
- // Target is the object to bind to.
- Target ObjectReference
-}
-
-// Preconditions must be fulfilled before an operation (update, delete, etc.) is carried out.
-type Preconditions struct {
- // Specifies the target UID.
- // +optional
- UID *types.UID
-}
-
-// DeletionPropagation decides whether and how garbage collection will be performed.
-type DeletionPropagation string
-
-const (
- // Orphans the dependents.
- DeletePropagationOrphan DeletionPropagation = "Orphan"
- // Deletes the object from the key-value store, the garbage collector will delete the dependents in the background.
- DeletePropagationBackground DeletionPropagation = "Background"
- // The object exists in the key-value store until the garbage collector deletes all the dependents whose ownerReference.blockOwnerDeletion=true from the key-value store.
- // API sever will put the "DeletingDependents" finalizer on the object, and sets its deletionTimestamp.
- // This policy is cascading, i.e., the dependents will be deleted with Foreground.
- DeletePropagationForeground DeletionPropagation = "Foreground"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// DeleteOptions may be provided when deleting an API object
-// DEPRECATED: This type has been moved to meta/v1 and will be removed soon.
-type DeleteOptions struct {
- metav1.TypeMeta
-
- // Optional duration in seconds before the object should be deleted. Value must be non-negative integer.
- // The value zero indicates delete immediately. If this value is nil, the default grace period for the
- // specified type will be used.
- // +optional
- GracePeriodSeconds *int64
-
- // Must be fulfilled before a deletion is carried out. If not possible, a 409 Conflict status will be
- // returned.
- // +optional
- Preconditions *Preconditions
-
- // Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7.
- // Should the dependent objects be orphaned. If true/false, the "orphan"
- // finalizer will be added to/removed from the object's finalizers list.
- // Either this field or PropagationPolicy may be set, but not both.
- // +optional
- OrphanDependents *bool
-
- // Whether and how garbage collection will be performed.
- // Either this field or OrphanDependents may be set, but not both.
- // The default policy is decided by the existing finalizer set in the
- // metadata.finalizers and the resource-specific default policy.
- // Acceptable values are: 'Orphan' - orphan the dependents; 'Background' -
- // allow the garbage collector to delete the dependents in the background;
- // 'Foreground' - a cascading policy that deletes all dependents in the
- // foreground.
- // +optional
- PropagationPolicy *DeletionPropagation
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ListOptions is the query options to a standard REST list call, and has future support for
-// watch calls.
-// DEPRECATED: This type has been moved to meta/v1 and will be removed soon.
-type ListOptions struct {
- metav1.TypeMeta
-
- // A selector based on labels
- LabelSelector labels.Selector
- // A selector based on fields
- FieldSelector fields.Selector
-
- // If true, partially initialized resources are included in the response.
- IncludeUninitialized bool
-
- // If true, watch for changes to this list
- Watch bool
- // When specified with a watch call, shows changes that occur after that particular version of a resource.
- // Defaults to changes from the beginning of history.
- // When specified for list:
- // - if unset, then the result is returned from remote storage based on quorum-read flag;
- // - if it's 0, then we simply return what we currently have in cache, no guarantee;
- // - if set to non zero, then the result is at least as fresh as given rv.
- ResourceVersion string
- // Timeout for the list/watch call.
- TimeoutSeconds *int64
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodLogOptions is the query options for a Pod's logs REST call
-type PodLogOptions struct {
- metav1.TypeMeta
-
- // Container for which to return logs
- Container string
- // If true, follow the logs for the pod
- Follow bool
- // If true, return previous terminated container logs
- Previous bool
- // A relative time in seconds before the current time from which to show logs. If this value
- // precedes the time a pod was started, only logs since the pod start will be returned.
- // If this value is in the future, no logs will be returned.
- // Only one of sinceSeconds or sinceTime may be specified.
- SinceSeconds *int64
- // An RFC3339 timestamp from which to show logs. If this value
- // precedes the time a pod was started, only logs since the pod start will be returned.
- // If this value is in the future, no logs will be returned.
- // Only one of sinceSeconds or sinceTime may be specified.
- SinceTime *metav1.Time
- // If true, add an RFC3339 or RFC3339Nano timestamp at the beginning of every line
- // of log output.
- Timestamps bool
- // If set, the number of lines from the end of the logs to show. If not specified,
- // logs are shown from the creation of the container or sinceSeconds or sinceTime
- TailLines *int64
- // If set, the number of bytes to read from the server before terminating the
- // log output. This may not display a complete final line of logging, and may return
- // slightly more or slightly less than the specified limit.
- LimitBytes *int64
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodAttachOptions is the query options to a Pod's remote attach call
-// TODO: merge w/ PodExecOptions below for stdin, stdout, etc
-type PodAttachOptions struct {
- metav1.TypeMeta
-
- // Stdin if true indicates that stdin is to be redirected for the attach call
- // +optional
- Stdin bool
-
- // Stdout if true indicates that stdout is to be redirected for the attach call
- // +optional
- Stdout bool
-
- // Stderr if true indicates that stderr is to be redirected for the attach call
- // +optional
- Stderr bool
-
- // TTY if true indicates that a tty will be allocated for the attach call
- // +optional
- TTY bool
-
- // Container to attach to.
- // +optional
- Container string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodExecOptions is the query options to a Pod's remote exec call
-type PodExecOptions struct {
- metav1.TypeMeta
-
- // Stdin if true indicates that stdin is to be redirected for the exec call
- Stdin bool
-
- // Stdout if true indicates that stdout is to be redirected for the exec call
- Stdout bool
-
- // Stderr if true indicates that stderr is to be redirected for the exec call
- Stderr bool
-
- // TTY if true indicates that a tty will be allocated for the exec call
- TTY bool
-
- // Container in which to execute the command.
- Container string
-
- // Command is the remote command to execute; argv array; not executed within a shell.
- Command []string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodPortForwardOptions is the query options to a Pod's port forward call
-type PodPortForwardOptions struct {
- metav1.TypeMeta
-
- // The list of ports to forward
- // +optional
- Ports []int32
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodProxyOptions is the query options to a Pod's proxy call
-type PodProxyOptions struct {
- metav1.TypeMeta
-
- // Path is the URL path to use for the current proxy request
- Path string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NodeProxyOptions is the query options to a Node's proxy call
-type NodeProxyOptions struct {
- metav1.TypeMeta
-
- // Path is the URL path to use for the current proxy request
- Path string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ServiceProxyOptions is the query options to a Service's proxy call.
-type ServiceProxyOptions struct {
- metav1.TypeMeta
-
- // Path is the part of URLs that include service endpoints, suffixes,
- // and parameters to use for the current proxy request to service.
- // For example, the whole request URL is
- // http://localhost/api/v1/namespaces/kube-system/services/elasticsearch-logging/_search?q=user:kimchy.
- // Path is _search?q=user:kimchy.
- Path string
-}
-
-// ObjectReference contains enough information to let you inspect or modify the referred object.
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-type ObjectReference struct {
- // +optional
- Kind string
- // +optional
- Namespace string
- // +optional
- Name string
- // +optional
- UID types.UID
- // +optional
- APIVersion string
- // +optional
- ResourceVersion string
-
- // Optional. If referring to a piece of an object instead of an entire object, this string
- // should contain information to identify the sub-object. For example, if the object
- // reference is to a container within a pod, this would take on a value like:
- // "spec.containers{name}" (where "name" refers to the name of the container that triggered
- // the event) or if no container name is specified "spec.containers[2]" (container with
- // index 2 in this pod). This syntax is chosen only to have some well-defined way of
- // referencing a part of an object.
- // TODO: this design is not final and this field is subject to change in the future.
- // +optional
- FieldPath string
-}
-
-// LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.
-type LocalObjectReference struct {
- //TODO: Add other useful fields. apiVersion, kind, uid?
- Name string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type SerializedReference struct {
- metav1.TypeMeta
- // +optional
- Reference ObjectReference
-}
-
-type EventSource struct {
- // Component from which the event is generated.
- // +optional
- Component string
- // Node name on which the event is generated.
- // +optional
- Host string
-}
-
-// Valid values for event types (new types could be added in future)
-const (
- // Information only and will not cause any problems
- EventTypeNormal string = "Normal"
- // These events are to warn that something might go wrong
- EventTypeWarning string = "Warning"
-)
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Event is a report of an event somewhere in the cluster.
-// TODO: Decide whether to store these separately or with the object they apply to.
-type Event struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Required. The object that this event is about. Mapped to events.Event.regarding
- // +optional
- InvolvedObject ObjectReference
-
- // Optional; this should be a short, machine understandable string that gives the reason
- // for this event being generated. For example, if the event is reporting that a container
- // can't start, the Reason might be "ImageNotFound".
- // TODO: provide exact specification for format.
- // +optional
- Reason string
-
- // Optional. A human-readable description of the status of this operation.
- // TODO: decide on maximum length. Mapped to events.Event.note
- // +optional
- Message string
-
- // Optional. The component reporting this event. Should be a short machine understandable string.
- // +optional
- Source EventSource
-
- // The time at which the event was first recorded. (Time of server receipt is in TypeMeta.)
- // +optional
- FirstTimestamp metav1.Time
-
- // The time at which the most recent occurrence of this event was recorded.
- // +optional
- LastTimestamp metav1.Time
-
- // The number of times this event has occurred.
- // +optional
- Count int32
-
- // Type of this event (Normal, Warning), new types could be added in the future.
- // +optional
- Type string
-
- // Time when this Event was first observed.
- // +optional
- EventTime metav1.MicroTime
-
- // Data about the Event series this event represents or nil if it's a singleton Event.
- // +optional
- Series *EventSeries
-
- // What action was taken/failed regarding to the Regarding object.
- // +optional
- Action string
-
- // Optional secondary object for more complex actions.
- // +optional
- Related *ObjectReference
-
- // Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`.
- // +optional
- ReportingController string
-
- // ID of the controller instance, e.g. `kubelet-xyzf`.
- // +optional
- ReportingInstance string
-}
-
-type EventSeries struct {
- // Number of occurrences in this series up to the last heartbeat time
- Count int32
- // Time of the last occurrence observed
- LastObservedTime metav1.MicroTime
- // State of this Series: Ongoing or Finished
- State EventSeriesState
-}
-
-type EventSeriesState string
-
-const (
- EventSeriesStateOngoing EventSeriesState = "Ongoing"
- EventSeriesStateFinished EventSeriesState = "Finished"
- EventSeriesStateUnknown EventSeriesState = "Unknown"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// EventList is a list of events.
-type EventList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Event
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// List holds a list of objects, which may not be known by the server.
-type List metainternalversion.List
-
-// A type of object that is limited
-type LimitType string
-
-const (
- // Limit that applies to all pods in a namespace
- LimitTypePod LimitType = "Pod"
- // Limit that applies to all containers in a namespace
- LimitTypeContainer LimitType = "Container"
- // Limit that applies to all persistent volume claims in a namespace
- LimitTypePersistentVolumeClaim LimitType = "PersistentVolumeClaim"
-)
-
-// LimitRangeItem defines a min/max usage limit for any resource that matches on kind
-type LimitRangeItem struct {
- // Type of resource that this limit applies to
- // +optional
- Type LimitType
- // Max usage constraints on this kind by resource name
- // +optional
- Max ResourceList
- // Min usage constraints on this kind by resource name
- // +optional
- Min ResourceList
- // Default resource requirement limit value by resource name.
- // +optional
- Default ResourceList
- // DefaultRequest resource requirement request value by resource name.
- // +optional
- DefaultRequest ResourceList
- // MaxLimitRequestRatio represents the max burst value for the named resource
- // +optional
- MaxLimitRequestRatio ResourceList
-}
-
-// LimitRangeSpec defines a min/max usage limit for resources that match on kind
-type LimitRangeSpec struct {
- // Limits is the list of LimitRangeItem objects that are enforced
- Limits []LimitRangeItem
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// LimitRange sets resource usage limits for each kind of resource in a Namespace
-type LimitRange struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the limits enforced
- // +optional
- Spec LimitRangeSpec
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// LimitRangeList is a list of LimitRange items.
-type LimitRangeList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- // Items is a list of LimitRange objects
- Items []LimitRange
-}
-
-// The following identify resource constants for Kubernetes object types
-const (
- // Pods, number
- ResourcePods ResourceName = "pods"
- // Services, number
- ResourceServices ResourceName = "services"
- // ReplicationControllers, number
- ResourceReplicationControllers ResourceName = "replicationcontrollers"
- // ResourceQuotas, number
- ResourceQuotas ResourceName = "resourcequotas"
- // ResourceSecrets, number
- ResourceSecrets ResourceName = "secrets"
- // ResourceConfigMaps, number
- ResourceConfigMaps ResourceName = "configmaps"
- // ResourcePersistentVolumeClaims, number
- ResourcePersistentVolumeClaims ResourceName = "persistentvolumeclaims"
- // ResourceServicesNodePorts, number
- ResourceServicesNodePorts ResourceName = "services.nodeports"
- // ResourceServicesLoadBalancers, number
- ResourceServicesLoadBalancers ResourceName = "services.loadbalancers"
- // CPU request, in cores. (500m = .5 cores)
- ResourceRequestsCPU ResourceName = "requests.cpu"
- // Memory request, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- ResourceRequestsMemory ResourceName = "requests.memory"
- // Storage request, in bytes
- ResourceRequestsStorage ResourceName = "requests.storage"
- // Local ephemeral storage request, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- ResourceRequestsEphemeralStorage ResourceName = "requests.ephemeral-storage"
- // CPU limit, in cores. (500m = .5 cores)
- ResourceLimitsCPU ResourceName = "limits.cpu"
- // Memory limit, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- ResourceLimitsMemory ResourceName = "limits.memory"
- // Local ephemeral storage limit, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- ResourceLimitsEphemeralStorage ResourceName = "limits.ephemeral-storage"
-)
-
-// The following identify resource prefix for Kubernetes object types
-const (
- // HugePages request, in bytes. (500Gi = 500GiB = 500 * 1024 * 1024 * 1024)
- // As burst is not supported for HugePages, we would only quota its request, and ignore the limit.
- ResourceRequestsHugePagesPrefix = "requests.hugepages-"
- // Default resource requests prefix
- DefaultResourceRequestsPrefix = "requests."
-)
-
-// A ResourceQuotaScope defines a filter that must match each object tracked by a quota
-type ResourceQuotaScope string
-
-const (
- // Match all pod objects where spec.activeDeadlineSeconds
- ResourceQuotaScopeTerminating ResourceQuotaScope = "Terminating"
- // Match all pod objects where !spec.activeDeadlineSeconds
- ResourceQuotaScopeNotTerminating ResourceQuotaScope = "NotTerminating"
- // Match all pod objects that have best effort quality of service
- ResourceQuotaScopeBestEffort ResourceQuotaScope = "BestEffort"
- // Match all pod objects that do not have best effort quality of service
- ResourceQuotaScopeNotBestEffort ResourceQuotaScope = "NotBestEffort"
-)
-
-// ResourceQuotaSpec defines the desired hard limits to enforce for Quota
-type ResourceQuotaSpec struct {
- // Hard is the set of desired hard limits for each named resource
- // +optional
- Hard ResourceList
- // A collection of filters that must match each object tracked by a quota.
- // If not specified, the quota matches all objects.
- // +optional
- Scopes []ResourceQuotaScope
-}
-
-// ResourceQuotaStatus defines the enforced hard limits and observed use
-type ResourceQuotaStatus struct {
- // Hard is the set of enforced hard limits for each named resource
- // +optional
- Hard ResourceList
- // Used is the current observed total usage of the resource in the namespace
- // +optional
- Used ResourceList
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ResourceQuota sets aggregate quota restrictions enforced per namespace
-type ResourceQuota struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the desired quota
- // +optional
- Spec ResourceQuotaSpec
-
- // Status defines the actual enforced quota and its current usage
- // +optional
- Status ResourceQuotaStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ResourceQuotaList is a list of ResourceQuota items
-type ResourceQuotaList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- // Items is a list of ResourceQuota objects
- Items []ResourceQuota
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Secret holds secret data of a certain type. The total bytes of the values in
-// the Data field must be less than MaxSecretSize bytes.
-type Secret struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Data contains the secret data. Each key must consist of alphanumeric
- // characters, '-', '_' or '.'. The serialized form of the secret data is a
- // base64 encoded string, representing the arbitrary (possibly non-string)
- // data value here.
- // +optional
- Data map[string][]byte
-
- // Used to facilitate programmatic handling of secret data.
- // +optional
- Type SecretType
-}
-
-const MaxSecretSize = 1 * 1024 * 1024
-
-type SecretType string
-
-const (
- // SecretTypeOpaque is the default; arbitrary user-defined data
- SecretTypeOpaque SecretType = "Opaque"
-
- // SecretTypeServiceAccountToken contains a token that identifies a service account to the API
- //
- // Required fields:
- // - Secret.Annotations["kubernetes.io/service-account.name"] - the name of the ServiceAccount the token identifies
- // - Secret.Annotations["kubernetes.io/service-account.uid"] - the UID of the ServiceAccount the token identifies
- // - Secret.Data["token"] - a token that identifies the service account to the API
- SecretTypeServiceAccountToken SecretType = "kubernetes.io/service-account-token"
-
- // ServiceAccountNameKey is the key of the required annotation for SecretTypeServiceAccountToken secrets
- ServiceAccountNameKey = "kubernetes.io/service-account.name"
- // ServiceAccountUIDKey is the key of the required annotation for SecretTypeServiceAccountToken secrets
- ServiceAccountUIDKey = "kubernetes.io/service-account.uid"
- // ServiceAccountTokenKey is the key of the required data for SecretTypeServiceAccountToken secrets
- ServiceAccountTokenKey = "token"
- // ServiceAccountKubeconfigKey is the key of the optional kubeconfig data for SecretTypeServiceAccountToken secrets
- ServiceAccountKubeconfigKey = "kubernetes.kubeconfig"
- // ServiceAccountRootCAKey is the key of the optional root certificate authority for SecretTypeServiceAccountToken secrets
- ServiceAccountRootCAKey = "ca.crt"
- // ServiceAccountNamespaceKey is the key of the optional namespace to use as the default for namespaced API calls
- ServiceAccountNamespaceKey = "namespace"
-
- // SecretTypeDockercfg contains a dockercfg file that follows the same format rules as ~/.dockercfg
- //
- // Required fields:
- // - Secret.Data[".dockercfg"] - a serialized ~/.dockercfg file
- SecretTypeDockercfg SecretType = "kubernetes.io/dockercfg"
-
- // DockerConfigKey is the key of the required data for SecretTypeDockercfg secrets
- DockerConfigKey = ".dockercfg"
-
- // SecretTypeDockerConfigJson contains a dockercfg file that follows the same format rules as ~/.docker/config.json
- //
- // Required fields:
- // - Secret.Data[".dockerconfigjson"] - a serialized ~/.docker/config.json file
- SecretTypeDockerConfigJson SecretType = "kubernetes.io/dockerconfigjson"
-
- // DockerConfigJsonKey is the key of the required data for SecretTypeDockerConfigJson secrets
- DockerConfigJsonKey = ".dockerconfigjson"
-
- // SecretTypeBasicAuth contains data needed for basic authentication.
- //
- // Required at least one of fields:
- // - Secret.Data["username"] - username used for authentication
- // - Secret.Data["password"] - password or token needed for authentication
- SecretTypeBasicAuth SecretType = "kubernetes.io/basic-auth"
-
- // BasicAuthUsernameKey is the key of the username for SecretTypeBasicAuth secrets
- BasicAuthUsernameKey = "username"
- // BasicAuthPasswordKey is the key of the password or token for SecretTypeBasicAuth secrets
- BasicAuthPasswordKey = "password"
-
- // SecretTypeSSHAuth contains data needed for SSH authentication.
- //
- // Required field:
- // - Secret.Data["ssh-privatekey"] - private SSH key needed for authentication
- SecretTypeSSHAuth SecretType = "kubernetes.io/ssh-auth"
-
- // SSHAuthPrivateKey is the key of the required SSH private key for SecretTypeSSHAuth secrets
- SSHAuthPrivateKey = "ssh-privatekey"
-
- // SecretTypeTLS contains information about a TLS client or server secret. It
- // is primarily used with TLS termination of the Ingress resource, but may be
- // used in other types.
- //
- // Required fields:
- // - Secret.Data["tls.key"] - TLS private key.
- // Secret.Data["tls.crt"] - TLS certificate.
- // TODO: Consider supporting different formats, specifying CA/destinationCA.
- SecretTypeTLS SecretType = "kubernetes.io/tls"
-
- // TLSCertKey is the key for tls certificates in a TLS secret.
- TLSCertKey = "tls.crt"
- // TLSPrivateKeyKey is the key for the private key field in a TLS secret.
- TLSPrivateKeyKey = "tls.key"
- // SecretTypeBootstrapToken is used during the automated bootstrap process (first
- // implemented by kubeadm). It stores tokens that are used to sign well known
- // ConfigMaps. They are used for authn.
- SecretTypeBootstrapToken SecretType = "bootstrap.kubernetes.io/token"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type SecretList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []Secret
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ConfigMap holds configuration data for components or applications to consume.
-type ConfigMap struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Data contains the configuration data.
- // Each key must consist of alphanumeric characters, '-', '_' or '.'.
- // Values with non-UTF-8 byte sequences must use the BinaryData field.
- // The keys stored in Data must not overlap with the keys in
- // the BinaryData field, this is enforced during validation process.
- // +optional
- Data map[string]string
-
- // BinaryData contains the binary data.
- // Each key must consist of alphanumeric characters, '-', '_' or '.'.
- // BinaryData can contain byte sequences that are not in the UTF-8 range.
- // The keys stored in BinaryData must not overlap with the ones in
- // the Data field, this is enforced during validation process.
- // Using this field will require 1.10+ apiserver and
- // kubelet.
- // +optional
- BinaryData map[string][]byte
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ConfigMapList is a resource containing a list of ConfigMap objects.
-type ConfigMapList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- // Items is the list of ConfigMaps.
- Items []ConfigMap
-}
-
-// These constants are for remote command execution and port forwarding and are
-// used by both the client side and server side components.
-//
-// This is probably not the ideal place for them, but it didn't seem worth it
-// to create pkg/exec and pkg/portforward just to contain a single file with
-// constants in it. Suggestions for more appropriate alternatives are
-// definitely welcome!
-const (
- // Enable stdin for remote command execution
- ExecStdinParam = "input"
- // Enable stdout for remote command execution
- ExecStdoutParam = "output"
- // Enable stderr for remote command execution
- ExecStderrParam = "error"
- // Enable TTY for remote command execution
- ExecTTYParam = "tty"
- // Command to run for remote command execution
- ExecCommandParam = "command"
-
- // Name of header that specifies stream type
- StreamType = "streamType"
- // Value for streamType header for stdin stream
- StreamTypeStdin = "stdin"
- // Value for streamType header for stdout stream
- StreamTypeStdout = "stdout"
- // Value for streamType header for stderr stream
- StreamTypeStderr = "stderr"
- // Value for streamType header for data stream
- StreamTypeData = "data"
- // Value for streamType header for error stream
- StreamTypeError = "error"
- // Value for streamType header for terminal resize stream
- StreamTypeResize = "resize"
-
- // Name of header that specifies the port being forwarded
- PortHeader = "port"
- // Name of header that specifies a request ID used to associate the error
- // and data streams for a single forwarded connection
- PortForwardRequestIDHeader = "requestID"
-)
-
-// Type and constants for component health validation.
-type ComponentConditionType string
-
-// These are the valid conditions for the component.
-const (
- ComponentHealthy ComponentConditionType = "Healthy"
-)
-
-type ComponentCondition struct {
- Type ComponentConditionType
- Status ConditionStatus
- // +optional
- Message string
- // +optional
- Error string
-}
-
-// +genclient
-// +genclient:nonNamespaced
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ComponentStatus (and ComponentStatusList) holds the cluster validation info.
-type ComponentStatus struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // +optional
- Conditions []ComponentCondition
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type ComponentStatusList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []ComponentStatus
-}
-
-// SecurityContext holds security configuration that will be applied to a container.
-// Some fields are present in both SecurityContext and PodSecurityContext. When both
-// are set, the values in SecurityContext take precedence.
-type SecurityContext struct {
- // The capabilities to add/drop when running containers.
- // Defaults to the default set of capabilities granted by the container runtime.
- // +optional
- Capabilities *Capabilities
- // Run container in privileged mode.
- // Processes in privileged containers are essentially equivalent to root on the host.
- // Defaults to false.
- // +optional
- Privileged *bool
- // The SELinux context to be applied to the container.
- // If unspecified, the container runtime will allocate a random SELinux context for each
- // container. May also be set in PodSecurityContext. If set in both SecurityContext and
- // PodSecurityContext, the value specified in SecurityContext takes precedence.
- // +optional
- SELinuxOptions *SELinuxOptions
- // The UID to run the entrypoint of the container process.
- // Defaults to user specified in image metadata if unspecified.
- // May also be set in PodSecurityContext. If set in both SecurityContext and
- // PodSecurityContext, the value specified in SecurityContext takes precedence.
- // +optional
- RunAsUser *int64
- // Indicates that the container must run as a non-root user.
- // If true, the Kubelet will validate the image at runtime to ensure that it
- // does not run as UID 0 (root) and fail to start the container if it does.
- // If unset or false, no such validation will be performed.
- // May also be set in PodSecurityContext. If set in both SecurityContext and
- // PodSecurityContext, the value specified in SecurityContext takes precedence.
- // +optional
- RunAsNonRoot *bool
- // The read-only root filesystem allows you to restrict the locations that an application can write
- // files to, ensuring the persistent data can only be written to mounts.
- // +optional
- ReadOnlyRootFilesystem *bool
- // AllowPrivilegeEscalation controls whether a process can gain more
- // privileges than its parent process. This bool directly controls if
- // the no_new_privs flag will be set on the container process.
- // +optional
- AllowPrivilegeEscalation *bool
-}
-
-// SELinuxOptions are the labels to be applied to the container.
-type SELinuxOptions struct {
- // SELinux user label
- // +optional
- User string
- // SELinux role label
- // +optional
- Role string
- // SELinux type label
- // +optional
- Type string
- // SELinux level label.
- // +optional
- Level string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// RangeAllocation is an opaque API object (not exposed to end users) that can be persisted to record
-// the global allocation state of the cluster. The schema of Range and Data generic, in that Range
-// should be a string representation of the inputs to a range (for instance, for IP allocation it
-// might be a CIDR) and Data is an opaque blob understood by an allocator which is typically a
-// binary range. Consumers should use annotations to record additional information (schema version,
-// data encoding hints). A range allocation should *ALWAYS* be recreatable at any time by observation
-// of the cluster, thus the object is less strongly typed than most.
-type RangeAllocation struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
- // A string representing a unique label for a range of resources, such as a CIDR "10.0.0.0/8" or
- // port range "10000-30000". Range is not strongly schema'd here. The Range is expected to define
- // a start and end unless there is an implicit end.
- Range string
- // A byte array representing the serialized state of a range allocation. Additional clarifiers on
- // the type or format of data should be represented with annotations. For IP allocations, this is
- // represented as a bit array starting at the base IP of the CIDR in Range, with each bit representing
- // a single allocated address (the fifth bit on CIDR 10.0.0.0/8 is 10.0.0.4).
- Data []byte
-}
-
-const (
- // "default-scheduler" is the name of default scheduler.
- DefaultSchedulerName = "default-scheduler"
-
- // RequiredDuringScheduling affinity is not symmetric, but there is an implicit PreferredDuringScheduling affinity rule
- // corresponding to every RequiredDuringScheduling affinity rule.
- // When the --hard-pod-affinity-weight scheduler flag is not specified,
- // DefaultHardPodAffinityWeight defines the weight of the implicit PreferredDuringScheduling affinity rule.
- DefaultHardPodAffinitySymmetricWeight int32 = 1
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/conversion.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/conversion.go
deleted file mode 100644
index d888ebca4..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/conversion.go
+++ /dev/null
@@ -1,619 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package v1
-
-import (
- "fmt"
- "reflect"
-
- "k8s.io/api/core/v1"
-
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/conversion"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/util/validation/field"
- "k8s.io/kubernetes/pkg/apis/core"
- "k8s.io/kubernetes/pkg/apis/extensions"
-)
-
-// This is a "fast-path" that avoids reflection for common types. It focuses on the objects that are
-// converted the most in the cluster.
-// TODO: generate one of these for every external API group - this is to prove the impact
-func addFastPathConversionFuncs(scheme *runtime.Scheme) error {
- scheme.AddGenericConversionFunc(func(objA, objB interface{}, s conversion.Scope) (bool, error) {
- switch a := objA.(type) {
- case *v1.Pod:
- switch b := objB.(type) {
- case *core.Pod:
- return true, Convert_v1_Pod_To_core_Pod(a, b, s)
- }
- case *core.Pod:
- switch b := objB.(type) {
- case *v1.Pod:
- return true, Convert_core_Pod_To_v1_Pod(a, b, s)
- }
-
- case *v1.Event:
- switch b := objB.(type) {
- case *core.Event:
- return true, Convert_v1_Event_To_core_Event(a, b, s)
- }
- case *core.Event:
- switch b := objB.(type) {
- case *v1.Event:
- return true, Convert_core_Event_To_v1_Event(a, b, s)
- }
-
- case *v1.ReplicationController:
- switch b := objB.(type) {
- case *core.ReplicationController:
- return true, Convert_v1_ReplicationController_To_core_ReplicationController(a, b, s)
- }
- case *core.ReplicationController:
- switch b := objB.(type) {
- case *v1.ReplicationController:
- return true, Convert_core_ReplicationController_To_v1_ReplicationController(a, b, s)
- }
-
- case *v1.Node:
- switch b := objB.(type) {
- case *core.Node:
- return true, Convert_v1_Node_To_core_Node(a, b, s)
- }
- case *core.Node:
- switch b := objB.(type) {
- case *v1.Node:
- return true, Convert_core_Node_To_v1_Node(a, b, s)
- }
-
- case *v1.Namespace:
- switch b := objB.(type) {
- case *core.Namespace:
- return true, Convert_v1_Namespace_To_core_Namespace(a, b, s)
- }
- case *core.Namespace:
- switch b := objB.(type) {
- case *v1.Namespace:
- return true, Convert_core_Namespace_To_v1_Namespace(a, b, s)
- }
-
- case *v1.Service:
- switch b := objB.(type) {
- case *core.Service:
- return true, Convert_v1_Service_To_core_Service(a, b, s)
- }
- case *core.Service:
- switch b := objB.(type) {
- case *v1.Service:
- return true, Convert_core_Service_To_v1_Service(a, b, s)
- }
-
- case *v1.Endpoints:
- switch b := objB.(type) {
- case *core.Endpoints:
- return true, Convert_v1_Endpoints_To_core_Endpoints(a, b, s)
- }
- case *core.Endpoints:
- switch b := objB.(type) {
- case *v1.Endpoints:
- return true, Convert_core_Endpoints_To_v1_Endpoints(a, b, s)
- }
-
- case *metav1.WatchEvent:
- switch b := objB.(type) {
- case *metav1.InternalEvent:
- return true, metav1.Convert_versioned_Event_to_versioned_InternalEvent(a, b, s)
- }
- case *metav1.InternalEvent:
- switch b := objB.(type) {
- case *metav1.WatchEvent:
- return true, metav1.Convert_versioned_InternalEvent_to_versioned_Event(a, b, s)
- }
- }
- return false, nil
- })
- return nil
-}
-
-func addConversionFuncs(scheme *runtime.Scheme) error {
- // Add non-generated conversion functions
- err := scheme.AddConversionFuncs(
- Convert_core_Pod_To_v1_Pod,
- Convert_core_PodSpec_To_v1_PodSpec,
- Convert_core_ReplicationControllerSpec_To_v1_ReplicationControllerSpec,
- Convert_core_ServiceSpec_To_v1_ServiceSpec,
- Convert_v1_Pod_To_core_Pod,
- Convert_v1_PodSpec_To_core_PodSpec,
- Convert_v1_ReplicationControllerSpec_To_core_ReplicationControllerSpec,
- Convert_v1_Secret_To_core_Secret,
- Convert_v1_ServiceSpec_To_core_ServiceSpec,
- Convert_v1_ResourceList_To_core_ResourceList,
- Convert_v1_ReplicationController_to_extensions_ReplicaSet,
- Convert_v1_ReplicationControllerSpec_to_extensions_ReplicaSetSpec,
- Convert_v1_ReplicationControllerStatus_to_extensions_ReplicaSetStatus,
- Convert_extensions_ReplicaSet_to_v1_ReplicationController,
- Convert_extensions_ReplicaSetSpec_to_v1_ReplicationControllerSpec,
- Convert_extensions_ReplicaSetStatus_to_v1_ReplicationControllerStatus,
- )
- if err != nil {
- return err
- }
-
- // Add field conversion funcs.
- err = scheme.AddFieldLabelConversionFunc("v1", "Pod",
- func(label, value string) (string, string, error) {
- switch label {
- case "metadata.name",
- "metadata.namespace",
- "spec.nodeName",
- "spec.restartPolicy",
- "spec.schedulerName",
- "status.phase",
- "status.podIP",
- "status.nominatedNodeName":
- return label, value, nil
- // This is for backwards compatibility with old v1 clients which send spec.host
- case "spec.host":
- return "spec.nodeName", value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- },
- )
- if err != nil {
- return err
- }
- err = scheme.AddFieldLabelConversionFunc("v1", "Node",
- func(label, value string) (string, string, error) {
- switch label {
- case "metadata.name":
- return label, value, nil
- case "spec.unschedulable":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- },
- )
- if err != nil {
- return err
- }
- err = scheme.AddFieldLabelConversionFunc("v1", "ReplicationController",
- func(label, value string) (string, string, error) {
- switch label {
- case "metadata.name",
- "metadata.namespace",
- "status.replicas":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- })
- if err != nil {
- return err
- }
- if err := AddFieldLabelConversionsForEvent(scheme); err != nil {
- return err
- }
- if err := AddFieldLabelConversionsForNamespace(scheme); err != nil {
- return err
- }
- if err := AddFieldLabelConversionsForSecret(scheme); err != nil {
- return err
- }
- return nil
-}
-
-func Convert_v1_ReplicationController_to_extensions_ReplicaSet(in *v1.ReplicationController, out *extensions.ReplicaSet, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ReplicationControllerSpec_to_extensions_ReplicaSetSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_ReplicationControllerStatus_to_extensions_ReplicaSetStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-func Convert_v1_ReplicationControllerSpec_to_extensions_ReplicaSetSpec(in *v1.ReplicationControllerSpec, out *extensions.ReplicaSetSpec, s conversion.Scope) error {
- out.Replicas = *in.Replicas
- out.MinReadySeconds = in.MinReadySeconds
- if in.Selector != nil {
- out.Selector = new(metav1.LabelSelector)
- metav1.Convert_map_to_unversioned_LabelSelector(&in.Selector, out.Selector, s)
- }
- if in.Template != nil {
- if err := Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec(in.Template, &out.Template, s); err != nil {
- return err
- }
- }
- return nil
-}
-
-func Convert_v1_ReplicationControllerStatus_to_extensions_ReplicaSetStatus(in *v1.ReplicationControllerStatus, out *extensions.ReplicaSetStatus, s conversion.Scope) error {
- out.Replicas = in.Replicas
- out.FullyLabeledReplicas = in.FullyLabeledReplicas
- out.ReadyReplicas = in.ReadyReplicas
- out.AvailableReplicas = in.AvailableReplicas
- out.ObservedGeneration = in.ObservedGeneration
- for _, cond := range in.Conditions {
- out.Conditions = append(out.Conditions, extensions.ReplicaSetCondition{
- Type: extensions.ReplicaSetConditionType(cond.Type),
- Status: core.ConditionStatus(cond.Status),
- LastTransitionTime: cond.LastTransitionTime,
- Reason: cond.Reason,
- Message: cond.Message,
- })
- }
- return nil
-}
-
-func Convert_extensions_ReplicaSet_to_v1_ReplicationController(in *extensions.ReplicaSet, out *v1.ReplicationController, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_extensions_ReplicaSetSpec_to_v1_ReplicationControllerSpec(&in.Spec, &out.Spec, s); err != nil {
- fieldErr, ok := err.(*field.Error)
- if !ok {
- return err
- }
- if out.Annotations == nil {
- out.Annotations = make(map[string]string)
- }
- out.Annotations[v1.NonConvertibleAnnotationPrefix+"/"+fieldErr.Field] = reflect.ValueOf(fieldErr.BadValue).String()
- }
- if err := Convert_extensions_ReplicaSetStatus_to_v1_ReplicationControllerStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-func Convert_extensions_ReplicaSetSpec_to_v1_ReplicationControllerSpec(in *extensions.ReplicaSetSpec, out *v1.ReplicationControllerSpec, s conversion.Scope) error {
- out.Replicas = new(int32)
- *out.Replicas = in.Replicas
- out.MinReadySeconds = in.MinReadySeconds
- var invalidErr error
- if in.Selector != nil {
- invalidErr = metav1.Convert_unversioned_LabelSelector_to_map(in.Selector, &out.Selector, s)
- }
- out.Template = new(v1.PodTemplateSpec)
- if err := Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec(&in.Template, out.Template, s); err != nil {
- return err
- }
- return invalidErr
-}
-
-func Convert_extensions_ReplicaSetStatus_to_v1_ReplicationControllerStatus(in *extensions.ReplicaSetStatus, out *v1.ReplicationControllerStatus, s conversion.Scope) error {
- out.Replicas = in.Replicas
- out.FullyLabeledReplicas = in.FullyLabeledReplicas
- out.ReadyReplicas = in.ReadyReplicas
- out.AvailableReplicas = in.AvailableReplicas
- out.ObservedGeneration = in.ObservedGeneration
- for _, cond := range in.Conditions {
- out.Conditions = append(out.Conditions, v1.ReplicationControllerCondition{
- Type: v1.ReplicationControllerConditionType(cond.Type),
- Status: v1.ConditionStatus(cond.Status),
- LastTransitionTime: cond.LastTransitionTime,
- Reason: cond.Reason,
- Message: cond.Message,
- })
- }
- return nil
-}
-
-func Convert_core_ReplicationControllerSpec_To_v1_ReplicationControllerSpec(in *core.ReplicationControllerSpec, out *v1.ReplicationControllerSpec, s conversion.Scope) error {
- out.Replicas = &in.Replicas
- out.MinReadySeconds = in.MinReadySeconds
- out.Selector = in.Selector
- if in.Template != nil {
- out.Template = new(v1.PodTemplateSpec)
- if err := Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec(in.Template, out.Template, s); err != nil {
- return err
- }
- } else {
- out.Template = nil
- }
- return nil
-}
-
-func Convert_v1_ReplicationControllerSpec_To_core_ReplicationControllerSpec(in *v1.ReplicationControllerSpec, out *core.ReplicationControllerSpec, s conversion.Scope) error {
- if in.Replicas != nil {
- out.Replicas = *in.Replicas
- }
- out.MinReadySeconds = in.MinReadySeconds
- out.Selector = in.Selector
- if in.Template != nil {
- out.Template = new(core.PodTemplateSpec)
- if err := Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec(in.Template, out.Template, s); err != nil {
- return err
- }
- } else {
- out.Template = nil
- }
- return nil
-}
-
-func Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec(in *core.PodTemplateSpec, out *v1.PodTemplateSpec, s conversion.Scope) error {
- if err := autoConvert_core_PodTemplateSpec_To_v1_PodTemplateSpec(in, out, s); err != nil {
- return err
- }
-
- // drop init container annotations so they don't take effect on legacy kubelets.
- // remove this once the oldest supported kubelet no longer honors the annotations over the field.
- out.Annotations = dropInitContainerAnnotations(out.Annotations)
-
- return nil
-}
-
-func Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec(in *v1.PodTemplateSpec, out *core.PodTemplateSpec, s conversion.Scope) error {
- if err := autoConvert_v1_PodTemplateSpec_To_core_PodTemplateSpec(in, out, s); err != nil {
- return err
- }
-
- // drop init container annotations so they don't show up as differences when receiving requests from old clients
- out.Annotations = dropInitContainerAnnotations(out.Annotations)
-
- return nil
-}
-
-// The following two v1.PodSpec conversions are done here to support v1.ServiceAccount
-// as an alias for ServiceAccountName.
-func Convert_core_PodSpec_To_v1_PodSpec(in *core.PodSpec, out *v1.PodSpec, s conversion.Scope) error {
- if err := autoConvert_core_PodSpec_To_v1_PodSpec(in, out, s); err != nil {
- return err
- }
-
- // DeprecatedServiceAccount is an alias for ServiceAccountName.
- out.DeprecatedServiceAccount = in.ServiceAccountName
-
- if in.SecurityContext != nil {
- // the host namespace fields have to be handled here for backward compatibility
- // with v1.0.0
- out.HostPID = in.SecurityContext.HostPID
- out.HostNetwork = in.SecurityContext.HostNetwork
- out.HostIPC = in.SecurityContext.HostIPC
- out.ShareProcessNamespace = in.SecurityContext.ShareProcessNamespace
- }
-
- return nil
-}
-
-func Convert_v1_PodSpec_To_core_PodSpec(in *v1.PodSpec, out *core.PodSpec, s conversion.Scope) error {
- if err := autoConvert_v1_PodSpec_To_core_PodSpec(in, out, s); err != nil {
- return err
- }
-
- // We support DeprecatedServiceAccount as an alias for ServiceAccountName.
- // If both are specified, ServiceAccountName (the new field) wins.
- if in.ServiceAccountName == "" {
- out.ServiceAccountName = in.DeprecatedServiceAccount
- }
-
- // the host namespace fields have to be handled specially for backward compatibility
- // with v1.0.0
- if out.SecurityContext == nil {
- out.SecurityContext = new(core.PodSecurityContext)
- }
- out.SecurityContext.HostNetwork = in.HostNetwork
- out.SecurityContext.HostPID = in.HostPID
- out.SecurityContext.HostIPC = in.HostIPC
- out.SecurityContext.ShareProcessNamespace = in.ShareProcessNamespace
-
- return nil
-}
-
-func Convert_v1_Pod_To_core_Pod(in *v1.Pod, out *core.Pod, s conversion.Scope) error {
- if err := autoConvert_v1_Pod_To_core_Pod(in, out, s); err != nil {
- return err
- }
-
- // drop init container annotations so they don't show up as differences when receiving requests from old clients
- out.Annotations = dropInitContainerAnnotations(out.Annotations)
-
- return nil
-}
-
-func Convert_core_Pod_To_v1_Pod(in *core.Pod, out *v1.Pod, s conversion.Scope) error {
- if err := autoConvert_core_Pod_To_v1_Pod(in, out, s); err != nil {
- return err
- }
-
- // drop init container annotations so they don't take effect on legacy kubelets.
- // remove this once the oldest supported kubelet no longer honors the annotations over the field.
- out.Annotations = dropInitContainerAnnotations(out.Annotations)
-
- return nil
-}
-
-func Convert_v1_Secret_To_core_Secret(in *v1.Secret, out *core.Secret, s conversion.Scope) error {
- if err := autoConvert_v1_Secret_To_core_Secret(in, out, s); err != nil {
- return err
- }
-
- // StringData overwrites Data
- if len(in.StringData) > 0 {
- if out.Data == nil {
- out.Data = map[string][]byte{}
- }
- for k, v := range in.StringData {
- out.Data[k] = []byte(v)
- }
- }
-
- return nil
-}
-func Convert_core_SecurityContext_To_v1_SecurityContext(in *core.SecurityContext, out *v1.SecurityContext, s conversion.Scope) error {
- if in.Capabilities != nil {
- out.Capabilities = new(v1.Capabilities)
- if err := Convert_core_Capabilities_To_v1_Capabilities(in.Capabilities, out.Capabilities, s); err != nil {
- return err
- }
- } else {
- out.Capabilities = nil
- }
- out.Privileged = in.Privileged
- if in.SELinuxOptions != nil {
- out.SELinuxOptions = new(v1.SELinuxOptions)
- if err := Convert_core_SELinuxOptions_To_v1_SELinuxOptions(in.SELinuxOptions, out.SELinuxOptions, s); err != nil {
- return err
- }
- } else {
- out.SELinuxOptions = nil
- }
- out.RunAsUser = in.RunAsUser
- out.RunAsNonRoot = in.RunAsNonRoot
- out.ReadOnlyRootFilesystem = in.ReadOnlyRootFilesystem
- out.AllowPrivilegeEscalation = in.AllowPrivilegeEscalation
- return nil
-}
-
-func Convert_core_PodSecurityContext_To_v1_PodSecurityContext(in *core.PodSecurityContext, out *v1.PodSecurityContext, s conversion.Scope) error {
- out.SupplementalGroups = in.SupplementalGroups
- if in.SELinuxOptions != nil {
- out.SELinuxOptions = new(v1.SELinuxOptions)
- if err := Convert_core_SELinuxOptions_To_v1_SELinuxOptions(in.SELinuxOptions, out.SELinuxOptions, s); err != nil {
- return err
- }
- } else {
- out.SELinuxOptions = nil
- }
- out.RunAsUser = in.RunAsUser
- out.RunAsNonRoot = in.RunAsNonRoot
- out.FSGroup = in.FSGroup
- return nil
-}
-
-func Convert_v1_PodSecurityContext_To_core_PodSecurityContext(in *v1.PodSecurityContext, out *core.PodSecurityContext, s conversion.Scope) error {
- out.SupplementalGroups = in.SupplementalGroups
- if in.SELinuxOptions != nil {
- out.SELinuxOptions = new(core.SELinuxOptions)
- if err := Convert_v1_SELinuxOptions_To_core_SELinuxOptions(in.SELinuxOptions, out.SELinuxOptions, s); err != nil {
- return err
- }
- } else {
- out.SELinuxOptions = nil
- }
- out.RunAsUser = in.RunAsUser
- out.RunAsNonRoot = in.RunAsNonRoot
- out.FSGroup = in.FSGroup
- return nil
-}
-
-// +k8s:conversion-fn=copy-only
-func Convert_v1_ResourceList_To_core_ResourceList(in *v1.ResourceList, out *core.ResourceList, s conversion.Scope) error {
- if *in == nil {
- return nil
- }
- if *out == nil {
- *out = make(core.ResourceList, len(*in))
- }
- for key, val := range *in {
- // Moved to defaults
- // TODO(#18538): We round up resource values to milli scale to maintain API compatibility.
- // In the future, we should instead reject values that need rounding.
- // const milliScale = -3
- // val.RoundUp(milliScale)
-
- (*out)[core.ResourceName(key)] = val
- }
- return nil
-}
-
-func AddFieldLabelConversionsForEvent(scheme *runtime.Scheme) error {
- return scheme.AddFieldLabelConversionFunc("v1", "Event",
- func(label, value string) (string, string, error) {
- switch label {
- case "involvedObject.kind",
- "involvedObject.namespace",
- "involvedObject.name",
- "involvedObject.uid",
- "involvedObject.apiVersion",
- "involvedObject.resourceVersion",
- "involvedObject.fieldPath",
- "reason",
- "source",
- "type",
- "metadata.namespace",
- "metadata.name":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- })
-}
-
-func AddFieldLabelConversionsForNamespace(scheme *runtime.Scheme) error {
- return scheme.AddFieldLabelConversionFunc("v1", "Namespace",
- func(label, value string) (string, string, error) {
- switch label {
- case "status.phase",
- "metadata.name":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- })
-}
-
-func AddFieldLabelConversionsForSecret(scheme *runtime.Scheme) error {
- return scheme.AddFieldLabelConversionFunc("v1", "Secret",
- func(label, value string) (string, string, error) {
- switch label {
- case "type",
- "metadata.namespace",
- "metadata.name":
- return label, value, nil
- default:
- return "", "", fmt.Errorf("field label not supported: %s", label)
- }
- })
-}
-
-var initContainerAnnotations = map[string]bool{
- "pod.beta.kubernetes.io/init-containers": true,
- "pod.alpha.kubernetes.io/init-containers": true,
- "pod.beta.kubernetes.io/init-container-statuses": true,
- "pod.alpha.kubernetes.io/init-container-statuses": true,
-}
-
-// dropInitContainerAnnotations returns a copy of the annotations with init container annotations removed,
-// or the original annotations if no init container annotations were present.
-//
-// this can be removed once no clients prior to 1.8 are supported, and no kubelets prior to 1.8 can be run
-// (we don't support kubelets older than 2 versions skewed from the apiserver, but we don't prevent them, either)
-func dropInitContainerAnnotations(oldAnnotations map[string]string) map[string]string {
- if len(oldAnnotations) == 0 {
- return oldAnnotations
- }
-
- found := false
- for k := range initContainerAnnotations {
- if _, ok := oldAnnotations[k]; ok {
- found = true
- break
- }
- }
- if !found {
- return oldAnnotations
- }
-
- newAnnotations := make(map[string]string, len(oldAnnotations))
- for k, v := range oldAnnotations {
- if !initContainerAnnotations[k] {
- newAnnotations[k] = v
- }
- }
- return newAnnotations
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/defaults.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/defaults.go
deleted file mode 100644
index c2aeafc3d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/defaults.go
+++ /dev/null
@@ -1,413 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package v1
-
-import (
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/util/intstr"
- utilfeature "k8s.io/apiserver/pkg/util/feature"
- "k8s.io/kubernetes/pkg/features"
- "k8s.io/kubernetes/pkg/util/parsers"
- utilpointer "k8s.io/kubernetes/pkg/util/pointer"
-)
-
-func addDefaultingFuncs(scheme *runtime.Scheme) error {
- return RegisterDefaults(scheme)
-}
-
-func SetDefaults_ResourceList(obj *v1.ResourceList) {
- for key, val := range *obj {
- // TODO(#18538): We round up resource values to milli scale to maintain API compatibility.
- // In the future, we should instead reject values that need rounding.
- const milliScale = -3
- val.RoundUp(milliScale)
-
- (*obj)[v1.ResourceName(key)] = val
- }
-}
-
-func SetDefaults_ReplicationController(obj *v1.ReplicationController) {
- var labels map[string]string
- if obj.Spec.Template != nil {
- labels = obj.Spec.Template.Labels
- }
- // TODO: support templates defined elsewhere when we support them in the API
- if labels != nil {
- if len(obj.Spec.Selector) == 0 {
- obj.Spec.Selector = labels
- }
- if len(obj.Labels) == 0 {
- obj.Labels = labels
- }
- }
- if obj.Spec.Replicas == nil {
- obj.Spec.Replicas = new(int32)
- *obj.Spec.Replicas = 1
- }
-}
-func SetDefaults_Volume(obj *v1.Volume) {
- if utilpointer.AllPtrFieldsNil(&obj.VolumeSource) {
- obj.VolumeSource = v1.VolumeSource{
- EmptyDir: &v1.EmptyDirVolumeSource{},
- }
- }
-}
-func SetDefaults_ContainerPort(obj *v1.ContainerPort) {
- if obj.Protocol == "" {
- obj.Protocol = v1.ProtocolTCP
- }
-}
-func SetDefaults_Container(obj *v1.Container) {
- if obj.ImagePullPolicy == "" {
- // Ignore error and assume it has been validated elsewhere
- _, tag, _, _ := parsers.ParseImageName(obj.Image)
-
- // Check image tag
- if tag == "latest" {
- obj.ImagePullPolicy = v1.PullAlways
- } else {
- obj.ImagePullPolicy = v1.PullIfNotPresent
- }
- }
- if obj.TerminationMessagePath == "" {
- obj.TerminationMessagePath = v1.TerminationMessagePathDefault
- }
- if obj.TerminationMessagePolicy == "" {
- obj.TerminationMessagePolicy = v1.TerminationMessageReadFile
- }
-}
-func SetDefaults_Service(obj *v1.Service) {
- if obj.Spec.SessionAffinity == "" {
- obj.Spec.SessionAffinity = v1.ServiceAffinityNone
- }
- if obj.Spec.SessionAffinity == v1.ServiceAffinityNone {
- obj.Spec.SessionAffinityConfig = nil
- }
- if obj.Spec.SessionAffinity == v1.ServiceAffinityClientIP {
- if obj.Spec.SessionAffinityConfig == nil || obj.Spec.SessionAffinityConfig.ClientIP == nil || obj.Spec.SessionAffinityConfig.ClientIP.TimeoutSeconds == nil {
- timeoutSeconds := v1.DefaultClientIPServiceAffinitySeconds
- obj.Spec.SessionAffinityConfig = &v1.SessionAffinityConfig{
- ClientIP: &v1.ClientIPConfig{
- TimeoutSeconds: &timeoutSeconds,
- },
- }
- }
- }
- if obj.Spec.Type == "" {
- obj.Spec.Type = v1.ServiceTypeClusterIP
- }
- for i := range obj.Spec.Ports {
- sp := &obj.Spec.Ports[i]
- if sp.Protocol == "" {
- sp.Protocol = v1.ProtocolTCP
- }
- if sp.TargetPort == intstr.FromInt(0) || sp.TargetPort == intstr.FromString("") {
- sp.TargetPort = intstr.FromInt(int(sp.Port))
- }
- }
- // Defaults ExternalTrafficPolicy field for NodePort / LoadBalancer service
- // to Global for consistency.
- if (obj.Spec.Type == v1.ServiceTypeNodePort ||
- obj.Spec.Type == v1.ServiceTypeLoadBalancer) &&
- obj.Spec.ExternalTrafficPolicy == "" {
- obj.Spec.ExternalTrafficPolicy = v1.ServiceExternalTrafficPolicyTypeCluster
- }
-}
-func SetDefaults_Pod(obj *v1.Pod) {
- // If limits are specified, but requests are not, default requests to limits
- // This is done here rather than a more specific defaulting pass on v1.ResourceRequirements
- // because we only want this defaulting semantic to take place on a v1.Pod and not a v1.PodTemplate
- for i := range obj.Spec.Containers {
- // set requests to limits if requests are not specified, but limits are
- if obj.Spec.Containers[i].Resources.Limits != nil {
- if obj.Spec.Containers[i].Resources.Requests == nil {
- obj.Spec.Containers[i].Resources.Requests = make(v1.ResourceList)
- }
- for key, value := range obj.Spec.Containers[i].Resources.Limits {
- if _, exists := obj.Spec.Containers[i].Resources.Requests[key]; !exists {
- obj.Spec.Containers[i].Resources.Requests[key] = *(value.Copy())
- }
- }
- }
- }
- for i := range obj.Spec.InitContainers {
- if obj.Spec.InitContainers[i].Resources.Limits != nil {
- if obj.Spec.InitContainers[i].Resources.Requests == nil {
- obj.Spec.InitContainers[i].Resources.Requests = make(v1.ResourceList)
- }
- for key, value := range obj.Spec.InitContainers[i].Resources.Limits {
- if _, exists := obj.Spec.InitContainers[i].Resources.Requests[key]; !exists {
- obj.Spec.InitContainers[i].Resources.Requests[key] = *(value.Copy())
- }
- }
- }
- }
-}
-func SetDefaults_PodSpec(obj *v1.PodSpec) {
- if obj.DNSPolicy == "" {
- obj.DNSPolicy = v1.DNSClusterFirst
- }
- if obj.RestartPolicy == "" {
- obj.RestartPolicy = v1.RestartPolicyAlways
- }
- if obj.HostNetwork {
- defaultHostNetworkPorts(&obj.Containers)
- defaultHostNetworkPorts(&obj.InitContainers)
- }
- if obj.SecurityContext == nil {
- obj.SecurityContext = &v1.PodSecurityContext{}
- }
- if obj.TerminationGracePeriodSeconds == nil {
- period := int64(v1.DefaultTerminationGracePeriodSeconds)
- obj.TerminationGracePeriodSeconds = &period
- }
- if obj.SchedulerName == "" {
- obj.SchedulerName = v1.DefaultSchedulerName
- }
-}
-func SetDefaults_Probe(obj *v1.Probe) {
- if obj.TimeoutSeconds == 0 {
- obj.TimeoutSeconds = 1
- }
- if obj.PeriodSeconds == 0 {
- obj.PeriodSeconds = 10
- }
- if obj.SuccessThreshold == 0 {
- obj.SuccessThreshold = 1
- }
- if obj.FailureThreshold == 0 {
- obj.FailureThreshold = 3
- }
-}
-func SetDefaults_SecretVolumeSource(obj *v1.SecretVolumeSource) {
- if obj.DefaultMode == nil {
- perm := int32(v1.SecretVolumeSourceDefaultMode)
- obj.DefaultMode = &perm
- }
-}
-func SetDefaults_ConfigMapVolumeSource(obj *v1.ConfigMapVolumeSource) {
- if obj.DefaultMode == nil {
- perm := int32(v1.ConfigMapVolumeSourceDefaultMode)
- obj.DefaultMode = &perm
- }
-}
-func SetDefaults_DownwardAPIVolumeSource(obj *v1.DownwardAPIVolumeSource) {
- if obj.DefaultMode == nil {
- perm := int32(v1.DownwardAPIVolumeSourceDefaultMode)
- obj.DefaultMode = &perm
- }
-}
-func SetDefaults_Secret(obj *v1.Secret) {
- if obj.Type == "" {
- obj.Type = v1.SecretTypeOpaque
- }
-}
-func SetDefaults_ProjectedVolumeSource(obj *v1.ProjectedVolumeSource) {
- if obj.DefaultMode == nil {
- perm := int32(v1.ProjectedVolumeSourceDefaultMode)
- obj.DefaultMode = &perm
- }
-}
-func SetDefaults_PersistentVolume(obj *v1.PersistentVolume) {
- if obj.Status.Phase == "" {
- obj.Status.Phase = v1.VolumePending
- }
- if obj.Spec.PersistentVolumeReclaimPolicy == "" {
- obj.Spec.PersistentVolumeReclaimPolicy = v1.PersistentVolumeReclaimRetain
- }
- if obj.Spec.VolumeMode == nil && utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- obj.Spec.VolumeMode = new(v1.PersistentVolumeMode)
- *obj.Spec.VolumeMode = v1.PersistentVolumeFilesystem
- }
-}
-func SetDefaults_PersistentVolumeClaim(obj *v1.PersistentVolumeClaim) {
- if obj.Status.Phase == "" {
- obj.Status.Phase = v1.ClaimPending
- }
- if obj.Spec.VolumeMode == nil && utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- obj.Spec.VolumeMode = new(v1.PersistentVolumeMode)
- *obj.Spec.VolumeMode = v1.PersistentVolumeFilesystem
- }
-}
-func SetDefaults_ISCSIVolumeSource(obj *v1.ISCSIVolumeSource) {
- if obj.ISCSIInterface == "" {
- obj.ISCSIInterface = "default"
- }
-}
-func SetDefaults_ISCSIPersistentVolumeSource(obj *v1.ISCSIPersistentVolumeSource) {
- if obj.ISCSIInterface == "" {
- obj.ISCSIInterface = "default"
- }
-}
-func SetDefaults_AzureDiskVolumeSource(obj *v1.AzureDiskVolumeSource) {
- if obj.CachingMode == nil {
- obj.CachingMode = new(v1.AzureDataDiskCachingMode)
- *obj.CachingMode = v1.AzureDataDiskCachingReadWrite
- }
- if obj.Kind == nil {
- obj.Kind = new(v1.AzureDataDiskKind)
- *obj.Kind = v1.AzureSharedBlobDisk
- }
- if obj.FSType == nil {
- obj.FSType = new(string)
- *obj.FSType = "ext4"
- }
- if obj.ReadOnly == nil {
- obj.ReadOnly = new(bool)
- *obj.ReadOnly = false
- }
-}
-func SetDefaults_Endpoints(obj *v1.Endpoints) {
- for i := range obj.Subsets {
- ss := &obj.Subsets[i]
- for i := range ss.Ports {
- ep := &ss.Ports[i]
- if ep.Protocol == "" {
- ep.Protocol = v1.ProtocolTCP
- }
- }
- }
-}
-func SetDefaults_HTTPGetAction(obj *v1.HTTPGetAction) {
- if obj.Path == "" {
- obj.Path = "/"
- }
- if obj.Scheme == "" {
- obj.Scheme = v1.URISchemeHTTP
- }
-}
-func SetDefaults_NamespaceStatus(obj *v1.NamespaceStatus) {
- if obj.Phase == "" {
- obj.Phase = v1.NamespaceActive
- }
-}
-func SetDefaults_Node(obj *v1.Node) {
- if obj.Spec.ExternalID == "" {
- obj.Spec.ExternalID = obj.Name
- }
-}
-func SetDefaults_NodeStatus(obj *v1.NodeStatus) {
- if obj.Allocatable == nil && obj.Capacity != nil {
- obj.Allocatable = make(v1.ResourceList, len(obj.Capacity))
- for key, value := range obj.Capacity {
- obj.Allocatable[key] = *(value.Copy())
- }
- obj.Allocatable = obj.Capacity
- }
-}
-func SetDefaults_ObjectFieldSelector(obj *v1.ObjectFieldSelector) {
- if obj.APIVersion == "" {
- obj.APIVersion = "v1"
- }
-}
-func SetDefaults_LimitRangeItem(obj *v1.LimitRangeItem) {
- // for container limits, we apply default values
- if obj.Type == v1.LimitTypeContainer {
-
- if obj.Default == nil {
- obj.Default = make(v1.ResourceList)
- }
- if obj.DefaultRequest == nil {
- obj.DefaultRequest = make(v1.ResourceList)
- }
-
- // If a default limit is unspecified, but the max is specified, default the limit to the max
- for key, value := range obj.Max {
- if _, exists := obj.Default[key]; !exists {
- obj.Default[key] = *(value.Copy())
- }
- }
- // If a default limit is specified, but the default request is not, default request to limit
- for key, value := range obj.Default {
- if _, exists := obj.DefaultRequest[key]; !exists {
- obj.DefaultRequest[key] = *(value.Copy())
- }
- }
- // If a default request is not specified, but the min is provided, default request to the min
- for key, value := range obj.Min {
- if _, exists := obj.DefaultRequest[key]; !exists {
- obj.DefaultRequest[key] = *(value.Copy())
- }
- }
- }
-}
-func SetDefaults_ConfigMap(obj *v1.ConfigMap) {
- if obj.Data == nil {
- obj.Data = make(map[string]string)
- }
-}
-
-// With host networking default all container ports to host ports.
-func defaultHostNetworkPorts(containers *[]v1.Container) {
- for i := range *containers {
- for j := range (*containers)[i].Ports {
- if (*containers)[i].Ports[j].HostPort == 0 {
- (*containers)[i].Ports[j].HostPort = (*containers)[i].Ports[j].ContainerPort
- }
- }
- }
-}
-
-func SetDefaults_RBDVolumeSource(obj *v1.RBDVolumeSource) {
- if obj.RBDPool == "" {
- obj.RBDPool = "rbd"
- }
- if obj.RadosUser == "" {
- obj.RadosUser = "admin"
- }
- if obj.Keyring == "" {
- obj.Keyring = "/etc/ceph/keyring"
- }
-}
-
-func SetDefaults_RBDPersistentVolumeSource(obj *v1.RBDPersistentVolumeSource) {
- if obj.RBDPool == "" {
- obj.RBDPool = "rbd"
- }
- if obj.RadosUser == "" {
- obj.RadosUser = "admin"
- }
- if obj.Keyring == "" {
- obj.Keyring = "/etc/ceph/keyring"
- }
-}
-
-func SetDefaults_ScaleIOVolumeSource(obj *v1.ScaleIOVolumeSource) {
- if obj.StorageMode == "" {
- obj.StorageMode = "ThinProvisioned"
- }
- if obj.FSType == "" {
- obj.FSType = "xfs"
- }
-}
-
-func SetDefaults_ScaleIOPersistentVolumeSource(obj *v1.ScaleIOPersistentVolumeSource) {
- if obj.StorageMode == "" {
- obj.StorageMode = "ThinProvisioned"
- }
- if obj.FSType == "" {
- obj.FSType = "xfs"
- }
-}
-
-func SetDefaults_HostPathVolumeSource(obj *v1.HostPathVolumeSource) {
- typeVol := v1.HostPathUnset
- if obj.Type == nil {
- obj.Type = &typeVol
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/doc.go
deleted file mode 100644
index 454e30183..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/doc.go
+++ /dev/null
@@ -1,23 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:conversion-gen=k8s.io/kubernetes/pkg/apis/core
-// +k8s:conversion-gen-external-types=k8s.io/api/core/v1
-// +k8s:defaulter-gen=TypeMeta
-// +k8s:defaulter-gen-input=../../../../vendor/k8s.io/api/core/v1
-
-// Package v1 is the v1 version of the API.
-package v1 // import "k8s.io/kubernetes/pkg/apis/core/v1"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/helper/helpers.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/helper/helpers.go
deleted file mode 100644
index bb6d9385f..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/helper/helpers.go
+++ /dev/null
@@ -1,450 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package helper
-
-import (
- "encoding/json"
- "fmt"
- "strings"
-
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/api/resource"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/selection"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/validation"
- "k8s.io/kubernetes/pkg/apis/core/helper"
-)
-
-// IsExtendedResourceName returns true if:
-// 1. the resource name is not in the default namespace;
-// 2. resource name does not have "requests." prefix,
-// to avoid confusion with the convention in quota
-// 3. it satisfies the rules in IsQualifiedName() after converted into quota resource name
-func IsExtendedResourceName(name v1.ResourceName) bool {
- if IsDefaultNamespaceResource(name) || strings.HasPrefix(string(name), v1.DefaultResourceRequestsPrefix) {
- return false
- }
- // Ensure it satisfies the rules in IsQualifiedName() after converted into quota resource name
- nameForQuota := fmt.Sprintf("%s%s", v1.DefaultResourceRequestsPrefix, string(name))
- if errs := validation.IsQualifiedName(string(nameForQuota)); len(errs) != 0 {
- return false
- }
- return true
-}
-
-// IsDefaultNamespaceResource returns true if the resource name is in the
-// *kubernetes.io/ namespace. Partially-qualified (unprefixed) names are
-// implicitly in the kubernetes.io/ namespace.
-func IsDefaultNamespaceResource(name v1.ResourceName) bool {
- return !strings.Contains(string(name), "/") ||
- strings.Contains(string(name), v1.ResourceDefaultNamespacePrefix)
-
-}
-
-// IsHugePageResourceName returns true if the resource name has the huge page
-// resource prefix.
-func IsHugePageResourceName(name v1.ResourceName) bool {
- return strings.HasPrefix(string(name), v1.ResourceHugePagesPrefix)
-}
-
-// HugePageResourceName returns a ResourceName with the canonical hugepage
-// prefix prepended for the specified page size. The page size is converted
-// to its canonical representation.
-func HugePageResourceName(pageSize resource.Quantity) v1.ResourceName {
- return v1.ResourceName(fmt.Sprintf("%s%s", v1.ResourceHugePagesPrefix, pageSize.String()))
-}
-
-// HugePageSizeFromResourceName returns the page size for the specified huge page
-// resource name. If the specified input is not a valid huge page resource name
-// an error is returned.
-func HugePageSizeFromResourceName(name v1.ResourceName) (resource.Quantity, error) {
- if !IsHugePageResourceName(name) {
- return resource.Quantity{}, fmt.Errorf("resource name: %s is not valid hugepage name", name)
- }
- pageSize := strings.TrimPrefix(string(name), v1.ResourceHugePagesPrefix)
- return resource.ParseQuantity(pageSize)
-}
-
-var overcommitBlacklist = sets.NewString(string(v1.ResourceNvidiaGPU))
-
-// IsOvercommitAllowed returns true if the resource is in the default
-// namespace and not blacklisted and is not hugepages.
-func IsOvercommitAllowed(name v1.ResourceName) bool {
- return IsDefaultNamespaceResource(name) &&
- !IsHugePageResourceName(name) &&
- !overcommitBlacklist.Has(string(name))
-}
-
-// Extended and Hugepages resources
-func IsScalarResourceName(name v1.ResourceName) bool {
- return IsExtendedResourceName(name) || IsHugePageResourceName(name)
-}
-
-// this function aims to check if the service's ClusterIP is set or not
-// the objective is not to perform validation here
-func IsServiceIPSet(service *v1.Service) bool {
- return service.Spec.ClusterIP != v1.ClusterIPNone && service.Spec.ClusterIP != ""
-}
-
-// AddToNodeAddresses appends the NodeAddresses to the passed-by-pointer slice,
-// only if they do not already exist
-func AddToNodeAddresses(addresses *[]v1.NodeAddress, addAddresses ...v1.NodeAddress) {
- for _, add := range addAddresses {
- exists := false
- for _, existing := range *addresses {
- if existing.Address == add.Address && existing.Type == add.Type {
- exists = true
- break
- }
- }
- if !exists {
- *addresses = append(*addresses, add)
- }
- }
-}
-
-// TODO: make method on LoadBalancerStatus?
-func LoadBalancerStatusEqual(l, r *v1.LoadBalancerStatus) bool {
- return ingressSliceEqual(l.Ingress, r.Ingress)
-}
-
-func ingressSliceEqual(lhs, rhs []v1.LoadBalancerIngress) bool {
- if len(lhs) != len(rhs) {
- return false
- }
- for i := range lhs {
- if !ingressEqual(&lhs[i], &rhs[i]) {
- return false
- }
- }
- return true
-}
-
-func ingressEqual(lhs, rhs *v1.LoadBalancerIngress) bool {
- if lhs.IP != rhs.IP {
- return false
- }
- if lhs.Hostname != rhs.Hostname {
- return false
- }
- return true
-}
-
-// TODO: make method on LoadBalancerStatus?
-func LoadBalancerStatusDeepCopy(lb *v1.LoadBalancerStatus) *v1.LoadBalancerStatus {
- c := &v1.LoadBalancerStatus{}
- c.Ingress = make([]v1.LoadBalancerIngress, len(lb.Ingress))
- for i := range lb.Ingress {
- c.Ingress[i] = lb.Ingress[i]
- }
- return c
-}
-
-// GetAccessModesAsString returns a string representation of an array of access modes.
-// modes, when present, are always in the same order: RWO,ROX,RWX.
-func GetAccessModesAsString(modes []v1.PersistentVolumeAccessMode) string {
- modes = removeDuplicateAccessModes(modes)
- modesStr := []string{}
- if containsAccessMode(modes, v1.ReadWriteOnce) {
- modesStr = append(modesStr, "RWO")
- }
- if containsAccessMode(modes, v1.ReadOnlyMany) {
- modesStr = append(modesStr, "ROX")
- }
- if containsAccessMode(modes, v1.ReadWriteMany) {
- modesStr = append(modesStr, "RWX")
- }
- return strings.Join(modesStr, ",")
-}
-
-// GetAccessModesAsString returns an array of AccessModes from a string created by GetAccessModesAsString
-func GetAccessModesFromString(modes string) []v1.PersistentVolumeAccessMode {
- strmodes := strings.Split(modes, ",")
- accessModes := []v1.PersistentVolumeAccessMode{}
- for _, s := range strmodes {
- s = strings.Trim(s, " ")
- switch {
- case s == "RWO":
- accessModes = append(accessModes, v1.ReadWriteOnce)
- case s == "ROX":
- accessModes = append(accessModes, v1.ReadOnlyMany)
- case s == "RWX":
- accessModes = append(accessModes, v1.ReadWriteMany)
- }
- }
- return accessModes
-}
-
-// removeDuplicateAccessModes returns an array of access modes without any duplicates
-func removeDuplicateAccessModes(modes []v1.PersistentVolumeAccessMode) []v1.PersistentVolumeAccessMode {
- accessModes := []v1.PersistentVolumeAccessMode{}
- for _, m := range modes {
- if !containsAccessMode(accessModes, m) {
- accessModes = append(accessModes, m)
- }
- }
- return accessModes
-}
-
-func containsAccessMode(modes []v1.PersistentVolumeAccessMode, mode v1.PersistentVolumeAccessMode) bool {
- for _, m := range modes {
- if m == mode {
- return true
- }
- }
- return false
-}
-
-// NodeSelectorRequirementsAsSelector converts the []NodeSelectorRequirement api type into a struct that implements
-// labels.Selector.
-func NodeSelectorRequirementsAsSelector(nsm []v1.NodeSelectorRequirement) (labels.Selector, error) {
- if len(nsm) == 0 {
- return labels.Nothing(), nil
- }
- selector := labels.NewSelector()
- for _, expr := range nsm {
- var op selection.Operator
- switch expr.Operator {
- case v1.NodeSelectorOpIn:
- op = selection.In
- case v1.NodeSelectorOpNotIn:
- op = selection.NotIn
- case v1.NodeSelectorOpExists:
- op = selection.Exists
- case v1.NodeSelectorOpDoesNotExist:
- op = selection.DoesNotExist
- case v1.NodeSelectorOpGt:
- op = selection.GreaterThan
- case v1.NodeSelectorOpLt:
- op = selection.LessThan
- default:
- return nil, fmt.Errorf("%q is not a valid node selector operator", expr.Operator)
- }
- r, err := labels.NewRequirement(expr.Key, op, expr.Values)
- if err != nil {
- return nil, err
- }
- selector = selector.Add(*r)
- }
- return selector, nil
-}
-
-// AddOrUpdateTolerationInPodSpec tries to add a toleration to the toleration list in PodSpec.
-// Returns true if something was updated, false otherwise.
-func AddOrUpdateTolerationInPodSpec(spec *v1.PodSpec, toleration *v1.Toleration) bool {
- podTolerations := spec.Tolerations
-
- var newTolerations []v1.Toleration
- updated := false
- for i := range podTolerations {
- if toleration.MatchToleration(&podTolerations[i]) {
- if helper.Semantic.DeepEqual(toleration, podTolerations[i]) {
- return false
- }
- newTolerations = append(newTolerations, *toleration)
- updated = true
- continue
- }
-
- newTolerations = append(newTolerations, podTolerations[i])
- }
-
- if !updated {
- newTolerations = append(newTolerations, *toleration)
- }
-
- spec.Tolerations = newTolerations
- return true
-}
-
-// AddOrUpdateTolerationInPod tries to add a toleration to the pod's toleration list.
-// Returns true if something was updated, false otherwise.
-func AddOrUpdateTolerationInPod(pod *v1.Pod, toleration *v1.Toleration) bool {
- return AddOrUpdateTolerationInPodSpec(&pod.Spec, toleration)
-}
-
-// TolerationsTolerateTaint checks if taint is tolerated by any of the tolerations.
-func TolerationsTolerateTaint(tolerations []v1.Toleration, taint *v1.Taint) bool {
- for i := range tolerations {
- if tolerations[i].ToleratesTaint(taint) {
- return true
- }
- }
- return false
-}
-
-type taintsFilterFunc func(*v1.Taint) bool
-
-// TolerationsTolerateTaintsWithFilter checks if given tolerations tolerates
-// all the taints that apply to the filter in given taint list.
-func TolerationsTolerateTaintsWithFilter(tolerations []v1.Toleration, taints []v1.Taint, applyFilter taintsFilterFunc) bool {
- if len(taints) == 0 {
- return true
- }
-
- for i := range taints {
- if applyFilter != nil && !applyFilter(&taints[i]) {
- continue
- }
-
- if !TolerationsTolerateTaint(tolerations, &taints[i]) {
- return false
- }
- }
-
- return true
-}
-
-// Returns true and list of Tolerations matching all Taints if all are tolerated, or false otherwise.
-func GetMatchingTolerations(taints []v1.Taint, tolerations []v1.Toleration) (bool, []v1.Toleration) {
- if len(taints) == 0 {
- return true, []v1.Toleration{}
- }
- if len(tolerations) == 0 && len(taints) > 0 {
- return false, []v1.Toleration{}
- }
- result := []v1.Toleration{}
- for i := range taints {
- tolerated := false
- for j := range tolerations {
- if tolerations[j].ToleratesTaint(&taints[i]) {
- result = append(result, tolerations[j])
- tolerated = true
- break
- }
- }
- if !tolerated {
- return false, []v1.Toleration{}
- }
- }
- return true, result
-}
-
-func GetAvoidPodsFromNodeAnnotations(annotations map[string]string) (v1.AvoidPods, error) {
- var avoidPods v1.AvoidPods
- if len(annotations) > 0 && annotations[v1.PreferAvoidPodsAnnotationKey] != "" {
- err := json.Unmarshal([]byte(annotations[v1.PreferAvoidPodsAnnotationKey]), &avoidPods)
- if err != nil {
- return avoidPods, err
- }
- }
- return avoidPods, nil
-}
-
-// SysctlsFromPodAnnotations parses the sysctl annotations into a slice of safe Sysctls
-// and a slice of unsafe Sysctls. This is only a convenience wrapper around
-// SysctlsFromPodAnnotation.
-func SysctlsFromPodAnnotations(a map[string]string) ([]v1.Sysctl, []v1.Sysctl, error) {
- safe, err := SysctlsFromPodAnnotation(a[v1.SysctlsPodAnnotationKey])
- if err != nil {
- return nil, nil, err
- }
- unsafe, err := SysctlsFromPodAnnotation(a[v1.UnsafeSysctlsPodAnnotationKey])
- if err != nil {
- return nil, nil, err
- }
-
- return safe, unsafe, nil
-}
-
-// SysctlsFromPodAnnotation parses an annotation value into a slice of Sysctls.
-func SysctlsFromPodAnnotation(annotation string) ([]v1.Sysctl, error) {
- if len(annotation) == 0 {
- return nil, nil
- }
-
- kvs := strings.Split(annotation, ",")
- sysctls := make([]v1.Sysctl, len(kvs))
- for i, kv := range kvs {
- cs := strings.Split(kv, "=")
- if len(cs) != 2 || len(cs[0]) == 0 {
- return nil, fmt.Errorf("sysctl %q not of the format sysctl_name=value", kv)
- }
- sysctls[i].Name = cs[0]
- sysctls[i].Value = cs[1]
- }
- return sysctls, nil
-}
-
-// PodAnnotationsFromSysctls creates an annotation value for a slice of Sysctls.
-func PodAnnotationsFromSysctls(sysctls []v1.Sysctl) string {
- if len(sysctls) == 0 {
- return ""
- }
-
- kvs := make([]string, len(sysctls))
- for i := range sysctls {
- kvs[i] = fmt.Sprintf("%s=%s", sysctls[i].Name, sysctls[i].Value)
- }
- return strings.Join(kvs, ",")
-}
-
-// GetPersistentVolumeClass returns StorageClassName.
-func GetPersistentVolumeClass(volume *v1.PersistentVolume) string {
- // Use beta annotation first
- if class, found := volume.Annotations[v1.BetaStorageClassAnnotation]; found {
- return class
- }
-
- return volume.Spec.StorageClassName
-}
-
-// GetPersistentVolumeClaimClass returns StorageClassName. If no storage class was
-// requested, it returns "".
-func GetPersistentVolumeClaimClass(claim *v1.PersistentVolumeClaim) string {
- // Use beta annotation first
- if class, found := claim.Annotations[v1.BetaStorageClassAnnotation]; found {
- return class
- }
-
- if claim.Spec.StorageClassName != nil {
- return *claim.Spec.StorageClassName
- }
-
- return ""
-}
-
-// GetStorageNodeAffinityFromAnnotation gets the json serialized data from PersistentVolume.Annotations
-// and converts it to the NodeAffinity type in api.
-// TODO: update when storage node affinity graduates to beta
-func GetStorageNodeAffinityFromAnnotation(annotations map[string]string) (*v1.NodeAffinity, error) {
- if len(annotations) > 0 && annotations[v1.AlphaStorageNodeAffinityAnnotation] != "" {
- var affinity v1.NodeAffinity
- err := json.Unmarshal([]byte(annotations[v1.AlphaStorageNodeAffinityAnnotation]), &affinity)
- if err != nil {
- return nil, err
- }
- return &affinity, nil
- }
- return nil, nil
-}
-
-// Converts NodeAffinity type to Alpha annotation for use in PersistentVolumes
-// TODO: update when storage node affinity graduates to beta
-func StorageNodeAffinityToAlphaAnnotation(annotations map[string]string, affinity *v1.NodeAffinity) error {
- if affinity == nil {
- return nil
- }
-
- json, err := json.Marshal(*affinity)
- if err != nil {
- return err
- }
- annotations[v1.AlphaStorageNodeAffinityAnnotation] = string(json)
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/register.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/register.go
deleted file mode 100644
index 79611beab..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/register.go
+++ /dev/null
@@ -1,46 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package v1
-
-import (
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-var (
- localSchemeBuilder = &v1.SchemeBuilder
- AddToScheme = localSchemeBuilder.AddToScheme
-)
-
-func init() {
- // We only register manually written functions here. The registration of the
- // generated functions takes place in the generated files. The separation
- // makes the code compile even when the generated files are missing.
- localSchemeBuilder.Register(addDefaultingFuncs, addConversionFuncs, addFastPathConversionFuncs)
-}
-
-// TODO: remove these global varialbes
-// GroupName is the group name use in this package
-const GroupName = ""
-
-// SchemeGroupVersion is group version used to register these objects
-var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: "v1"}
-
-// Resource takes an unqualified resource and returns a Group qualified GroupResource
-func Resource(resource string) schema.GroupResource {
- return SchemeGroupVersion.WithResource(resource).GroupResource()
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.conversion.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.conversion.go
deleted file mode 100644
index 17773f26e..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.conversion.go
+++ /dev/null
@@ -1,5686 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by conversion-gen. DO NOT EDIT.
-
-package v1
-
-import (
- unsafe "unsafe"
-
- v1 "k8s.io/api/core/v1"
- resource "k8s.io/apimachinery/pkg/api/resource"
- meta_v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- conversion "k8s.io/apimachinery/pkg/conversion"
- runtime "k8s.io/apimachinery/pkg/runtime"
- types "k8s.io/apimachinery/pkg/types"
- core "k8s.io/kubernetes/pkg/apis/core"
-)
-
-func init() {
- localSchemeBuilder.Register(RegisterConversions)
-}
-
-// RegisterConversions adds conversion functions to the given scheme.
-// Public to allow building arbitrary schemes.
-func RegisterConversions(scheme *runtime.Scheme) error {
- return scheme.AddGeneratedConversionFuncs(
- Convert_v1_AWSElasticBlockStoreVolumeSource_To_core_AWSElasticBlockStoreVolumeSource,
- Convert_core_AWSElasticBlockStoreVolumeSource_To_v1_AWSElasticBlockStoreVolumeSource,
- Convert_v1_Affinity_To_core_Affinity,
- Convert_core_Affinity_To_v1_Affinity,
- Convert_v1_AttachedVolume_To_core_AttachedVolume,
- Convert_core_AttachedVolume_To_v1_AttachedVolume,
- Convert_v1_AvoidPods_To_core_AvoidPods,
- Convert_core_AvoidPods_To_v1_AvoidPods,
- Convert_v1_AzureDiskVolumeSource_To_core_AzureDiskVolumeSource,
- Convert_core_AzureDiskVolumeSource_To_v1_AzureDiskVolumeSource,
- Convert_v1_AzureFilePersistentVolumeSource_To_core_AzureFilePersistentVolumeSource,
- Convert_core_AzureFilePersistentVolumeSource_To_v1_AzureFilePersistentVolumeSource,
- Convert_v1_AzureFileVolumeSource_To_core_AzureFileVolumeSource,
- Convert_core_AzureFileVolumeSource_To_v1_AzureFileVolumeSource,
- Convert_v1_Binding_To_core_Binding,
- Convert_core_Binding_To_v1_Binding,
- Convert_v1_CSIPersistentVolumeSource_To_core_CSIPersistentVolumeSource,
- Convert_core_CSIPersistentVolumeSource_To_v1_CSIPersistentVolumeSource,
- Convert_v1_Capabilities_To_core_Capabilities,
- Convert_core_Capabilities_To_v1_Capabilities,
- Convert_v1_CephFSPersistentVolumeSource_To_core_CephFSPersistentVolumeSource,
- Convert_core_CephFSPersistentVolumeSource_To_v1_CephFSPersistentVolumeSource,
- Convert_v1_CephFSVolumeSource_To_core_CephFSVolumeSource,
- Convert_core_CephFSVolumeSource_To_v1_CephFSVolumeSource,
- Convert_v1_CinderVolumeSource_To_core_CinderVolumeSource,
- Convert_core_CinderVolumeSource_To_v1_CinderVolumeSource,
- Convert_v1_ClientIPConfig_To_core_ClientIPConfig,
- Convert_core_ClientIPConfig_To_v1_ClientIPConfig,
- Convert_v1_ComponentCondition_To_core_ComponentCondition,
- Convert_core_ComponentCondition_To_v1_ComponentCondition,
- Convert_v1_ComponentStatus_To_core_ComponentStatus,
- Convert_core_ComponentStatus_To_v1_ComponentStatus,
- Convert_v1_ComponentStatusList_To_core_ComponentStatusList,
- Convert_core_ComponentStatusList_To_v1_ComponentStatusList,
- Convert_v1_ConfigMap_To_core_ConfigMap,
- Convert_core_ConfigMap_To_v1_ConfigMap,
- Convert_v1_ConfigMapEnvSource_To_core_ConfigMapEnvSource,
- Convert_core_ConfigMapEnvSource_To_v1_ConfigMapEnvSource,
- Convert_v1_ConfigMapKeySelector_To_core_ConfigMapKeySelector,
- Convert_core_ConfigMapKeySelector_To_v1_ConfigMapKeySelector,
- Convert_v1_ConfigMapList_To_core_ConfigMapList,
- Convert_core_ConfigMapList_To_v1_ConfigMapList,
- Convert_v1_ConfigMapProjection_To_core_ConfigMapProjection,
- Convert_core_ConfigMapProjection_To_v1_ConfigMapProjection,
- Convert_v1_ConfigMapVolumeSource_To_core_ConfigMapVolumeSource,
- Convert_core_ConfigMapVolumeSource_To_v1_ConfigMapVolumeSource,
- Convert_v1_Container_To_core_Container,
- Convert_core_Container_To_v1_Container,
- Convert_v1_ContainerImage_To_core_ContainerImage,
- Convert_core_ContainerImage_To_v1_ContainerImage,
- Convert_v1_ContainerPort_To_core_ContainerPort,
- Convert_core_ContainerPort_To_v1_ContainerPort,
- Convert_v1_ContainerState_To_core_ContainerState,
- Convert_core_ContainerState_To_v1_ContainerState,
- Convert_v1_ContainerStateRunning_To_core_ContainerStateRunning,
- Convert_core_ContainerStateRunning_To_v1_ContainerStateRunning,
- Convert_v1_ContainerStateTerminated_To_core_ContainerStateTerminated,
- Convert_core_ContainerStateTerminated_To_v1_ContainerStateTerminated,
- Convert_v1_ContainerStateWaiting_To_core_ContainerStateWaiting,
- Convert_core_ContainerStateWaiting_To_v1_ContainerStateWaiting,
- Convert_v1_ContainerStatus_To_core_ContainerStatus,
- Convert_core_ContainerStatus_To_v1_ContainerStatus,
- Convert_v1_DaemonEndpoint_To_core_DaemonEndpoint,
- Convert_core_DaemonEndpoint_To_v1_DaemonEndpoint,
- Convert_v1_DeleteOptions_To_core_DeleteOptions,
- Convert_core_DeleteOptions_To_v1_DeleteOptions,
- Convert_v1_DownwardAPIProjection_To_core_DownwardAPIProjection,
- Convert_core_DownwardAPIProjection_To_v1_DownwardAPIProjection,
- Convert_v1_DownwardAPIVolumeFile_To_core_DownwardAPIVolumeFile,
- Convert_core_DownwardAPIVolumeFile_To_v1_DownwardAPIVolumeFile,
- Convert_v1_DownwardAPIVolumeSource_To_core_DownwardAPIVolumeSource,
- Convert_core_DownwardAPIVolumeSource_To_v1_DownwardAPIVolumeSource,
- Convert_v1_EmptyDirVolumeSource_To_core_EmptyDirVolumeSource,
- Convert_core_EmptyDirVolumeSource_To_v1_EmptyDirVolumeSource,
- Convert_v1_EndpointAddress_To_core_EndpointAddress,
- Convert_core_EndpointAddress_To_v1_EndpointAddress,
- Convert_v1_EndpointPort_To_core_EndpointPort,
- Convert_core_EndpointPort_To_v1_EndpointPort,
- Convert_v1_EndpointSubset_To_core_EndpointSubset,
- Convert_core_EndpointSubset_To_v1_EndpointSubset,
- Convert_v1_Endpoints_To_core_Endpoints,
- Convert_core_Endpoints_To_v1_Endpoints,
- Convert_v1_EndpointsList_To_core_EndpointsList,
- Convert_core_EndpointsList_To_v1_EndpointsList,
- Convert_v1_EnvFromSource_To_core_EnvFromSource,
- Convert_core_EnvFromSource_To_v1_EnvFromSource,
- Convert_v1_EnvVar_To_core_EnvVar,
- Convert_core_EnvVar_To_v1_EnvVar,
- Convert_v1_EnvVarSource_To_core_EnvVarSource,
- Convert_core_EnvVarSource_To_v1_EnvVarSource,
- Convert_v1_Event_To_core_Event,
- Convert_core_Event_To_v1_Event,
- Convert_v1_EventList_To_core_EventList,
- Convert_core_EventList_To_v1_EventList,
- Convert_v1_EventSeries_To_core_EventSeries,
- Convert_core_EventSeries_To_v1_EventSeries,
- Convert_v1_EventSource_To_core_EventSource,
- Convert_core_EventSource_To_v1_EventSource,
- Convert_v1_ExecAction_To_core_ExecAction,
- Convert_core_ExecAction_To_v1_ExecAction,
- Convert_v1_FCVolumeSource_To_core_FCVolumeSource,
- Convert_core_FCVolumeSource_To_v1_FCVolumeSource,
- Convert_v1_FlexPersistentVolumeSource_To_core_FlexPersistentVolumeSource,
- Convert_core_FlexPersistentVolumeSource_To_v1_FlexPersistentVolumeSource,
- Convert_v1_FlexVolumeSource_To_core_FlexVolumeSource,
- Convert_core_FlexVolumeSource_To_v1_FlexVolumeSource,
- Convert_v1_FlockerVolumeSource_To_core_FlockerVolumeSource,
- Convert_core_FlockerVolumeSource_To_v1_FlockerVolumeSource,
- Convert_v1_GCEPersistentDiskVolumeSource_To_core_GCEPersistentDiskVolumeSource,
- Convert_core_GCEPersistentDiskVolumeSource_To_v1_GCEPersistentDiskVolumeSource,
- Convert_v1_GitRepoVolumeSource_To_core_GitRepoVolumeSource,
- Convert_core_GitRepoVolumeSource_To_v1_GitRepoVolumeSource,
- Convert_v1_GlusterfsVolumeSource_To_core_GlusterfsVolumeSource,
- Convert_core_GlusterfsVolumeSource_To_v1_GlusterfsVolumeSource,
- Convert_v1_HTTPGetAction_To_core_HTTPGetAction,
- Convert_core_HTTPGetAction_To_v1_HTTPGetAction,
- Convert_v1_HTTPHeader_To_core_HTTPHeader,
- Convert_core_HTTPHeader_To_v1_HTTPHeader,
- Convert_v1_Handler_To_core_Handler,
- Convert_core_Handler_To_v1_Handler,
- Convert_v1_HostAlias_To_core_HostAlias,
- Convert_core_HostAlias_To_v1_HostAlias,
- Convert_v1_HostPathVolumeSource_To_core_HostPathVolumeSource,
- Convert_core_HostPathVolumeSource_To_v1_HostPathVolumeSource,
- Convert_v1_ISCSIPersistentVolumeSource_To_core_ISCSIPersistentVolumeSource,
- Convert_core_ISCSIPersistentVolumeSource_To_v1_ISCSIPersistentVolumeSource,
- Convert_v1_ISCSIVolumeSource_To_core_ISCSIVolumeSource,
- Convert_core_ISCSIVolumeSource_To_v1_ISCSIVolumeSource,
- Convert_v1_KeyToPath_To_core_KeyToPath,
- Convert_core_KeyToPath_To_v1_KeyToPath,
- Convert_v1_Lifecycle_To_core_Lifecycle,
- Convert_core_Lifecycle_To_v1_Lifecycle,
- Convert_v1_LimitRange_To_core_LimitRange,
- Convert_core_LimitRange_To_v1_LimitRange,
- Convert_v1_LimitRangeItem_To_core_LimitRangeItem,
- Convert_core_LimitRangeItem_To_v1_LimitRangeItem,
- Convert_v1_LimitRangeList_To_core_LimitRangeList,
- Convert_core_LimitRangeList_To_v1_LimitRangeList,
- Convert_v1_LimitRangeSpec_To_core_LimitRangeSpec,
- Convert_core_LimitRangeSpec_To_v1_LimitRangeSpec,
- Convert_v1_List_To_core_List,
- Convert_core_List_To_v1_List,
- Convert_v1_ListOptions_To_core_ListOptions,
- Convert_core_ListOptions_To_v1_ListOptions,
- Convert_v1_LoadBalancerIngress_To_core_LoadBalancerIngress,
- Convert_core_LoadBalancerIngress_To_v1_LoadBalancerIngress,
- Convert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus,
- Convert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus,
- Convert_v1_LocalObjectReference_To_core_LocalObjectReference,
- Convert_core_LocalObjectReference_To_v1_LocalObjectReference,
- Convert_v1_LocalVolumeSource_To_core_LocalVolumeSource,
- Convert_core_LocalVolumeSource_To_v1_LocalVolumeSource,
- Convert_v1_NFSVolumeSource_To_core_NFSVolumeSource,
- Convert_core_NFSVolumeSource_To_v1_NFSVolumeSource,
- Convert_v1_Namespace_To_core_Namespace,
- Convert_core_Namespace_To_v1_Namespace,
- Convert_v1_NamespaceList_To_core_NamespaceList,
- Convert_core_NamespaceList_To_v1_NamespaceList,
- Convert_v1_NamespaceSpec_To_core_NamespaceSpec,
- Convert_core_NamespaceSpec_To_v1_NamespaceSpec,
- Convert_v1_NamespaceStatus_To_core_NamespaceStatus,
- Convert_core_NamespaceStatus_To_v1_NamespaceStatus,
- Convert_v1_Node_To_core_Node,
- Convert_core_Node_To_v1_Node,
- Convert_v1_NodeAddress_To_core_NodeAddress,
- Convert_core_NodeAddress_To_v1_NodeAddress,
- Convert_v1_NodeAffinity_To_core_NodeAffinity,
- Convert_core_NodeAffinity_To_v1_NodeAffinity,
- Convert_v1_NodeCondition_To_core_NodeCondition,
- Convert_core_NodeCondition_To_v1_NodeCondition,
- Convert_v1_NodeConfigSource_To_core_NodeConfigSource,
- Convert_core_NodeConfigSource_To_v1_NodeConfigSource,
- Convert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints,
- Convert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints,
- Convert_v1_NodeList_To_core_NodeList,
- Convert_core_NodeList_To_v1_NodeList,
- Convert_v1_NodeProxyOptions_To_core_NodeProxyOptions,
- Convert_core_NodeProxyOptions_To_v1_NodeProxyOptions,
- Convert_v1_NodeResources_To_core_NodeResources,
- Convert_core_NodeResources_To_v1_NodeResources,
- Convert_v1_NodeSelector_To_core_NodeSelector,
- Convert_core_NodeSelector_To_v1_NodeSelector,
- Convert_v1_NodeSelectorRequirement_To_core_NodeSelectorRequirement,
- Convert_core_NodeSelectorRequirement_To_v1_NodeSelectorRequirement,
- Convert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm,
- Convert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm,
- Convert_v1_NodeSpec_To_core_NodeSpec,
- Convert_core_NodeSpec_To_v1_NodeSpec,
- Convert_v1_NodeStatus_To_core_NodeStatus,
- Convert_core_NodeStatus_To_v1_NodeStatus,
- Convert_v1_NodeSystemInfo_To_core_NodeSystemInfo,
- Convert_core_NodeSystemInfo_To_v1_NodeSystemInfo,
- Convert_v1_ObjectFieldSelector_To_core_ObjectFieldSelector,
- Convert_core_ObjectFieldSelector_To_v1_ObjectFieldSelector,
- Convert_v1_ObjectMeta_To_core_ObjectMeta,
- Convert_core_ObjectMeta_To_v1_ObjectMeta,
- Convert_v1_ObjectReference_To_core_ObjectReference,
- Convert_core_ObjectReference_To_v1_ObjectReference,
- Convert_v1_PersistentVolume_To_core_PersistentVolume,
- Convert_core_PersistentVolume_To_v1_PersistentVolume,
- Convert_v1_PersistentVolumeClaim_To_core_PersistentVolumeClaim,
- Convert_core_PersistentVolumeClaim_To_v1_PersistentVolumeClaim,
- Convert_v1_PersistentVolumeClaimCondition_To_core_PersistentVolumeClaimCondition,
- Convert_core_PersistentVolumeClaimCondition_To_v1_PersistentVolumeClaimCondition,
- Convert_v1_PersistentVolumeClaimList_To_core_PersistentVolumeClaimList,
- Convert_core_PersistentVolumeClaimList_To_v1_PersistentVolumeClaimList,
- Convert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec,
- Convert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec,
- Convert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus,
- Convert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus,
- Convert_v1_PersistentVolumeClaimVolumeSource_To_core_PersistentVolumeClaimVolumeSource,
- Convert_core_PersistentVolumeClaimVolumeSource_To_v1_PersistentVolumeClaimVolumeSource,
- Convert_v1_PersistentVolumeList_To_core_PersistentVolumeList,
- Convert_core_PersistentVolumeList_To_v1_PersistentVolumeList,
- Convert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource,
- Convert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource,
- Convert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec,
- Convert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec,
- Convert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus,
- Convert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus,
- Convert_v1_PhotonPersistentDiskVolumeSource_To_core_PhotonPersistentDiskVolumeSource,
- Convert_core_PhotonPersistentDiskVolumeSource_To_v1_PhotonPersistentDiskVolumeSource,
- Convert_v1_Pod_To_core_Pod,
- Convert_core_Pod_To_v1_Pod,
- Convert_v1_PodAffinity_To_core_PodAffinity,
- Convert_core_PodAffinity_To_v1_PodAffinity,
- Convert_v1_PodAffinityTerm_To_core_PodAffinityTerm,
- Convert_core_PodAffinityTerm_To_v1_PodAffinityTerm,
- Convert_v1_PodAntiAffinity_To_core_PodAntiAffinity,
- Convert_core_PodAntiAffinity_To_v1_PodAntiAffinity,
- Convert_v1_PodAttachOptions_To_core_PodAttachOptions,
- Convert_core_PodAttachOptions_To_v1_PodAttachOptions,
- Convert_v1_PodCondition_To_core_PodCondition,
- Convert_core_PodCondition_To_v1_PodCondition,
- Convert_v1_PodDNSConfig_To_core_PodDNSConfig,
- Convert_core_PodDNSConfig_To_v1_PodDNSConfig,
- Convert_v1_PodDNSConfigOption_To_core_PodDNSConfigOption,
- Convert_core_PodDNSConfigOption_To_v1_PodDNSConfigOption,
- Convert_v1_PodExecOptions_To_core_PodExecOptions,
- Convert_core_PodExecOptions_To_v1_PodExecOptions,
- Convert_v1_PodList_To_core_PodList,
- Convert_core_PodList_To_v1_PodList,
- Convert_v1_PodLogOptions_To_core_PodLogOptions,
- Convert_core_PodLogOptions_To_v1_PodLogOptions,
- Convert_v1_PodPortForwardOptions_To_core_PodPortForwardOptions,
- Convert_core_PodPortForwardOptions_To_v1_PodPortForwardOptions,
- Convert_v1_PodProxyOptions_To_core_PodProxyOptions,
- Convert_core_PodProxyOptions_To_v1_PodProxyOptions,
- Convert_v1_PodSecurityContext_To_core_PodSecurityContext,
- Convert_core_PodSecurityContext_To_v1_PodSecurityContext,
- Convert_v1_PodSignature_To_core_PodSignature,
- Convert_core_PodSignature_To_v1_PodSignature,
- Convert_v1_PodSpec_To_core_PodSpec,
- Convert_core_PodSpec_To_v1_PodSpec,
- Convert_v1_PodStatus_To_core_PodStatus,
- Convert_core_PodStatus_To_v1_PodStatus,
- Convert_v1_PodStatusResult_To_core_PodStatusResult,
- Convert_core_PodStatusResult_To_v1_PodStatusResult,
- Convert_v1_PodTemplate_To_core_PodTemplate,
- Convert_core_PodTemplate_To_v1_PodTemplate,
- Convert_v1_PodTemplateList_To_core_PodTemplateList,
- Convert_core_PodTemplateList_To_v1_PodTemplateList,
- Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec,
- Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec,
- Convert_v1_PortworxVolumeSource_To_core_PortworxVolumeSource,
- Convert_core_PortworxVolumeSource_To_v1_PortworxVolumeSource,
- Convert_v1_Preconditions_To_core_Preconditions,
- Convert_core_Preconditions_To_v1_Preconditions,
- Convert_v1_PreferAvoidPodsEntry_To_core_PreferAvoidPodsEntry,
- Convert_core_PreferAvoidPodsEntry_To_v1_PreferAvoidPodsEntry,
- Convert_v1_PreferredSchedulingTerm_To_core_PreferredSchedulingTerm,
- Convert_core_PreferredSchedulingTerm_To_v1_PreferredSchedulingTerm,
- Convert_v1_Probe_To_core_Probe,
- Convert_core_Probe_To_v1_Probe,
- Convert_v1_ProjectedVolumeSource_To_core_ProjectedVolumeSource,
- Convert_core_ProjectedVolumeSource_To_v1_ProjectedVolumeSource,
- Convert_v1_QuobyteVolumeSource_To_core_QuobyteVolumeSource,
- Convert_core_QuobyteVolumeSource_To_v1_QuobyteVolumeSource,
- Convert_v1_RBDPersistentVolumeSource_To_core_RBDPersistentVolumeSource,
- Convert_core_RBDPersistentVolumeSource_To_v1_RBDPersistentVolumeSource,
- Convert_v1_RBDVolumeSource_To_core_RBDVolumeSource,
- Convert_core_RBDVolumeSource_To_v1_RBDVolumeSource,
- Convert_v1_RangeAllocation_To_core_RangeAllocation,
- Convert_core_RangeAllocation_To_v1_RangeAllocation,
- Convert_v1_ReplicationController_To_core_ReplicationController,
- Convert_core_ReplicationController_To_v1_ReplicationController,
- Convert_v1_ReplicationControllerCondition_To_core_ReplicationControllerCondition,
- Convert_core_ReplicationControllerCondition_To_v1_ReplicationControllerCondition,
- Convert_v1_ReplicationControllerList_To_core_ReplicationControllerList,
- Convert_core_ReplicationControllerList_To_v1_ReplicationControllerList,
- Convert_v1_ReplicationControllerSpec_To_core_ReplicationControllerSpec,
- Convert_core_ReplicationControllerSpec_To_v1_ReplicationControllerSpec,
- Convert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus,
- Convert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus,
- Convert_v1_ResourceFieldSelector_To_core_ResourceFieldSelector,
- Convert_core_ResourceFieldSelector_To_v1_ResourceFieldSelector,
- Convert_v1_ResourceQuota_To_core_ResourceQuota,
- Convert_core_ResourceQuota_To_v1_ResourceQuota,
- Convert_v1_ResourceQuotaList_To_core_ResourceQuotaList,
- Convert_core_ResourceQuotaList_To_v1_ResourceQuotaList,
- Convert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec,
- Convert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec,
- Convert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus,
- Convert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus,
- Convert_v1_ResourceRequirements_To_core_ResourceRequirements,
- Convert_core_ResourceRequirements_To_v1_ResourceRequirements,
- Convert_v1_SELinuxOptions_To_core_SELinuxOptions,
- Convert_core_SELinuxOptions_To_v1_SELinuxOptions,
- Convert_v1_ScaleIOPersistentVolumeSource_To_core_ScaleIOPersistentVolumeSource,
- Convert_core_ScaleIOPersistentVolumeSource_To_v1_ScaleIOPersistentVolumeSource,
- Convert_v1_ScaleIOVolumeSource_To_core_ScaleIOVolumeSource,
- Convert_core_ScaleIOVolumeSource_To_v1_ScaleIOVolumeSource,
- Convert_v1_Secret_To_core_Secret,
- Convert_core_Secret_To_v1_Secret,
- Convert_v1_SecretEnvSource_To_core_SecretEnvSource,
- Convert_core_SecretEnvSource_To_v1_SecretEnvSource,
- Convert_v1_SecretKeySelector_To_core_SecretKeySelector,
- Convert_core_SecretKeySelector_To_v1_SecretKeySelector,
- Convert_v1_SecretList_To_core_SecretList,
- Convert_core_SecretList_To_v1_SecretList,
- Convert_v1_SecretProjection_To_core_SecretProjection,
- Convert_core_SecretProjection_To_v1_SecretProjection,
- Convert_v1_SecretReference_To_core_SecretReference,
- Convert_core_SecretReference_To_v1_SecretReference,
- Convert_v1_SecretVolumeSource_To_core_SecretVolumeSource,
- Convert_core_SecretVolumeSource_To_v1_SecretVolumeSource,
- Convert_v1_SecurityContext_To_core_SecurityContext,
- Convert_core_SecurityContext_To_v1_SecurityContext,
- Convert_v1_SerializedReference_To_core_SerializedReference,
- Convert_core_SerializedReference_To_v1_SerializedReference,
- Convert_v1_Service_To_core_Service,
- Convert_core_Service_To_v1_Service,
- Convert_v1_ServiceAccount_To_core_ServiceAccount,
- Convert_core_ServiceAccount_To_v1_ServiceAccount,
- Convert_v1_ServiceAccountList_To_core_ServiceAccountList,
- Convert_core_ServiceAccountList_To_v1_ServiceAccountList,
- Convert_v1_ServiceList_To_core_ServiceList,
- Convert_core_ServiceList_To_v1_ServiceList,
- Convert_v1_ServicePort_To_core_ServicePort,
- Convert_core_ServicePort_To_v1_ServicePort,
- Convert_v1_ServiceProxyOptions_To_core_ServiceProxyOptions,
- Convert_core_ServiceProxyOptions_To_v1_ServiceProxyOptions,
- Convert_v1_ServiceSpec_To_core_ServiceSpec,
- Convert_core_ServiceSpec_To_v1_ServiceSpec,
- Convert_v1_ServiceStatus_To_core_ServiceStatus,
- Convert_core_ServiceStatus_To_v1_ServiceStatus,
- Convert_v1_SessionAffinityConfig_To_core_SessionAffinityConfig,
- Convert_core_SessionAffinityConfig_To_v1_SessionAffinityConfig,
- Convert_v1_StorageOSPersistentVolumeSource_To_core_StorageOSPersistentVolumeSource,
- Convert_core_StorageOSPersistentVolumeSource_To_v1_StorageOSPersistentVolumeSource,
- Convert_v1_StorageOSVolumeSource_To_core_StorageOSVolumeSource,
- Convert_core_StorageOSVolumeSource_To_v1_StorageOSVolumeSource,
- Convert_v1_Sysctl_To_core_Sysctl,
- Convert_core_Sysctl_To_v1_Sysctl,
- Convert_v1_TCPSocketAction_To_core_TCPSocketAction,
- Convert_core_TCPSocketAction_To_v1_TCPSocketAction,
- Convert_v1_Taint_To_core_Taint,
- Convert_core_Taint_To_v1_Taint,
- Convert_v1_Toleration_To_core_Toleration,
- Convert_core_Toleration_To_v1_Toleration,
- Convert_v1_Volume_To_core_Volume,
- Convert_core_Volume_To_v1_Volume,
- Convert_v1_VolumeDevice_To_core_VolumeDevice,
- Convert_core_VolumeDevice_To_v1_VolumeDevice,
- Convert_v1_VolumeMount_To_core_VolumeMount,
- Convert_core_VolumeMount_To_v1_VolumeMount,
- Convert_v1_VolumeNodeAffinity_To_core_VolumeNodeAffinity,
- Convert_core_VolumeNodeAffinity_To_v1_VolumeNodeAffinity,
- Convert_v1_VolumeProjection_To_core_VolumeProjection,
- Convert_core_VolumeProjection_To_v1_VolumeProjection,
- Convert_v1_VolumeSource_To_core_VolumeSource,
- Convert_core_VolumeSource_To_v1_VolumeSource,
- Convert_v1_VsphereVirtualDiskVolumeSource_To_core_VsphereVirtualDiskVolumeSource,
- Convert_core_VsphereVirtualDiskVolumeSource_To_v1_VsphereVirtualDiskVolumeSource,
- Convert_v1_WeightedPodAffinityTerm_To_core_WeightedPodAffinityTerm,
- Convert_core_WeightedPodAffinityTerm_To_v1_WeightedPodAffinityTerm,
- )
-}
-
-func autoConvert_v1_AWSElasticBlockStoreVolumeSource_To_core_AWSElasticBlockStoreVolumeSource(in *v1.AWSElasticBlockStoreVolumeSource, out *core.AWSElasticBlockStoreVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.Partition = in.Partition
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_AWSElasticBlockStoreVolumeSource_To_core_AWSElasticBlockStoreVolumeSource is an autogenerated conversion function.
-func Convert_v1_AWSElasticBlockStoreVolumeSource_To_core_AWSElasticBlockStoreVolumeSource(in *v1.AWSElasticBlockStoreVolumeSource, out *core.AWSElasticBlockStoreVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_AWSElasticBlockStoreVolumeSource_To_core_AWSElasticBlockStoreVolumeSource(in, out, s)
-}
-
-func autoConvert_core_AWSElasticBlockStoreVolumeSource_To_v1_AWSElasticBlockStoreVolumeSource(in *core.AWSElasticBlockStoreVolumeSource, out *v1.AWSElasticBlockStoreVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.Partition = in.Partition
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_AWSElasticBlockStoreVolumeSource_To_v1_AWSElasticBlockStoreVolumeSource is an autogenerated conversion function.
-func Convert_core_AWSElasticBlockStoreVolumeSource_To_v1_AWSElasticBlockStoreVolumeSource(in *core.AWSElasticBlockStoreVolumeSource, out *v1.AWSElasticBlockStoreVolumeSource, s conversion.Scope) error {
- return autoConvert_core_AWSElasticBlockStoreVolumeSource_To_v1_AWSElasticBlockStoreVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Affinity_To_core_Affinity(in *v1.Affinity, out *core.Affinity, s conversion.Scope) error {
- out.NodeAffinity = (*core.NodeAffinity)(unsafe.Pointer(in.NodeAffinity))
- out.PodAffinity = (*core.PodAffinity)(unsafe.Pointer(in.PodAffinity))
- out.PodAntiAffinity = (*core.PodAntiAffinity)(unsafe.Pointer(in.PodAntiAffinity))
- return nil
-}
-
-// Convert_v1_Affinity_To_core_Affinity is an autogenerated conversion function.
-func Convert_v1_Affinity_To_core_Affinity(in *v1.Affinity, out *core.Affinity, s conversion.Scope) error {
- return autoConvert_v1_Affinity_To_core_Affinity(in, out, s)
-}
-
-func autoConvert_core_Affinity_To_v1_Affinity(in *core.Affinity, out *v1.Affinity, s conversion.Scope) error {
- out.NodeAffinity = (*v1.NodeAffinity)(unsafe.Pointer(in.NodeAffinity))
- out.PodAffinity = (*v1.PodAffinity)(unsafe.Pointer(in.PodAffinity))
- out.PodAntiAffinity = (*v1.PodAntiAffinity)(unsafe.Pointer(in.PodAntiAffinity))
- return nil
-}
-
-// Convert_core_Affinity_To_v1_Affinity is an autogenerated conversion function.
-func Convert_core_Affinity_To_v1_Affinity(in *core.Affinity, out *v1.Affinity, s conversion.Scope) error {
- return autoConvert_core_Affinity_To_v1_Affinity(in, out, s)
-}
-
-func autoConvert_v1_AttachedVolume_To_core_AttachedVolume(in *v1.AttachedVolume, out *core.AttachedVolume, s conversion.Scope) error {
- out.Name = core.UniqueVolumeName(in.Name)
- out.DevicePath = in.DevicePath
- return nil
-}
-
-// Convert_v1_AttachedVolume_To_core_AttachedVolume is an autogenerated conversion function.
-func Convert_v1_AttachedVolume_To_core_AttachedVolume(in *v1.AttachedVolume, out *core.AttachedVolume, s conversion.Scope) error {
- return autoConvert_v1_AttachedVolume_To_core_AttachedVolume(in, out, s)
-}
-
-func autoConvert_core_AttachedVolume_To_v1_AttachedVolume(in *core.AttachedVolume, out *v1.AttachedVolume, s conversion.Scope) error {
- out.Name = v1.UniqueVolumeName(in.Name)
- out.DevicePath = in.DevicePath
- return nil
-}
-
-// Convert_core_AttachedVolume_To_v1_AttachedVolume is an autogenerated conversion function.
-func Convert_core_AttachedVolume_To_v1_AttachedVolume(in *core.AttachedVolume, out *v1.AttachedVolume, s conversion.Scope) error {
- return autoConvert_core_AttachedVolume_To_v1_AttachedVolume(in, out, s)
-}
-
-func autoConvert_v1_AvoidPods_To_core_AvoidPods(in *v1.AvoidPods, out *core.AvoidPods, s conversion.Scope) error {
- out.PreferAvoidPods = *(*[]core.PreferAvoidPodsEntry)(unsafe.Pointer(&in.PreferAvoidPods))
- return nil
-}
-
-// Convert_v1_AvoidPods_To_core_AvoidPods is an autogenerated conversion function.
-func Convert_v1_AvoidPods_To_core_AvoidPods(in *v1.AvoidPods, out *core.AvoidPods, s conversion.Scope) error {
- return autoConvert_v1_AvoidPods_To_core_AvoidPods(in, out, s)
-}
-
-func autoConvert_core_AvoidPods_To_v1_AvoidPods(in *core.AvoidPods, out *v1.AvoidPods, s conversion.Scope) error {
- out.PreferAvoidPods = *(*[]v1.PreferAvoidPodsEntry)(unsafe.Pointer(&in.PreferAvoidPods))
- return nil
-}
-
-// Convert_core_AvoidPods_To_v1_AvoidPods is an autogenerated conversion function.
-func Convert_core_AvoidPods_To_v1_AvoidPods(in *core.AvoidPods, out *v1.AvoidPods, s conversion.Scope) error {
- return autoConvert_core_AvoidPods_To_v1_AvoidPods(in, out, s)
-}
-
-func autoConvert_v1_AzureDiskVolumeSource_To_core_AzureDiskVolumeSource(in *v1.AzureDiskVolumeSource, out *core.AzureDiskVolumeSource, s conversion.Scope) error {
- out.DiskName = in.DiskName
- out.DataDiskURI = in.DataDiskURI
- out.CachingMode = (*core.AzureDataDiskCachingMode)(unsafe.Pointer(in.CachingMode))
- out.FSType = (*string)(unsafe.Pointer(in.FSType))
- out.ReadOnly = (*bool)(unsafe.Pointer(in.ReadOnly))
- out.Kind = (*core.AzureDataDiskKind)(unsafe.Pointer(in.Kind))
- return nil
-}
-
-// Convert_v1_AzureDiskVolumeSource_To_core_AzureDiskVolumeSource is an autogenerated conversion function.
-func Convert_v1_AzureDiskVolumeSource_To_core_AzureDiskVolumeSource(in *v1.AzureDiskVolumeSource, out *core.AzureDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_AzureDiskVolumeSource_To_core_AzureDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_core_AzureDiskVolumeSource_To_v1_AzureDiskVolumeSource(in *core.AzureDiskVolumeSource, out *v1.AzureDiskVolumeSource, s conversion.Scope) error {
- out.DiskName = in.DiskName
- out.DataDiskURI = in.DataDiskURI
- out.CachingMode = (*v1.AzureDataDiskCachingMode)(unsafe.Pointer(in.CachingMode))
- out.FSType = (*string)(unsafe.Pointer(in.FSType))
- out.ReadOnly = (*bool)(unsafe.Pointer(in.ReadOnly))
- out.Kind = (*v1.AzureDataDiskKind)(unsafe.Pointer(in.Kind))
- return nil
-}
-
-// Convert_core_AzureDiskVolumeSource_To_v1_AzureDiskVolumeSource is an autogenerated conversion function.
-func Convert_core_AzureDiskVolumeSource_To_v1_AzureDiskVolumeSource(in *core.AzureDiskVolumeSource, out *v1.AzureDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_core_AzureDiskVolumeSource_To_v1_AzureDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_AzureFilePersistentVolumeSource_To_core_AzureFilePersistentVolumeSource(in *v1.AzureFilePersistentVolumeSource, out *core.AzureFilePersistentVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.ShareName = in.ShareName
- out.ReadOnly = in.ReadOnly
- out.SecretNamespace = (*string)(unsafe.Pointer(in.SecretNamespace))
- return nil
-}
-
-// Convert_v1_AzureFilePersistentVolumeSource_To_core_AzureFilePersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_AzureFilePersistentVolumeSource_To_core_AzureFilePersistentVolumeSource(in *v1.AzureFilePersistentVolumeSource, out *core.AzureFilePersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_AzureFilePersistentVolumeSource_To_core_AzureFilePersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_AzureFilePersistentVolumeSource_To_v1_AzureFilePersistentVolumeSource(in *core.AzureFilePersistentVolumeSource, out *v1.AzureFilePersistentVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.ShareName = in.ShareName
- out.ReadOnly = in.ReadOnly
- out.SecretNamespace = (*string)(unsafe.Pointer(in.SecretNamespace))
- return nil
-}
-
-// Convert_core_AzureFilePersistentVolumeSource_To_v1_AzureFilePersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_AzureFilePersistentVolumeSource_To_v1_AzureFilePersistentVolumeSource(in *core.AzureFilePersistentVolumeSource, out *v1.AzureFilePersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_AzureFilePersistentVolumeSource_To_v1_AzureFilePersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_AzureFileVolumeSource_To_core_AzureFileVolumeSource(in *v1.AzureFileVolumeSource, out *core.AzureFileVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.ShareName = in.ShareName
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_AzureFileVolumeSource_To_core_AzureFileVolumeSource is an autogenerated conversion function.
-func Convert_v1_AzureFileVolumeSource_To_core_AzureFileVolumeSource(in *v1.AzureFileVolumeSource, out *core.AzureFileVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_AzureFileVolumeSource_To_core_AzureFileVolumeSource(in, out, s)
-}
-
-func autoConvert_core_AzureFileVolumeSource_To_v1_AzureFileVolumeSource(in *core.AzureFileVolumeSource, out *v1.AzureFileVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.ShareName = in.ShareName
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_AzureFileVolumeSource_To_v1_AzureFileVolumeSource is an autogenerated conversion function.
-func Convert_core_AzureFileVolumeSource_To_v1_AzureFileVolumeSource(in *core.AzureFileVolumeSource, out *v1.AzureFileVolumeSource, s conversion.Scope) error {
- return autoConvert_core_AzureFileVolumeSource_To_v1_AzureFileVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Binding_To_core_Binding(in *v1.Binding, out *core.Binding, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ObjectReference_To_core_ObjectReference(&in.Target, &out.Target, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_Binding_To_core_Binding is an autogenerated conversion function.
-func Convert_v1_Binding_To_core_Binding(in *v1.Binding, out *core.Binding, s conversion.Scope) error {
- return autoConvert_v1_Binding_To_core_Binding(in, out, s)
-}
-
-func autoConvert_core_Binding_To_v1_Binding(in *core.Binding, out *v1.Binding, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_ObjectReference_To_v1_ObjectReference(&in.Target, &out.Target, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_Binding_To_v1_Binding is an autogenerated conversion function.
-func Convert_core_Binding_To_v1_Binding(in *core.Binding, out *v1.Binding, s conversion.Scope) error {
- return autoConvert_core_Binding_To_v1_Binding(in, out, s)
-}
-
-func autoConvert_v1_CSIPersistentVolumeSource_To_core_CSIPersistentVolumeSource(in *v1.CSIPersistentVolumeSource, out *core.CSIPersistentVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.VolumeHandle = in.VolumeHandle
- out.ReadOnly = in.ReadOnly
- out.FSType = in.FSType
- out.VolumeAttributes = *(*map[string]string)(unsafe.Pointer(&in.VolumeAttributes))
- out.ControllerPublishSecretRef = (*core.SecretReference)(unsafe.Pointer(in.ControllerPublishSecretRef))
- out.NodeStageSecretRef = (*core.SecretReference)(unsafe.Pointer(in.NodeStageSecretRef))
- out.NodePublishSecretRef = (*core.SecretReference)(unsafe.Pointer(in.NodePublishSecretRef))
- return nil
-}
-
-// Convert_v1_CSIPersistentVolumeSource_To_core_CSIPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_CSIPersistentVolumeSource_To_core_CSIPersistentVolumeSource(in *v1.CSIPersistentVolumeSource, out *core.CSIPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_CSIPersistentVolumeSource_To_core_CSIPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_CSIPersistentVolumeSource_To_v1_CSIPersistentVolumeSource(in *core.CSIPersistentVolumeSource, out *v1.CSIPersistentVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.VolumeHandle = in.VolumeHandle
- out.ReadOnly = in.ReadOnly
- out.FSType = in.FSType
- out.VolumeAttributes = *(*map[string]string)(unsafe.Pointer(&in.VolumeAttributes))
- out.ControllerPublishSecretRef = (*v1.SecretReference)(unsafe.Pointer(in.ControllerPublishSecretRef))
- out.NodeStageSecretRef = (*v1.SecretReference)(unsafe.Pointer(in.NodeStageSecretRef))
- out.NodePublishSecretRef = (*v1.SecretReference)(unsafe.Pointer(in.NodePublishSecretRef))
- return nil
-}
-
-// Convert_core_CSIPersistentVolumeSource_To_v1_CSIPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_CSIPersistentVolumeSource_To_v1_CSIPersistentVolumeSource(in *core.CSIPersistentVolumeSource, out *v1.CSIPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_CSIPersistentVolumeSource_To_v1_CSIPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Capabilities_To_core_Capabilities(in *v1.Capabilities, out *core.Capabilities, s conversion.Scope) error {
- out.Add = *(*[]core.Capability)(unsafe.Pointer(&in.Add))
- out.Drop = *(*[]core.Capability)(unsafe.Pointer(&in.Drop))
- return nil
-}
-
-// Convert_v1_Capabilities_To_core_Capabilities is an autogenerated conversion function.
-func Convert_v1_Capabilities_To_core_Capabilities(in *v1.Capabilities, out *core.Capabilities, s conversion.Scope) error {
- return autoConvert_v1_Capabilities_To_core_Capabilities(in, out, s)
-}
-
-func autoConvert_core_Capabilities_To_v1_Capabilities(in *core.Capabilities, out *v1.Capabilities, s conversion.Scope) error {
- out.Add = *(*[]v1.Capability)(unsafe.Pointer(&in.Add))
- out.Drop = *(*[]v1.Capability)(unsafe.Pointer(&in.Drop))
- return nil
-}
-
-// Convert_core_Capabilities_To_v1_Capabilities is an autogenerated conversion function.
-func Convert_core_Capabilities_To_v1_Capabilities(in *core.Capabilities, out *v1.Capabilities, s conversion.Scope) error {
- return autoConvert_core_Capabilities_To_v1_Capabilities(in, out, s)
-}
-
-func autoConvert_v1_CephFSPersistentVolumeSource_To_core_CephFSPersistentVolumeSource(in *v1.CephFSPersistentVolumeSource, out *core.CephFSPersistentVolumeSource, s conversion.Scope) error {
- out.Monitors = *(*[]string)(unsafe.Pointer(&in.Monitors))
- out.Path = in.Path
- out.User = in.User
- out.SecretFile = in.SecretFile
- out.SecretRef = (*core.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_CephFSPersistentVolumeSource_To_core_CephFSPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_CephFSPersistentVolumeSource_To_core_CephFSPersistentVolumeSource(in *v1.CephFSPersistentVolumeSource, out *core.CephFSPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_CephFSPersistentVolumeSource_To_core_CephFSPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_CephFSPersistentVolumeSource_To_v1_CephFSPersistentVolumeSource(in *core.CephFSPersistentVolumeSource, out *v1.CephFSPersistentVolumeSource, s conversion.Scope) error {
- out.Monitors = *(*[]string)(unsafe.Pointer(&in.Monitors))
- out.Path = in.Path
- out.User = in.User
- out.SecretFile = in.SecretFile
- out.SecretRef = (*v1.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_CephFSPersistentVolumeSource_To_v1_CephFSPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_CephFSPersistentVolumeSource_To_v1_CephFSPersistentVolumeSource(in *core.CephFSPersistentVolumeSource, out *v1.CephFSPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_CephFSPersistentVolumeSource_To_v1_CephFSPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_CephFSVolumeSource_To_core_CephFSVolumeSource(in *v1.CephFSVolumeSource, out *core.CephFSVolumeSource, s conversion.Scope) error {
- out.Monitors = *(*[]string)(unsafe.Pointer(&in.Monitors))
- out.Path = in.Path
- out.User = in.User
- out.SecretFile = in.SecretFile
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_CephFSVolumeSource_To_core_CephFSVolumeSource is an autogenerated conversion function.
-func Convert_v1_CephFSVolumeSource_To_core_CephFSVolumeSource(in *v1.CephFSVolumeSource, out *core.CephFSVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_CephFSVolumeSource_To_core_CephFSVolumeSource(in, out, s)
-}
-
-func autoConvert_core_CephFSVolumeSource_To_v1_CephFSVolumeSource(in *core.CephFSVolumeSource, out *v1.CephFSVolumeSource, s conversion.Scope) error {
- out.Monitors = *(*[]string)(unsafe.Pointer(&in.Monitors))
- out.Path = in.Path
- out.User = in.User
- out.SecretFile = in.SecretFile
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_CephFSVolumeSource_To_v1_CephFSVolumeSource is an autogenerated conversion function.
-func Convert_core_CephFSVolumeSource_To_v1_CephFSVolumeSource(in *core.CephFSVolumeSource, out *v1.CephFSVolumeSource, s conversion.Scope) error {
- return autoConvert_core_CephFSVolumeSource_To_v1_CephFSVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_CinderVolumeSource_To_core_CinderVolumeSource(in *v1.CinderVolumeSource, out *core.CinderVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_CinderVolumeSource_To_core_CinderVolumeSource is an autogenerated conversion function.
-func Convert_v1_CinderVolumeSource_To_core_CinderVolumeSource(in *v1.CinderVolumeSource, out *core.CinderVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_CinderVolumeSource_To_core_CinderVolumeSource(in, out, s)
-}
-
-func autoConvert_core_CinderVolumeSource_To_v1_CinderVolumeSource(in *core.CinderVolumeSource, out *v1.CinderVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_CinderVolumeSource_To_v1_CinderVolumeSource is an autogenerated conversion function.
-func Convert_core_CinderVolumeSource_To_v1_CinderVolumeSource(in *core.CinderVolumeSource, out *v1.CinderVolumeSource, s conversion.Scope) error {
- return autoConvert_core_CinderVolumeSource_To_v1_CinderVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_ClientIPConfig_To_core_ClientIPConfig(in *v1.ClientIPConfig, out *core.ClientIPConfig, s conversion.Scope) error {
- out.TimeoutSeconds = (*int32)(unsafe.Pointer(in.TimeoutSeconds))
- return nil
-}
-
-// Convert_v1_ClientIPConfig_To_core_ClientIPConfig is an autogenerated conversion function.
-func Convert_v1_ClientIPConfig_To_core_ClientIPConfig(in *v1.ClientIPConfig, out *core.ClientIPConfig, s conversion.Scope) error {
- return autoConvert_v1_ClientIPConfig_To_core_ClientIPConfig(in, out, s)
-}
-
-func autoConvert_core_ClientIPConfig_To_v1_ClientIPConfig(in *core.ClientIPConfig, out *v1.ClientIPConfig, s conversion.Scope) error {
- out.TimeoutSeconds = (*int32)(unsafe.Pointer(in.TimeoutSeconds))
- return nil
-}
-
-// Convert_core_ClientIPConfig_To_v1_ClientIPConfig is an autogenerated conversion function.
-func Convert_core_ClientIPConfig_To_v1_ClientIPConfig(in *core.ClientIPConfig, out *v1.ClientIPConfig, s conversion.Scope) error {
- return autoConvert_core_ClientIPConfig_To_v1_ClientIPConfig(in, out, s)
-}
-
-func autoConvert_v1_ComponentCondition_To_core_ComponentCondition(in *v1.ComponentCondition, out *core.ComponentCondition, s conversion.Scope) error {
- out.Type = core.ComponentConditionType(in.Type)
- out.Status = core.ConditionStatus(in.Status)
- out.Message = in.Message
- out.Error = in.Error
- return nil
-}
-
-// Convert_v1_ComponentCondition_To_core_ComponentCondition is an autogenerated conversion function.
-func Convert_v1_ComponentCondition_To_core_ComponentCondition(in *v1.ComponentCondition, out *core.ComponentCondition, s conversion.Scope) error {
- return autoConvert_v1_ComponentCondition_To_core_ComponentCondition(in, out, s)
-}
-
-func autoConvert_core_ComponentCondition_To_v1_ComponentCondition(in *core.ComponentCondition, out *v1.ComponentCondition, s conversion.Scope) error {
- out.Type = v1.ComponentConditionType(in.Type)
- out.Status = v1.ConditionStatus(in.Status)
- out.Message = in.Message
- out.Error = in.Error
- return nil
-}
-
-// Convert_core_ComponentCondition_To_v1_ComponentCondition is an autogenerated conversion function.
-func Convert_core_ComponentCondition_To_v1_ComponentCondition(in *core.ComponentCondition, out *v1.ComponentCondition, s conversion.Scope) error {
- return autoConvert_core_ComponentCondition_To_v1_ComponentCondition(in, out, s)
-}
-
-func autoConvert_v1_ComponentStatus_To_core_ComponentStatus(in *v1.ComponentStatus, out *core.ComponentStatus, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Conditions = *(*[]core.ComponentCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_v1_ComponentStatus_To_core_ComponentStatus is an autogenerated conversion function.
-func Convert_v1_ComponentStatus_To_core_ComponentStatus(in *v1.ComponentStatus, out *core.ComponentStatus, s conversion.Scope) error {
- return autoConvert_v1_ComponentStatus_To_core_ComponentStatus(in, out, s)
-}
-
-func autoConvert_core_ComponentStatus_To_v1_ComponentStatus(in *core.ComponentStatus, out *v1.ComponentStatus, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Conditions = *(*[]v1.ComponentCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_core_ComponentStatus_To_v1_ComponentStatus is an autogenerated conversion function.
-func Convert_core_ComponentStatus_To_v1_ComponentStatus(in *core.ComponentStatus, out *v1.ComponentStatus, s conversion.Scope) error {
- return autoConvert_core_ComponentStatus_To_v1_ComponentStatus(in, out, s)
-}
-
-func autoConvert_v1_ComponentStatusList_To_core_ComponentStatusList(in *v1.ComponentStatusList, out *core.ComponentStatusList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.ComponentStatus)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_ComponentStatusList_To_core_ComponentStatusList is an autogenerated conversion function.
-func Convert_v1_ComponentStatusList_To_core_ComponentStatusList(in *v1.ComponentStatusList, out *core.ComponentStatusList, s conversion.Scope) error {
- return autoConvert_v1_ComponentStatusList_To_core_ComponentStatusList(in, out, s)
-}
-
-func autoConvert_core_ComponentStatusList_To_v1_ComponentStatusList(in *core.ComponentStatusList, out *v1.ComponentStatusList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.ComponentStatus)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_ComponentStatusList_To_v1_ComponentStatusList is an autogenerated conversion function.
-func Convert_core_ComponentStatusList_To_v1_ComponentStatusList(in *core.ComponentStatusList, out *v1.ComponentStatusList, s conversion.Scope) error {
- return autoConvert_core_ComponentStatusList_To_v1_ComponentStatusList(in, out, s)
-}
-
-func autoConvert_v1_ConfigMap_To_core_ConfigMap(in *v1.ConfigMap, out *core.ConfigMap, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Data = *(*map[string]string)(unsafe.Pointer(&in.Data))
- out.BinaryData = *(*map[string][]byte)(unsafe.Pointer(&in.BinaryData))
- return nil
-}
-
-// Convert_v1_ConfigMap_To_core_ConfigMap is an autogenerated conversion function.
-func Convert_v1_ConfigMap_To_core_ConfigMap(in *v1.ConfigMap, out *core.ConfigMap, s conversion.Scope) error {
- return autoConvert_v1_ConfigMap_To_core_ConfigMap(in, out, s)
-}
-
-func autoConvert_core_ConfigMap_To_v1_ConfigMap(in *core.ConfigMap, out *v1.ConfigMap, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Data = *(*map[string]string)(unsafe.Pointer(&in.Data))
- out.BinaryData = *(*map[string][]byte)(unsafe.Pointer(&in.BinaryData))
- return nil
-}
-
-// Convert_core_ConfigMap_To_v1_ConfigMap is an autogenerated conversion function.
-func Convert_core_ConfigMap_To_v1_ConfigMap(in *core.ConfigMap, out *v1.ConfigMap, s conversion.Scope) error {
- return autoConvert_core_ConfigMap_To_v1_ConfigMap(in, out, s)
-}
-
-func autoConvert_v1_ConfigMapEnvSource_To_core_ConfigMapEnvSource(in *v1.ConfigMapEnvSource, out *core.ConfigMapEnvSource, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_ConfigMapEnvSource_To_core_ConfigMapEnvSource is an autogenerated conversion function.
-func Convert_v1_ConfigMapEnvSource_To_core_ConfigMapEnvSource(in *v1.ConfigMapEnvSource, out *core.ConfigMapEnvSource, s conversion.Scope) error {
- return autoConvert_v1_ConfigMapEnvSource_To_core_ConfigMapEnvSource(in, out, s)
-}
-
-func autoConvert_core_ConfigMapEnvSource_To_v1_ConfigMapEnvSource(in *core.ConfigMapEnvSource, out *v1.ConfigMapEnvSource, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_ConfigMapEnvSource_To_v1_ConfigMapEnvSource is an autogenerated conversion function.
-func Convert_core_ConfigMapEnvSource_To_v1_ConfigMapEnvSource(in *core.ConfigMapEnvSource, out *v1.ConfigMapEnvSource, s conversion.Scope) error {
- return autoConvert_core_ConfigMapEnvSource_To_v1_ConfigMapEnvSource(in, out, s)
-}
-
-func autoConvert_v1_ConfigMapKeySelector_To_core_ConfigMapKeySelector(in *v1.ConfigMapKeySelector, out *core.ConfigMapKeySelector, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Key = in.Key
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_ConfigMapKeySelector_To_core_ConfigMapKeySelector is an autogenerated conversion function.
-func Convert_v1_ConfigMapKeySelector_To_core_ConfigMapKeySelector(in *v1.ConfigMapKeySelector, out *core.ConfigMapKeySelector, s conversion.Scope) error {
- return autoConvert_v1_ConfigMapKeySelector_To_core_ConfigMapKeySelector(in, out, s)
-}
-
-func autoConvert_core_ConfigMapKeySelector_To_v1_ConfigMapKeySelector(in *core.ConfigMapKeySelector, out *v1.ConfigMapKeySelector, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Key = in.Key
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_ConfigMapKeySelector_To_v1_ConfigMapKeySelector is an autogenerated conversion function.
-func Convert_core_ConfigMapKeySelector_To_v1_ConfigMapKeySelector(in *core.ConfigMapKeySelector, out *v1.ConfigMapKeySelector, s conversion.Scope) error {
- return autoConvert_core_ConfigMapKeySelector_To_v1_ConfigMapKeySelector(in, out, s)
-}
-
-func autoConvert_v1_ConfigMapList_To_core_ConfigMapList(in *v1.ConfigMapList, out *core.ConfigMapList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.ConfigMap)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_ConfigMapList_To_core_ConfigMapList is an autogenerated conversion function.
-func Convert_v1_ConfigMapList_To_core_ConfigMapList(in *v1.ConfigMapList, out *core.ConfigMapList, s conversion.Scope) error {
- return autoConvert_v1_ConfigMapList_To_core_ConfigMapList(in, out, s)
-}
-
-func autoConvert_core_ConfigMapList_To_v1_ConfigMapList(in *core.ConfigMapList, out *v1.ConfigMapList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.ConfigMap)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_ConfigMapList_To_v1_ConfigMapList is an autogenerated conversion function.
-func Convert_core_ConfigMapList_To_v1_ConfigMapList(in *core.ConfigMapList, out *v1.ConfigMapList, s conversion.Scope) error {
- return autoConvert_core_ConfigMapList_To_v1_ConfigMapList(in, out, s)
-}
-
-func autoConvert_v1_ConfigMapProjection_To_core_ConfigMapProjection(in *v1.ConfigMapProjection, out *core.ConfigMapProjection, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]core.KeyToPath)(unsafe.Pointer(&in.Items))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_ConfigMapProjection_To_core_ConfigMapProjection is an autogenerated conversion function.
-func Convert_v1_ConfigMapProjection_To_core_ConfigMapProjection(in *v1.ConfigMapProjection, out *core.ConfigMapProjection, s conversion.Scope) error {
- return autoConvert_v1_ConfigMapProjection_To_core_ConfigMapProjection(in, out, s)
-}
-
-func autoConvert_core_ConfigMapProjection_To_v1_ConfigMapProjection(in *core.ConfigMapProjection, out *v1.ConfigMapProjection, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]v1.KeyToPath)(unsafe.Pointer(&in.Items))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_ConfigMapProjection_To_v1_ConfigMapProjection is an autogenerated conversion function.
-func Convert_core_ConfigMapProjection_To_v1_ConfigMapProjection(in *core.ConfigMapProjection, out *v1.ConfigMapProjection, s conversion.Scope) error {
- return autoConvert_core_ConfigMapProjection_To_v1_ConfigMapProjection(in, out, s)
-}
-
-func autoConvert_v1_ConfigMapVolumeSource_To_core_ConfigMapVolumeSource(in *v1.ConfigMapVolumeSource, out *core.ConfigMapVolumeSource, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]core.KeyToPath)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_ConfigMapVolumeSource_To_core_ConfigMapVolumeSource is an autogenerated conversion function.
-func Convert_v1_ConfigMapVolumeSource_To_core_ConfigMapVolumeSource(in *v1.ConfigMapVolumeSource, out *core.ConfigMapVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ConfigMapVolumeSource_To_core_ConfigMapVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ConfigMapVolumeSource_To_v1_ConfigMapVolumeSource(in *core.ConfigMapVolumeSource, out *v1.ConfigMapVolumeSource, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]v1.KeyToPath)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_ConfigMapVolumeSource_To_v1_ConfigMapVolumeSource is an autogenerated conversion function.
-func Convert_core_ConfigMapVolumeSource_To_v1_ConfigMapVolumeSource(in *core.ConfigMapVolumeSource, out *v1.ConfigMapVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ConfigMapVolumeSource_To_v1_ConfigMapVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Container_To_core_Container(in *v1.Container, out *core.Container, s conversion.Scope) error {
- out.Name = in.Name
- out.Image = in.Image
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- out.Args = *(*[]string)(unsafe.Pointer(&in.Args))
- out.WorkingDir = in.WorkingDir
- out.Ports = *(*[]core.ContainerPort)(unsafe.Pointer(&in.Ports))
- out.EnvFrom = *(*[]core.EnvFromSource)(unsafe.Pointer(&in.EnvFrom))
- out.Env = *(*[]core.EnvVar)(unsafe.Pointer(&in.Env))
- if err := Convert_v1_ResourceRequirements_To_core_ResourceRequirements(&in.Resources, &out.Resources, s); err != nil {
- return err
- }
- out.VolumeMounts = *(*[]core.VolumeMount)(unsafe.Pointer(&in.VolumeMounts))
- out.VolumeDevices = *(*[]core.VolumeDevice)(unsafe.Pointer(&in.VolumeDevices))
- out.LivenessProbe = (*core.Probe)(unsafe.Pointer(in.LivenessProbe))
- out.ReadinessProbe = (*core.Probe)(unsafe.Pointer(in.ReadinessProbe))
- out.Lifecycle = (*core.Lifecycle)(unsafe.Pointer(in.Lifecycle))
- out.TerminationMessagePath = in.TerminationMessagePath
- out.TerminationMessagePolicy = core.TerminationMessagePolicy(in.TerminationMessagePolicy)
- out.ImagePullPolicy = core.PullPolicy(in.ImagePullPolicy)
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- *out = new(core.SecurityContext)
- if err := Convert_v1_SecurityContext_To_core_SecurityContext(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.SecurityContext = nil
- }
- out.Stdin = in.Stdin
- out.StdinOnce = in.StdinOnce
- out.TTY = in.TTY
- return nil
-}
-
-// Convert_v1_Container_To_core_Container is an autogenerated conversion function.
-func Convert_v1_Container_To_core_Container(in *v1.Container, out *core.Container, s conversion.Scope) error {
- return autoConvert_v1_Container_To_core_Container(in, out, s)
-}
-
-func autoConvert_core_Container_To_v1_Container(in *core.Container, out *v1.Container, s conversion.Scope) error {
- out.Name = in.Name
- out.Image = in.Image
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- out.Args = *(*[]string)(unsafe.Pointer(&in.Args))
- out.WorkingDir = in.WorkingDir
- out.Ports = *(*[]v1.ContainerPort)(unsafe.Pointer(&in.Ports))
- out.EnvFrom = *(*[]v1.EnvFromSource)(unsafe.Pointer(&in.EnvFrom))
- out.Env = *(*[]v1.EnvVar)(unsafe.Pointer(&in.Env))
- if err := Convert_core_ResourceRequirements_To_v1_ResourceRequirements(&in.Resources, &out.Resources, s); err != nil {
- return err
- }
- out.VolumeMounts = *(*[]v1.VolumeMount)(unsafe.Pointer(&in.VolumeMounts))
- out.VolumeDevices = *(*[]v1.VolumeDevice)(unsafe.Pointer(&in.VolumeDevices))
- out.LivenessProbe = (*v1.Probe)(unsafe.Pointer(in.LivenessProbe))
- out.ReadinessProbe = (*v1.Probe)(unsafe.Pointer(in.ReadinessProbe))
- out.Lifecycle = (*v1.Lifecycle)(unsafe.Pointer(in.Lifecycle))
- out.TerminationMessagePath = in.TerminationMessagePath
- out.TerminationMessagePolicy = v1.TerminationMessagePolicy(in.TerminationMessagePolicy)
- out.ImagePullPolicy = v1.PullPolicy(in.ImagePullPolicy)
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- *out = new(v1.SecurityContext)
- if err := Convert_core_SecurityContext_To_v1_SecurityContext(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.SecurityContext = nil
- }
- out.Stdin = in.Stdin
- out.StdinOnce = in.StdinOnce
- out.TTY = in.TTY
- return nil
-}
-
-// Convert_core_Container_To_v1_Container is an autogenerated conversion function.
-func Convert_core_Container_To_v1_Container(in *core.Container, out *v1.Container, s conversion.Scope) error {
- return autoConvert_core_Container_To_v1_Container(in, out, s)
-}
-
-func autoConvert_v1_ContainerImage_To_core_ContainerImage(in *v1.ContainerImage, out *core.ContainerImage, s conversion.Scope) error {
- out.Names = *(*[]string)(unsafe.Pointer(&in.Names))
- out.SizeBytes = in.SizeBytes
- return nil
-}
-
-// Convert_v1_ContainerImage_To_core_ContainerImage is an autogenerated conversion function.
-func Convert_v1_ContainerImage_To_core_ContainerImage(in *v1.ContainerImage, out *core.ContainerImage, s conversion.Scope) error {
- return autoConvert_v1_ContainerImage_To_core_ContainerImage(in, out, s)
-}
-
-func autoConvert_core_ContainerImage_To_v1_ContainerImage(in *core.ContainerImage, out *v1.ContainerImage, s conversion.Scope) error {
- out.Names = *(*[]string)(unsafe.Pointer(&in.Names))
- out.SizeBytes = in.SizeBytes
- return nil
-}
-
-// Convert_core_ContainerImage_To_v1_ContainerImage is an autogenerated conversion function.
-func Convert_core_ContainerImage_To_v1_ContainerImage(in *core.ContainerImage, out *v1.ContainerImage, s conversion.Scope) error {
- return autoConvert_core_ContainerImage_To_v1_ContainerImage(in, out, s)
-}
-
-func autoConvert_v1_ContainerPort_To_core_ContainerPort(in *v1.ContainerPort, out *core.ContainerPort, s conversion.Scope) error {
- out.Name = in.Name
- out.HostPort = in.HostPort
- out.ContainerPort = in.ContainerPort
- out.Protocol = core.Protocol(in.Protocol)
- out.HostIP = in.HostIP
- return nil
-}
-
-// Convert_v1_ContainerPort_To_core_ContainerPort is an autogenerated conversion function.
-func Convert_v1_ContainerPort_To_core_ContainerPort(in *v1.ContainerPort, out *core.ContainerPort, s conversion.Scope) error {
- return autoConvert_v1_ContainerPort_To_core_ContainerPort(in, out, s)
-}
-
-func autoConvert_core_ContainerPort_To_v1_ContainerPort(in *core.ContainerPort, out *v1.ContainerPort, s conversion.Scope) error {
- out.Name = in.Name
- out.HostPort = in.HostPort
- out.ContainerPort = in.ContainerPort
- out.Protocol = v1.Protocol(in.Protocol)
- out.HostIP = in.HostIP
- return nil
-}
-
-// Convert_core_ContainerPort_To_v1_ContainerPort is an autogenerated conversion function.
-func Convert_core_ContainerPort_To_v1_ContainerPort(in *core.ContainerPort, out *v1.ContainerPort, s conversion.Scope) error {
- return autoConvert_core_ContainerPort_To_v1_ContainerPort(in, out, s)
-}
-
-func autoConvert_v1_ContainerState_To_core_ContainerState(in *v1.ContainerState, out *core.ContainerState, s conversion.Scope) error {
- out.Waiting = (*core.ContainerStateWaiting)(unsafe.Pointer(in.Waiting))
- out.Running = (*core.ContainerStateRunning)(unsafe.Pointer(in.Running))
- out.Terminated = (*core.ContainerStateTerminated)(unsafe.Pointer(in.Terminated))
- return nil
-}
-
-// Convert_v1_ContainerState_To_core_ContainerState is an autogenerated conversion function.
-func Convert_v1_ContainerState_To_core_ContainerState(in *v1.ContainerState, out *core.ContainerState, s conversion.Scope) error {
- return autoConvert_v1_ContainerState_To_core_ContainerState(in, out, s)
-}
-
-func autoConvert_core_ContainerState_To_v1_ContainerState(in *core.ContainerState, out *v1.ContainerState, s conversion.Scope) error {
- out.Waiting = (*v1.ContainerStateWaiting)(unsafe.Pointer(in.Waiting))
- out.Running = (*v1.ContainerStateRunning)(unsafe.Pointer(in.Running))
- out.Terminated = (*v1.ContainerStateTerminated)(unsafe.Pointer(in.Terminated))
- return nil
-}
-
-// Convert_core_ContainerState_To_v1_ContainerState is an autogenerated conversion function.
-func Convert_core_ContainerState_To_v1_ContainerState(in *core.ContainerState, out *v1.ContainerState, s conversion.Scope) error {
- return autoConvert_core_ContainerState_To_v1_ContainerState(in, out, s)
-}
-
-func autoConvert_v1_ContainerStateRunning_To_core_ContainerStateRunning(in *v1.ContainerStateRunning, out *core.ContainerStateRunning, s conversion.Scope) error {
- out.StartedAt = in.StartedAt
- return nil
-}
-
-// Convert_v1_ContainerStateRunning_To_core_ContainerStateRunning is an autogenerated conversion function.
-func Convert_v1_ContainerStateRunning_To_core_ContainerStateRunning(in *v1.ContainerStateRunning, out *core.ContainerStateRunning, s conversion.Scope) error {
- return autoConvert_v1_ContainerStateRunning_To_core_ContainerStateRunning(in, out, s)
-}
-
-func autoConvert_core_ContainerStateRunning_To_v1_ContainerStateRunning(in *core.ContainerStateRunning, out *v1.ContainerStateRunning, s conversion.Scope) error {
- out.StartedAt = in.StartedAt
- return nil
-}
-
-// Convert_core_ContainerStateRunning_To_v1_ContainerStateRunning is an autogenerated conversion function.
-func Convert_core_ContainerStateRunning_To_v1_ContainerStateRunning(in *core.ContainerStateRunning, out *v1.ContainerStateRunning, s conversion.Scope) error {
- return autoConvert_core_ContainerStateRunning_To_v1_ContainerStateRunning(in, out, s)
-}
-
-func autoConvert_v1_ContainerStateTerminated_To_core_ContainerStateTerminated(in *v1.ContainerStateTerminated, out *core.ContainerStateTerminated, s conversion.Scope) error {
- out.ExitCode = in.ExitCode
- out.Signal = in.Signal
- out.Reason = in.Reason
- out.Message = in.Message
- out.StartedAt = in.StartedAt
- out.FinishedAt = in.FinishedAt
- out.ContainerID = in.ContainerID
- return nil
-}
-
-// Convert_v1_ContainerStateTerminated_To_core_ContainerStateTerminated is an autogenerated conversion function.
-func Convert_v1_ContainerStateTerminated_To_core_ContainerStateTerminated(in *v1.ContainerStateTerminated, out *core.ContainerStateTerminated, s conversion.Scope) error {
- return autoConvert_v1_ContainerStateTerminated_To_core_ContainerStateTerminated(in, out, s)
-}
-
-func autoConvert_core_ContainerStateTerminated_To_v1_ContainerStateTerminated(in *core.ContainerStateTerminated, out *v1.ContainerStateTerminated, s conversion.Scope) error {
- out.ExitCode = in.ExitCode
- out.Signal = in.Signal
- out.Reason = in.Reason
- out.Message = in.Message
- out.StartedAt = in.StartedAt
- out.FinishedAt = in.FinishedAt
- out.ContainerID = in.ContainerID
- return nil
-}
-
-// Convert_core_ContainerStateTerminated_To_v1_ContainerStateTerminated is an autogenerated conversion function.
-func Convert_core_ContainerStateTerminated_To_v1_ContainerStateTerminated(in *core.ContainerStateTerminated, out *v1.ContainerStateTerminated, s conversion.Scope) error {
- return autoConvert_core_ContainerStateTerminated_To_v1_ContainerStateTerminated(in, out, s)
-}
-
-func autoConvert_v1_ContainerStateWaiting_To_core_ContainerStateWaiting(in *v1.ContainerStateWaiting, out *core.ContainerStateWaiting, s conversion.Scope) error {
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_ContainerStateWaiting_To_core_ContainerStateWaiting is an autogenerated conversion function.
-func Convert_v1_ContainerStateWaiting_To_core_ContainerStateWaiting(in *v1.ContainerStateWaiting, out *core.ContainerStateWaiting, s conversion.Scope) error {
- return autoConvert_v1_ContainerStateWaiting_To_core_ContainerStateWaiting(in, out, s)
-}
-
-func autoConvert_core_ContainerStateWaiting_To_v1_ContainerStateWaiting(in *core.ContainerStateWaiting, out *v1.ContainerStateWaiting, s conversion.Scope) error {
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_ContainerStateWaiting_To_v1_ContainerStateWaiting is an autogenerated conversion function.
-func Convert_core_ContainerStateWaiting_To_v1_ContainerStateWaiting(in *core.ContainerStateWaiting, out *v1.ContainerStateWaiting, s conversion.Scope) error {
- return autoConvert_core_ContainerStateWaiting_To_v1_ContainerStateWaiting(in, out, s)
-}
-
-func autoConvert_v1_ContainerStatus_To_core_ContainerStatus(in *v1.ContainerStatus, out *core.ContainerStatus, s conversion.Scope) error {
- out.Name = in.Name
- if err := Convert_v1_ContainerState_To_core_ContainerState(&in.State, &out.State, s); err != nil {
- return err
- }
- if err := Convert_v1_ContainerState_To_core_ContainerState(&in.LastTerminationState, &out.LastTerminationState, s); err != nil {
- return err
- }
- out.Ready = in.Ready
- out.RestartCount = in.RestartCount
- out.Image = in.Image
- out.ImageID = in.ImageID
- out.ContainerID = in.ContainerID
- return nil
-}
-
-// Convert_v1_ContainerStatus_To_core_ContainerStatus is an autogenerated conversion function.
-func Convert_v1_ContainerStatus_To_core_ContainerStatus(in *v1.ContainerStatus, out *core.ContainerStatus, s conversion.Scope) error {
- return autoConvert_v1_ContainerStatus_To_core_ContainerStatus(in, out, s)
-}
-
-func autoConvert_core_ContainerStatus_To_v1_ContainerStatus(in *core.ContainerStatus, out *v1.ContainerStatus, s conversion.Scope) error {
- out.Name = in.Name
- if err := Convert_core_ContainerState_To_v1_ContainerState(&in.State, &out.State, s); err != nil {
- return err
- }
- if err := Convert_core_ContainerState_To_v1_ContainerState(&in.LastTerminationState, &out.LastTerminationState, s); err != nil {
- return err
- }
- out.Ready = in.Ready
- out.RestartCount = in.RestartCount
- out.Image = in.Image
- out.ImageID = in.ImageID
- out.ContainerID = in.ContainerID
- return nil
-}
-
-// Convert_core_ContainerStatus_To_v1_ContainerStatus is an autogenerated conversion function.
-func Convert_core_ContainerStatus_To_v1_ContainerStatus(in *core.ContainerStatus, out *v1.ContainerStatus, s conversion.Scope) error {
- return autoConvert_core_ContainerStatus_To_v1_ContainerStatus(in, out, s)
-}
-
-func autoConvert_v1_DaemonEndpoint_To_core_DaemonEndpoint(in *v1.DaemonEndpoint, out *core.DaemonEndpoint, s conversion.Scope) error {
- out.Port = in.Port
- return nil
-}
-
-// Convert_v1_DaemonEndpoint_To_core_DaemonEndpoint is an autogenerated conversion function.
-func Convert_v1_DaemonEndpoint_To_core_DaemonEndpoint(in *v1.DaemonEndpoint, out *core.DaemonEndpoint, s conversion.Scope) error {
- return autoConvert_v1_DaemonEndpoint_To_core_DaemonEndpoint(in, out, s)
-}
-
-func autoConvert_core_DaemonEndpoint_To_v1_DaemonEndpoint(in *core.DaemonEndpoint, out *v1.DaemonEndpoint, s conversion.Scope) error {
- out.Port = in.Port
- return nil
-}
-
-// Convert_core_DaemonEndpoint_To_v1_DaemonEndpoint is an autogenerated conversion function.
-func Convert_core_DaemonEndpoint_To_v1_DaemonEndpoint(in *core.DaemonEndpoint, out *v1.DaemonEndpoint, s conversion.Scope) error {
- return autoConvert_core_DaemonEndpoint_To_v1_DaemonEndpoint(in, out, s)
-}
-
-func autoConvert_v1_DeleteOptions_To_core_DeleteOptions(in *v1.DeleteOptions, out *core.DeleteOptions, s conversion.Scope) error {
- out.GracePeriodSeconds = (*int64)(unsafe.Pointer(in.GracePeriodSeconds))
- out.Preconditions = (*core.Preconditions)(unsafe.Pointer(in.Preconditions))
- out.OrphanDependents = (*bool)(unsafe.Pointer(in.OrphanDependents))
- out.PropagationPolicy = (*core.DeletionPropagation)(unsafe.Pointer(in.PropagationPolicy))
- return nil
-}
-
-// Convert_v1_DeleteOptions_To_core_DeleteOptions is an autogenerated conversion function.
-func Convert_v1_DeleteOptions_To_core_DeleteOptions(in *v1.DeleteOptions, out *core.DeleteOptions, s conversion.Scope) error {
- return autoConvert_v1_DeleteOptions_To_core_DeleteOptions(in, out, s)
-}
-
-func autoConvert_core_DeleteOptions_To_v1_DeleteOptions(in *core.DeleteOptions, out *v1.DeleteOptions, s conversion.Scope) error {
- out.GracePeriodSeconds = (*int64)(unsafe.Pointer(in.GracePeriodSeconds))
- out.Preconditions = (*v1.Preconditions)(unsafe.Pointer(in.Preconditions))
- out.OrphanDependents = (*bool)(unsafe.Pointer(in.OrphanDependents))
- out.PropagationPolicy = (*v1.DeletionPropagation)(unsafe.Pointer(in.PropagationPolicy))
- return nil
-}
-
-// Convert_core_DeleteOptions_To_v1_DeleteOptions is an autogenerated conversion function.
-func Convert_core_DeleteOptions_To_v1_DeleteOptions(in *core.DeleteOptions, out *v1.DeleteOptions, s conversion.Scope) error {
- return autoConvert_core_DeleteOptions_To_v1_DeleteOptions(in, out, s)
-}
-
-func autoConvert_v1_DownwardAPIProjection_To_core_DownwardAPIProjection(in *v1.DownwardAPIProjection, out *core.DownwardAPIProjection, s conversion.Scope) error {
- out.Items = *(*[]core.DownwardAPIVolumeFile)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_DownwardAPIProjection_To_core_DownwardAPIProjection is an autogenerated conversion function.
-func Convert_v1_DownwardAPIProjection_To_core_DownwardAPIProjection(in *v1.DownwardAPIProjection, out *core.DownwardAPIProjection, s conversion.Scope) error {
- return autoConvert_v1_DownwardAPIProjection_To_core_DownwardAPIProjection(in, out, s)
-}
-
-func autoConvert_core_DownwardAPIProjection_To_v1_DownwardAPIProjection(in *core.DownwardAPIProjection, out *v1.DownwardAPIProjection, s conversion.Scope) error {
- out.Items = *(*[]v1.DownwardAPIVolumeFile)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_DownwardAPIProjection_To_v1_DownwardAPIProjection is an autogenerated conversion function.
-func Convert_core_DownwardAPIProjection_To_v1_DownwardAPIProjection(in *core.DownwardAPIProjection, out *v1.DownwardAPIProjection, s conversion.Scope) error {
- return autoConvert_core_DownwardAPIProjection_To_v1_DownwardAPIProjection(in, out, s)
-}
-
-func autoConvert_v1_DownwardAPIVolumeFile_To_core_DownwardAPIVolumeFile(in *v1.DownwardAPIVolumeFile, out *core.DownwardAPIVolumeFile, s conversion.Scope) error {
- out.Path = in.Path
- out.FieldRef = (*core.ObjectFieldSelector)(unsafe.Pointer(in.FieldRef))
- out.ResourceFieldRef = (*core.ResourceFieldSelector)(unsafe.Pointer(in.ResourceFieldRef))
- out.Mode = (*int32)(unsafe.Pointer(in.Mode))
- return nil
-}
-
-// Convert_v1_DownwardAPIVolumeFile_To_core_DownwardAPIVolumeFile is an autogenerated conversion function.
-func Convert_v1_DownwardAPIVolumeFile_To_core_DownwardAPIVolumeFile(in *v1.DownwardAPIVolumeFile, out *core.DownwardAPIVolumeFile, s conversion.Scope) error {
- return autoConvert_v1_DownwardAPIVolumeFile_To_core_DownwardAPIVolumeFile(in, out, s)
-}
-
-func autoConvert_core_DownwardAPIVolumeFile_To_v1_DownwardAPIVolumeFile(in *core.DownwardAPIVolumeFile, out *v1.DownwardAPIVolumeFile, s conversion.Scope) error {
- out.Path = in.Path
- out.FieldRef = (*v1.ObjectFieldSelector)(unsafe.Pointer(in.FieldRef))
- out.ResourceFieldRef = (*v1.ResourceFieldSelector)(unsafe.Pointer(in.ResourceFieldRef))
- out.Mode = (*int32)(unsafe.Pointer(in.Mode))
- return nil
-}
-
-// Convert_core_DownwardAPIVolumeFile_To_v1_DownwardAPIVolumeFile is an autogenerated conversion function.
-func Convert_core_DownwardAPIVolumeFile_To_v1_DownwardAPIVolumeFile(in *core.DownwardAPIVolumeFile, out *v1.DownwardAPIVolumeFile, s conversion.Scope) error {
- return autoConvert_core_DownwardAPIVolumeFile_To_v1_DownwardAPIVolumeFile(in, out, s)
-}
-
-func autoConvert_v1_DownwardAPIVolumeSource_To_core_DownwardAPIVolumeSource(in *v1.DownwardAPIVolumeSource, out *core.DownwardAPIVolumeSource, s conversion.Scope) error {
- out.Items = *(*[]core.DownwardAPIVolumeFile)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- return nil
-}
-
-// Convert_v1_DownwardAPIVolumeSource_To_core_DownwardAPIVolumeSource is an autogenerated conversion function.
-func Convert_v1_DownwardAPIVolumeSource_To_core_DownwardAPIVolumeSource(in *v1.DownwardAPIVolumeSource, out *core.DownwardAPIVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_DownwardAPIVolumeSource_To_core_DownwardAPIVolumeSource(in, out, s)
-}
-
-func autoConvert_core_DownwardAPIVolumeSource_To_v1_DownwardAPIVolumeSource(in *core.DownwardAPIVolumeSource, out *v1.DownwardAPIVolumeSource, s conversion.Scope) error {
- out.Items = *(*[]v1.DownwardAPIVolumeFile)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- return nil
-}
-
-// Convert_core_DownwardAPIVolumeSource_To_v1_DownwardAPIVolumeSource is an autogenerated conversion function.
-func Convert_core_DownwardAPIVolumeSource_To_v1_DownwardAPIVolumeSource(in *core.DownwardAPIVolumeSource, out *v1.DownwardAPIVolumeSource, s conversion.Scope) error {
- return autoConvert_core_DownwardAPIVolumeSource_To_v1_DownwardAPIVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_EmptyDirVolumeSource_To_core_EmptyDirVolumeSource(in *v1.EmptyDirVolumeSource, out *core.EmptyDirVolumeSource, s conversion.Scope) error {
- out.Medium = core.StorageMedium(in.Medium)
- out.SizeLimit = (*resource.Quantity)(unsafe.Pointer(in.SizeLimit))
- return nil
-}
-
-// Convert_v1_EmptyDirVolumeSource_To_core_EmptyDirVolumeSource is an autogenerated conversion function.
-func Convert_v1_EmptyDirVolumeSource_To_core_EmptyDirVolumeSource(in *v1.EmptyDirVolumeSource, out *core.EmptyDirVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_EmptyDirVolumeSource_To_core_EmptyDirVolumeSource(in, out, s)
-}
-
-func autoConvert_core_EmptyDirVolumeSource_To_v1_EmptyDirVolumeSource(in *core.EmptyDirVolumeSource, out *v1.EmptyDirVolumeSource, s conversion.Scope) error {
- out.Medium = v1.StorageMedium(in.Medium)
- out.SizeLimit = (*resource.Quantity)(unsafe.Pointer(in.SizeLimit))
- return nil
-}
-
-// Convert_core_EmptyDirVolumeSource_To_v1_EmptyDirVolumeSource is an autogenerated conversion function.
-func Convert_core_EmptyDirVolumeSource_To_v1_EmptyDirVolumeSource(in *core.EmptyDirVolumeSource, out *v1.EmptyDirVolumeSource, s conversion.Scope) error {
- return autoConvert_core_EmptyDirVolumeSource_To_v1_EmptyDirVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_EndpointAddress_To_core_EndpointAddress(in *v1.EndpointAddress, out *core.EndpointAddress, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostname = in.Hostname
- out.NodeName = (*string)(unsafe.Pointer(in.NodeName))
- out.TargetRef = (*core.ObjectReference)(unsafe.Pointer(in.TargetRef))
- return nil
-}
-
-// Convert_v1_EndpointAddress_To_core_EndpointAddress is an autogenerated conversion function.
-func Convert_v1_EndpointAddress_To_core_EndpointAddress(in *v1.EndpointAddress, out *core.EndpointAddress, s conversion.Scope) error {
- return autoConvert_v1_EndpointAddress_To_core_EndpointAddress(in, out, s)
-}
-
-func autoConvert_core_EndpointAddress_To_v1_EndpointAddress(in *core.EndpointAddress, out *v1.EndpointAddress, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostname = in.Hostname
- out.NodeName = (*string)(unsafe.Pointer(in.NodeName))
- out.TargetRef = (*v1.ObjectReference)(unsafe.Pointer(in.TargetRef))
- return nil
-}
-
-// Convert_core_EndpointAddress_To_v1_EndpointAddress is an autogenerated conversion function.
-func Convert_core_EndpointAddress_To_v1_EndpointAddress(in *core.EndpointAddress, out *v1.EndpointAddress, s conversion.Scope) error {
- return autoConvert_core_EndpointAddress_To_v1_EndpointAddress(in, out, s)
-}
-
-func autoConvert_v1_EndpointPort_To_core_EndpointPort(in *v1.EndpointPort, out *core.EndpointPort, s conversion.Scope) error {
- out.Name = in.Name
- out.Port = in.Port
- out.Protocol = core.Protocol(in.Protocol)
- return nil
-}
-
-// Convert_v1_EndpointPort_To_core_EndpointPort is an autogenerated conversion function.
-func Convert_v1_EndpointPort_To_core_EndpointPort(in *v1.EndpointPort, out *core.EndpointPort, s conversion.Scope) error {
- return autoConvert_v1_EndpointPort_To_core_EndpointPort(in, out, s)
-}
-
-func autoConvert_core_EndpointPort_To_v1_EndpointPort(in *core.EndpointPort, out *v1.EndpointPort, s conversion.Scope) error {
- out.Name = in.Name
- out.Port = in.Port
- out.Protocol = v1.Protocol(in.Protocol)
- return nil
-}
-
-// Convert_core_EndpointPort_To_v1_EndpointPort is an autogenerated conversion function.
-func Convert_core_EndpointPort_To_v1_EndpointPort(in *core.EndpointPort, out *v1.EndpointPort, s conversion.Scope) error {
- return autoConvert_core_EndpointPort_To_v1_EndpointPort(in, out, s)
-}
-
-func autoConvert_v1_EndpointSubset_To_core_EndpointSubset(in *v1.EndpointSubset, out *core.EndpointSubset, s conversion.Scope) error {
- out.Addresses = *(*[]core.EndpointAddress)(unsafe.Pointer(&in.Addresses))
- out.NotReadyAddresses = *(*[]core.EndpointAddress)(unsafe.Pointer(&in.NotReadyAddresses))
- out.Ports = *(*[]core.EndpointPort)(unsafe.Pointer(&in.Ports))
- return nil
-}
-
-// Convert_v1_EndpointSubset_To_core_EndpointSubset is an autogenerated conversion function.
-func Convert_v1_EndpointSubset_To_core_EndpointSubset(in *v1.EndpointSubset, out *core.EndpointSubset, s conversion.Scope) error {
- return autoConvert_v1_EndpointSubset_To_core_EndpointSubset(in, out, s)
-}
-
-func autoConvert_core_EndpointSubset_To_v1_EndpointSubset(in *core.EndpointSubset, out *v1.EndpointSubset, s conversion.Scope) error {
- out.Addresses = *(*[]v1.EndpointAddress)(unsafe.Pointer(&in.Addresses))
- out.NotReadyAddresses = *(*[]v1.EndpointAddress)(unsafe.Pointer(&in.NotReadyAddresses))
- out.Ports = *(*[]v1.EndpointPort)(unsafe.Pointer(&in.Ports))
- return nil
-}
-
-// Convert_core_EndpointSubset_To_v1_EndpointSubset is an autogenerated conversion function.
-func Convert_core_EndpointSubset_To_v1_EndpointSubset(in *core.EndpointSubset, out *v1.EndpointSubset, s conversion.Scope) error {
- return autoConvert_core_EndpointSubset_To_v1_EndpointSubset(in, out, s)
-}
-
-func autoConvert_v1_Endpoints_To_core_Endpoints(in *v1.Endpoints, out *core.Endpoints, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Subsets = *(*[]core.EndpointSubset)(unsafe.Pointer(&in.Subsets))
- return nil
-}
-
-// Convert_v1_Endpoints_To_core_Endpoints is an autogenerated conversion function.
-func Convert_v1_Endpoints_To_core_Endpoints(in *v1.Endpoints, out *core.Endpoints, s conversion.Scope) error {
- return autoConvert_v1_Endpoints_To_core_Endpoints(in, out, s)
-}
-
-func autoConvert_core_Endpoints_To_v1_Endpoints(in *core.Endpoints, out *v1.Endpoints, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Subsets = *(*[]v1.EndpointSubset)(unsafe.Pointer(&in.Subsets))
- return nil
-}
-
-// Convert_core_Endpoints_To_v1_Endpoints is an autogenerated conversion function.
-func Convert_core_Endpoints_To_v1_Endpoints(in *core.Endpoints, out *v1.Endpoints, s conversion.Scope) error {
- return autoConvert_core_Endpoints_To_v1_Endpoints(in, out, s)
-}
-
-func autoConvert_v1_EndpointsList_To_core_EndpointsList(in *v1.EndpointsList, out *core.EndpointsList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.Endpoints)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_EndpointsList_To_core_EndpointsList is an autogenerated conversion function.
-func Convert_v1_EndpointsList_To_core_EndpointsList(in *v1.EndpointsList, out *core.EndpointsList, s conversion.Scope) error {
- return autoConvert_v1_EndpointsList_To_core_EndpointsList(in, out, s)
-}
-
-func autoConvert_core_EndpointsList_To_v1_EndpointsList(in *core.EndpointsList, out *v1.EndpointsList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.Endpoints)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_EndpointsList_To_v1_EndpointsList is an autogenerated conversion function.
-func Convert_core_EndpointsList_To_v1_EndpointsList(in *core.EndpointsList, out *v1.EndpointsList, s conversion.Scope) error {
- return autoConvert_core_EndpointsList_To_v1_EndpointsList(in, out, s)
-}
-
-func autoConvert_v1_EnvFromSource_To_core_EnvFromSource(in *v1.EnvFromSource, out *core.EnvFromSource, s conversion.Scope) error {
- out.Prefix = in.Prefix
- out.ConfigMapRef = (*core.ConfigMapEnvSource)(unsafe.Pointer(in.ConfigMapRef))
- out.SecretRef = (*core.SecretEnvSource)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_v1_EnvFromSource_To_core_EnvFromSource is an autogenerated conversion function.
-func Convert_v1_EnvFromSource_To_core_EnvFromSource(in *v1.EnvFromSource, out *core.EnvFromSource, s conversion.Scope) error {
- return autoConvert_v1_EnvFromSource_To_core_EnvFromSource(in, out, s)
-}
-
-func autoConvert_core_EnvFromSource_To_v1_EnvFromSource(in *core.EnvFromSource, out *v1.EnvFromSource, s conversion.Scope) error {
- out.Prefix = in.Prefix
- out.ConfigMapRef = (*v1.ConfigMapEnvSource)(unsafe.Pointer(in.ConfigMapRef))
- out.SecretRef = (*v1.SecretEnvSource)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_core_EnvFromSource_To_v1_EnvFromSource is an autogenerated conversion function.
-func Convert_core_EnvFromSource_To_v1_EnvFromSource(in *core.EnvFromSource, out *v1.EnvFromSource, s conversion.Scope) error {
- return autoConvert_core_EnvFromSource_To_v1_EnvFromSource(in, out, s)
-}
-
-func autoConvert_v1_EnvVar_To_core_EnvVar(in *v1.EnvVar, out *core.EnvVar, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- out.ValueFrom = (*core.EnvVarSource)(unsafe.Pointer(in.ValueFrom))
- return nil
-}
-
-// Convert_v1_EnvVar_To_core_EnvVar is an autogenerated conversion function.
-func Convert_v1_EnvVar_To_core_EnvVar(in *v1.EnvVar, out *core.EnvVar, s conversion.Scope) error {
- return autoConvert_v1_EnvVar_To_core_EnvVar(in, out, s)
-}
-
-func autoConvert_core_EnvVar_To_v1_EnvVar(in *core.EnvVar, out *v1.EnvVar, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- out.ValueFrom = (*v1.EnvVarSource)(unsafe.Pointer(in.ValueFrom))
- return nil
-}
-
-// Convert_core_EnvVar_To_v1_EnvVar is an autogenerated conversion function.
-func Convert_core_EnvVar_To_v1_EnvVar(in *core.EnvVar, out *v1.EnvVar, s conversion.Scope) error {
- return autoConvert_core_EnvVar_To_v1_EnvVar(in, out, s)
-}
-
-func autoConvert_v1_EnvVarSource_To_core_EnvVarSource(in *v1.EnvVarSource, out *core.EnvVarSource, s conversion.Scope) error {
- out.FieldRef = (*core.ObjectFieldSelector)(unsafe.Pointer(in.FieldRef))
- out.ResourceFieldRef = (*core.ResourceFieldSelector)(unsafe.Pointer(in.ResourceFieldRef))
- out.ConfigMapKeyRef = (*core.ConfigMapKeySelector)(unsafe.Pointer(in.ConfigMapKeyRef))
- out.SecretKeyRef = (*core.SecretKeySelector)(unsafe.Pointer(in.SecretKeyRef))
- return nil
-}
-
-// Convert_v1_EnvVarSource_To_core_EnvVarSource is an autogenerated conversion function.
-func Convert_v1_EnvVarSource_To_core_EnvVarSource(in *v1.EnvVarSource, out *core.EnvVarSource, s conversion.Scope) error {
- return autoConvert_v1_EnvVarSource_To_core_EnvVarSource(in, out, s)
-}
-
-func autoConvert_core_EnvVarSource_To_v1_EnvVarSource(in *core.EnvVarSource, out *v1.EnvVarSource, s conversion.Scope) error {
- out.FieldRef = (*v1.ObjectFieldSelector)(unsafe.Pointer(in.FieldRef))
- out.ResourceFieldRef = (*v1.ResourceFieldSelector)(unsafe.Pointer(in.ResourceFieldRef))
- out.ConfigMapKeyRef = (*v1.ConfigMapKeySelector)(unsafe.Pointer(in.ConfigMapKeyRef))
- out.SecretKeyRef = (*v1.SecretKeySelector)(unsafe.Pointer(in.SecretKeyRef))
- return nil
-}
-
-// Convert_core_EnvVarSource_To_v1_EnvVarSource is an autogenerated conversion function.
-func Convert_core_EnvVarSource_To_v1_EnvVarSource(in *core.EnvVarSource, out *v1.EnvVarSource, s conversion.Scope) error {
- return autoConvert_core_EnvVarSource_To_v1_EnvVarSource(in, out, s)
-}
-
-func autoConvert_v1_Event_To_core_Event(in *v1.Event, out *core.Event, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ObjectReference_To_core_ObjectReference(&in.InvolvedObject, &out.InvolvedObject, s); err != nil {
- return err
- }
- out.Reason = in.Reason
- out.Message = in.Message
- if err := Convert_v1_EventSource_To_core_EventSource(&in.Source, &out.Source, s); err != nil {
- return err
- }
- out.FirstTimestamp = in.FirstTimestamp
- out.LastTimestamp = in.LastTimestamp
- out.Count = in.Count
- out.Type = in.Type
- out.EventTime = in.EventTime
- out.Series = (*core.EventSeries)(unsafe.Pointer(in.Series))
- out.Action = in.Action
- out.Related = (*core.ObjectReference)(unsafe.Pointer(in.Related))
- out.ReportingController = in.ReportingController
- out.ReportingInstance = in.ReportingInstance
- return nil
-}
-
-// Convert_v1_Event_To_core_Event is an autogenerated conversion function.
-func Convert_v1_Event_To_core_Event(in *v1.Event, out *core.Event, s conversion.Scope) error {
- return autoConvert_v1_Event_To_core_Event(in, out, s)
-}
-
-func autoConvert_core_Event_To_v1_Event(in *core.Event, out *v1.Event, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_ObjectReference_To_v1_ObjectReference(&in.InvolvedObject, &out.InvolvedObject, s); err != nil {
- return err
- }
- out.Reason = in.Reason
- out.Message = in.Message
- if err := Convert_core_EventSource_To_v1_EventSource(&in.Source, &out.Source, s); err != nil {
- return err
- }
- out.FirstTimestamp = in.FirstTimestamp
- out.LastTimestamp = in.LastTimestamp
- out.Count = in.Count
- out.Type = in.Type
- out.EventTime = in.EventTime
- out.Series = (*v1.EventSeries)(unsafe.Pointer(in.Series))
- out.Action = in.Action
- out.Related = (*v1.ObjectReference)(unsafe.Pointer(in.Related))
- out.ReportingController = in.ReportingController
- out.ReportingInstance = in.ReportingInstance
- return nil
-}
-
-// Convert_core_Event_To_v1_Event is an autogenerated conversion function.
-func Convert_core_Event_To_v1_Event(in *core.Event, out *v1.Event, s conversion.Scope) error {
- return autoConvert_core_Event_To_v1_Event(in, out, s)
-}
-
-func autoConvert_v1_EventList_To_core_EventList(in *v1.EventList, out *core.EventList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.Event)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_EventList_To_core_EventList is an autogenerated conversion function.
-func Convert_v1_EventList_To_core_EventList(in *v1.EventList, out *core.EventList, s conversion.Scope) error {
- return autoConvert_v1_EventList_To_core_EventList(in, out, s)
-}
-
-func autoConvert_core_EventList_To_v1_EventList(in *core.EventList, out *v1.EventList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.Event)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_EventList_To_v1_EventList is an autogenerated conversion function.
-func Convert_core_EventList_To_v1_EventList(in *core.EventList, out *v1.EventList, s conversion.Scope) error {
- return autoConvert_core_EventList_To_v1_EventList(in, out, s)
-}
-
-func autoConvert_v1_EventSeries_To_core_EventSeries(in *v1.EventSeries, out *core.EventSeries, s conversion.Scope) error {
- out.Count = in.Count
- out.LastObservedTime = in.LastObservedTime
- out.State = core.EventSeriesState(in.State)
- return nil
-}
-
-// Convert_v1_EventSeries_To_core_EventSeries is an autogenerated conversion function.
-func Convert_v1_EventSeries_To_core_EventSeries(in *v1.EventSeries, out *core.EventSeries, s conversion.Scope) error {
- return autoConvert_v1_EventSeries_To_core_EventSeries(in, out, s)
-}
-
-func autoConvert_core_EventSeries_To_v1_EventSeries(in *core.EventSeries, out *v1.EventSeries, s conversion.Scope) error {
- out.Count = in.Count
- out.LastObservedTime = in.LastObservedTime
- out.State = v1.EventSeriesState(in.State)
- return nil
-}
-
-// Convert_core_EventSeries_To_v1_EventSeries is an autogenerated conversion function.
-func Convert_core_EventSeries_To_v1_EventSeries(in *core.EventSeries, out *v1.EventSeries, s conversion.Scope) error {
- return autoConvert_core_EventSeries_To_v1_EventSeries(in, out, s)
-}
-
-func autoConvert_v1_EventSource_To_core_EventSource(in *v1.EventSource, out *core.EventSource, s conversion.Scope) error {
- out.Component = in.Component
- out.Host = in.Host
- return nil
-}
-
-// Convert_v1_EventSource_To_core_EventSource is an autogenerated conversion function.
-func Convert_v1_EventSource_To_core_EventSource(in *v1.EventSource, out *core.EventSource, s conversion.Scope) error {
- return autoConvert_v1_EventSource_To_core_EventSource(in, out, s)
-}
-
-func autoConvert_core_EventSource_To_v1_EventSource(in *core.EventSource, out *v1.EventSource, s conversion.Scope) error {
- out.Component = in.Component
- out.Host = in.Host
- return nil
-}
-
-// Convert_core_EventSource_To_v1_EventSource is an autogenerated conversion function.
-func Convert_core_EventSource_To_v1_EventSource(in *core.EventSource, out *v1.EventSource, s conversion.Scope) error {
- return autoConvert_core_EventSource_To_v1_EventSource(in, out, s)
-}
-
-func autoConvert_v1_ExecAction_To_core_ExecAction(in *v1.ExecAction, out *core.ExecAction, s conversion.Scope) error {
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- return nil
-}
-
-// Convert_v1_ExecAction_To_core_ExecAction is an autogenerated conversion function.
-func Convert_v1_ExecAction_To_core_ExecAction(in *v1.ExecAction, out *core.ExecAction, s conversion.Scope) error {
- return autoConvert_v1_ExecAction_To_core_ExecAction(in, out, s)
-}
-
-func autoConvert_core_ExecAction_To_v1_ExecAction(in *core.ExecAction, out *v1.ExecAction, s conversion.Scope) error {
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- return nil
-}
-
-// Convert_core_ExecAction_To_v1_ExecAction is an autogenerated conversion function.
-func Convert_core_ExecAction_To_v1_ExecAction(in *core.ExecAction, out *v1.ExecAction, s conversion.Scope) error {
- return autoConvert_core_ExecAction_To_v1_ExecAction(in, out, s)
-}
-
-func autoConvert_v1_FCVolumeSource_To_core_FCVolumeSource(in *v1.FCVolumeSource, out *core.FCVolumeSource, s conversion.Scope) error {
- out.TargetWWNs = *(*[]string)(unsafe.Pointer(&in.TargetWWNs))
- out.Lun = (*int32)(unsafe.Pointer(in.Lun))
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.WWIDs = *(*[]string)(unsafe.Pointer(&in.WWIDs))
- return nil
-}
-
-// Convert_v1_FCVolumeSource_To_core_FCVolumeSource is an autogenerated conversion function.
-func Convert_v1_FCVolumeSource_To_core_FCVolumeSource(in *v1.FCVolumeSource, out *core.FCVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_FCVolumeSource_To_core_FCVolumeSource(in, out, s)
-}
-
-func autoConvert_core_FCVolumeSource_To_v1_FCVolumeSource(in *core.FCVolumeSource, out *v1.FCVolumeSource, s conversion.Scope) error {
- out.TargetWWNs = *(*[]string)(unsafe.Pointer(&in.TargetWWNs))
- out.Lun = (*int32)(unsafe.Pointer(in.Lun))
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.WWIDs = *(*[]string)(unsafe.Pointer(&in.WWIDs))
- return nil
-}
-
-// Convert_core_FCVolumeSource_To_v1_FCVolumeSource is an autogenerated conversion function.
-func Convert_core_FCVolumeSource_To_v1_FCVolumeSource(in *core.FCVolumeSource, out *v1.FCVolumeSource, s conversion.Scope) error {
- return autoConvert_core_FCVolumeSource_To_v1_FCVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_FlexPersistentVolumeSource_To_core_FlexPersistentVolumeSource(in *v1.FlexPersistentVolumeSource, out *core.FlexPersistentVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.FSType = in.FSType
- out.SecretRef = (*core.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- out.Options = *(*map[string]string)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_v1_FlexPersistentVolumeSource_To_core_FlexPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_FlexPersistentVolumeSource_To_core_FlexPersistentVolumeSource(in *v1.FlexPersistentVolumeSource, out *core.FlexPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_FlexPersistentVolumeSource_To_core_FlexPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_FlexPersistentVolumeSource_To_v1_FlexPersistentVolumeSource(in *core.FlexPersistentVolumeSource, out *v1.FlexPersistentVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.FSType = in.FSType
- out.SecretRef = (*v1.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- out.Options = *(*map[string]string)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_core_FlexPersistentVolumeSource_To_v1_FlexPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_FlexPersistentVolumeSource_To_v1_FlexPersistentVolumeSource(in *core.FlexPersistentVolumeSource, out *v1.FlexPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_FlexPersistentVolumeSource_To_v1_FlexPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_FlexVolumeSource_To_core_FlexVolumeSource(in *v1.FlexVolumeSource, out *core.FlexVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.FSType = in.FSType
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- out.Options = *(*map[string]string)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_v1_FlexVolumeSource_To_core_FlexVolumeSource is an autogenerated conversion function.
-func Convert_v1_FlexVolumeSource_To_core_FlexVolumeSource(in *v1.FlexVolumeSource, out *core.FlexVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_FlexVolumeSource_To_core_FlexVolumeSource(in, out, s)
-}
-
-func autoConvert_core_FlexVolumeSource_To_v1_FlexVolumeSource(in *core.FlexVolumeSource, out *v1.FlexVolumeSource, s conversion.Scope) error {
- out.Driver = in.Driver
- out.FSType = in.FSType
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- out.Options = *(*map[string]string)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_core_FlexVolumeSource_To_v1_FlexVolumeSource is an autogenerated conversion function.
-func Convert_core_FlexVolumeSource_To_v1_FlexVolumeSource(in *core.FlexVolumeSource, out *v1.FlexVolumeSource, s conversion.Scope) error {
- return autoConvert_core_FlexVolumeSource_To_v1_FlexVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_FlockerVolumeSource_To_core_FlockerVolumeSource(in *v1.FlockerVolumeSource, out *core.FlockerVolumeSource, s conversion.Scope) error {
- out.DatasetName = in.DatasetName
- out.DatasetUUID = in.DatasetUUID
- return nil
-}
-
-// Convert_v1_FlockerVolumeSource_To_core_FlockerVolumeSource is an autogenerated conversion function.
-func Convert_v1_FlockerVolumeSource_To_core_FlockerVolumeSource(in *v1.FlockerVolumeSource, out *core.FlockerVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_FlockerVolumeSource_To_core_FlockerVolumeSource(in, out, s)
-}
-
-func autoConvert_core_FlockerVolumeSource_To_v1_FlockerVolumeSource(in *core.FlockerVolumeSource, out *v1.FlockerVolumeSource, s conversion.Scope) error {
- out.DatasetName = in.DatasetName
- out.DatasetUUID = in.DatasetUUID
- return nil
-}
-
-// Convert_core_FlockerVolumeSource_To_v1_FlockerVolumeSource is an autogenerated conversion function.
-func Convert_core_FlockerVolumeSource_To_v1_FlockerVolumeSource(in *core.FlockerVolumeSource, out *v1.FlockerVolumeSource, s conversion.Scope) error {
- return autoConvert_core_FlockerVolumeSource_To_v1_FlockerVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_GCEPersistentDiskVolumeSource_To_core_GCEPersistentDiskVolumeSource(in *v1.GCEPersistentDiskVolumeSource, out *core.GCEPersistentDiskVolumeSource, s conversion.Scope) error {
- out.PDName = in.PDName
- out.FSType = in.FSType
- out.Partition = in.Partition
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_GCEPersistentDiskVolumeSource_To_core_GCEPersistentDiskVolumeSource is an autogenerated conversion function.
-func Convert_v1_GCEPersistentDiskVolumeSource_To_core_GCEPersistentDiskVolumeSource(in *v1.GCEPersistentDiskVolumeSource, out *core.GCEPersistentDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_GCEPersistentDiskVolumeSource_To_core_GCEPersistentDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_core_GCEPersistentDiskVolumeSource_To_v1_GCEPersistentDiskVolumeSource(in *core.GCEPersistentDiskVolumeSource, out *v1.GCEPersistentDiskVolumeSource, s conversion.Scope) error {
- out.PDName = in.PDName
- out.FSType = in.FSType
- out.Partition = in.Partition
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_GCEPersistentDiskVolumeSource_To_v1_GCEPersistentDiskVolumeSource is an autogenerated conversion function.
-func Convert_core_GCEPersistentDiskVolumeSource_To_v1_GCEPersistentDiskVolumeSource(in *core.GCEPersistentDiskVolumeSource, out *v1.GCEPersistentDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_core_GCEPersistentDiskVolumeSource_To_v1_GCEPersistentDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_GitRepoVolumeSource_To_core_GitRepoVolumeSource(in *v1.GitRepoVolumeSource, out *core.GitRepoVolumeSource, s conversion.Scope) error {
- out.Repository = in.Repository
- out.Revision = in.Revision
- out.Directory = in.Directory
- return nil
-}
-
-// Convert_v1_GitRepoVolumeSource_To_core_GitRepoVolumeSource is an autogenerated conversion function.
-func Convert_v1_GitRepoVolumeSource_To_core_GitRepoVolumeSource(in *v1.GitRepoVolumeSource, out *core.GitRepoVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_GitRepoVolumeSource_To_core_GitRepoVolumeSource(in, out, s)
-}
-
-func autoConvert_core_GitRepoVolumeSource_To_v1_GitRepoVolumeSource(in *core.GitRepoVolumeSource, out *v1.GitRepoVolumeSource, s conversion.Scope) error {
- out.Repository = in.Repository
- out.Revision = in.Revision
- out.Directory = in.Directory
- return nil
-}
-
-// Convert_core_GitRepoVolumeSource_To_v1_GitRepoVolumeSource is an autogenerated conversion function.
-func Convert_core_GitRepoVolumeSource_To_v1_GitRepoVolumeSource(in *core.GitRepoVolumeSource, out *v1.GitRepoVolumeSource, s conversion.Scope) error {
- return autoConvert_core_GitRepoVolumeSource_To_v1_GitRepoVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_GlusterfsVolumeSource_To_core_GlusterfsVolumeSource(in *v1.GlusterfsVolumeSource, out *core.GlusterfsVolumeSource, s conversion.Scope) error {
- out.EndpointsName = in.EndpointsName
- out.Path = in.Path
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_GlusterfsVolumeSource_To_core_GlusterfsVolumeSource is an autogenerated conversion function.
-func Convert_v1_GlusterfsVolumeSource_To_core_GlusterfsVolumeSource(in *v1.GlusterfsVolumeSource, out *core.GlusterfsVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_GlusterfsVolumeSource_To_core_GlusterfsVolumeSource(in, out, s)
-}
-
-func autoConvert_core_GlusterfsVolumeSource_To_v1_GlusterfsVolumeSource(in *core.GlusterfsVolumeSource, out *v1.GlusterfsVolumeSource, s conversion.Scope) error {
- out.EndpointsName = in.EndpointsName
- out.Path = in.Path
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_GlusterfsVolumeSource_To_v1_GlusterfsVolumeSource is an autogenerated conversion function.
-func Convert_core_GlusterfsVolumeSource_To_v1_GlusterfsVolumeSource(in *core.GlusterfsVolumeSource, out *v1.GlusterfsVolumeSource, s conversion.Scope) error {
- return autoConvert_core_GlusterfsVolumeSource_To_v1_GlusterfsVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_HTTPGetAction_To_core_HTTPGetAction(in *v1.HTTPGetAction, out *core.HTTPGetAction, s conversion.Scope) error {
- out.Path = in.Path
- out.Port = in.Port
- out.Host = in.Host
- out.Scheme = core.URIScheme(in.Scheme)
- out.HTTPHeaders = *(*[]core.HTTPHeader)(unsafe.Pointer(&in.HTTPHeaders))
- return nil
-}
-
-// Convert_v1_HTTPGetAction_To_core_HTTPGetAction is an autogenerated conversion function.
-func Convert_v1_HTTPGetAction_To_core_HTTPGetAction(in *v1.HTTPGetAction, out *core.HTTPGetAction, s conversion.Scope) error {
- return autoConvert_v1_HTTPGetAction_To_core_HTTPGetAction(in, out, s)
-}
-
-func autoConvert_core_HTTPGetAction_To_v1_HTTPGetAction(in *core.HTTPGetAction, out *v1.HTTPGetAction, s conversion.Scope) error {
- out.Path = in.Path
- out.Port = in.Port
- out.Host = in.Host
- out.Scheme = v1.URIScheme(in.Scheme)
- out.HTTPHeaders = *(*[]v1.HTTPHeader)(unsafe.Pointer(&in.HTTPHeaders))
- return nil
-}
-
-// Convert_core_HTTPGetAction_To_v1_HTTPGetAction is an autogenerated conversion function.
-func Convert_core_HTTPGetAction_To_v1_HTTPGetAction(in *core.HTTPGetAction, out *v1.HTTPGetAction, s conversion.Scope) error {
- return autoConvert_core_HTTPGetAction_To_v1_HTTPGetAction(in, out, s)
-}
-
-func autoConvert_v1_HTTPHeader_To_core_HTTPHeader(in *v1.HTTPHeader, out *core.HTTPHeader, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- return nil
-}
-
-// Convert_v1_HTTPHeader_To_core_HTTPHeader is an autogenerated conversion function.
-func Convert_v1_HTTPHeader_To_core_HTTPHeader(in *v1.HTTPHeader, out *core.HTTPHeader, s conversion.Scope) error {
- return autoConvert_v1_HTTPHeader_To_core_HTTPHeader(in, out, s)
-}
-
-func autoConvert_core_HTTPHeader_To_v1_HTTPHeader(in *core.HTTPHeader, out *v1.HTTPHeader, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- return nil
-}
-
-// Convert_core_HTTPHeader_To_v1_HTTPHeader is an autogenerated conversion function.
-func Convert_core_HTTPHeader_To_v1_HTTPHeader(in *core.HTTPHeader, out *v1.HTTPHeader, s conversion.Scope) error {
- return autoConvert_core_HTTPHeader_To_v1_HTTPHeader(in, out, s)
-}
-
-func autoConvert_v1_Handler_To_core_Handler(in *v1.Handler, out *core.Handler, s conversion.Scope) error {
- out.Exec = (*core.ExecAction)(unsafe.Pointer(in.Exec))
- out.HTTPGet = (*core.HTTPGetAction)(unsafe.Pointer(in.HTTPGet))
- out.TCPSocket = (*core.TCPSocketAction)(unsafe.Pointer(in.TCPSocket))
- return nil
-}
-
-// Convert_v1_Handler_To_core_Handler is an autogenerated conversion function.
-func Convert_v1_Handler_To_core_Handler(in *v1.Handler, out *core.Handler, s conversion.Scope) error {
- return autoConvert_v1_Handler_To_core_Handler(in, out, s)
-}
-
-func autoConvert_core_Handler_To_v1_Handler(in *core.Handler, out *v1.Handler, s conversion.Scope) error {
- out.Exec = (*v1.ExecAction)(unsafe.Pointer(in.Exec))
- out.HTTPGet = (*v1.HTTPGetAction)(unsafe.Pointer(in.HTTPGet))
- out.TCPSocket = (*v1.TCPSocketAction)(unsafe.Pointer(in.TCPSocket))
- return nil
-}
-
-// Convert_core_Handler_To_v1_Handler is an autogenerated conversion function.
-func Convert_core_Handler_To_v1_Handler(in *core.Handler, out *v1.Handler, s conversion.Scope) error {
- return autoConvert_core_Handler_To_v1_Handler(in, out, s)
-}
-
-func autoConvert_v1_HostAlias_To_core_HostAlias(in *v1.HostAlias, out *core.HostAlias, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostnames = *(*[]string)(unsafe.Pointer(&in.Hostnames))
- return nil
-}
-
-// Convert_v1_HostAlias_To_core_HostAlias is an autogenerated conversion function.
-func Convert_v1_HostAlias_To_core_HostAlias(in *v1.HostAlias, out *core.HostAlias, s conversion.Scope) error {
- return autoConvert_v1_HostAlias_To_core_HostAlias(in, out, s)
-}
-
-func autoConvert_core_HostAlias_To_v1_HostAlias(in *core.HostAlias, out *v1.HostAlias, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostnames = *(*[]string)(unsafe.Pointer(&in.Hostnames))
- return nil
-}
-
-// Convert_core_HostAlias_To_v1_HostAlias is an autogenerated conversion function.
-func Convert_core_HostAlias_To_v1_HostAlias(in *core.HostAlias, out *v1.HostAlias, s conversion.Scope) error {
- return autoConvert_core_HostAlias_To_v1_HostAlias(in, out, s)
-}
-
-func autoConvert_v1_HostPathVolumeSource_To_core_HostPathVolumeSource(in *v1.HostPathVolumeSource, out *core.HostPathVolumeSource, s conversion.Scope) error {
- out.Path = in.Path
- out.Type = (*core.HostPathType)(unsafe.Pointer(in.Type))
- return nil
-}
-
-// Convert_v1_HostPathVolumeSource_To_core_HostPathVolumeSource is an autogenerated conversion function.
-func Convert_v1_HostPathVolumeSource_To_core_HostPathVolumeSource(in *v1.HostPathVolumeSource, out *core.HostPathVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_HostPathVolumeSource_To_core_HostPathVolumeSource(in, out, s)
-}
-
-func autoConvert_core_HostPathVolumeSource_To_v1_HostPathVolumeSource(in *core.HostPathVolumeSource, out *v1.HostPathVolumeSource, s conversion.Scope) error {
- out.Path = in.Path
- out.Type = (*v1.HostPathType)(unsafe.Pointer(in.Type))
- return nil
-}
-
-// Convert_core_HostPathVolumeSource_To_v1_HostPathVolumeSource is an autogenerated conversion function.
-func Convert_core_HostPathVolumeSource_To_v1_HostPathVolumeSource(in *core.HostPathVolumeSource, out *v1.HostPathVolumeSource, s conversion.Scope) error {
- return autoConvert_core_HostPathVolumeSource_To_v1_HostPathVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_ISCSIPersistentVolumeSource_To_core_ISCSIPersistentVolumeSource(in *v1.ISCSIPersistentVolumeSource, out *core.ISCSIPersistentVolumeSource, s conversion.Scope) error {
- out.TargetPortal = in.TargetPortal
- out.IQN = in.IQN
- out.Lun = in.Lun
- out.ISCSIInterface = in.ISCSIInterface
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.Portals = *(*[]string)(unsafe.Pointer(&in.Portals))
- out.DiscoveryCHAPAuth = in.DiscoveryCHAPAuth
- out.SessionCHAPAuth = in.SessionCHAPAuth
- out.SecretRef = (*core.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.InitiatorName = (*string)(unsafe.Pointer(in.InitiatorName))
- return nil
-}
-
-// Convert_v1_ISCSIPersistentVolumeSource_To_core_ISCSIPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_ISCSIPersistentVolumeSource_To_core_ISCSIPersistentVolumeSource(in *v1.ISCSIPersistentVolumeSource, out *core.ISCSIPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ISCSIPersistentVolumeSource_To_core_ISCSIPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ISCSIPersistentVolumeSource_To_v1_ISCSIPersistentVolumeSource(in *core.ISCSIPersistentVolumeSource, out *v1.ISCSIPersistentVolumeSource, s conversion.Scope) error {
- out.TargetPortal = in.TargetPortal
- out.IQN = in.IQN
- out.Lun = in.Lun
- out.ISCSIInterface = in.ISCSIInterface
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.Portals = *(*[]string)(unsafe.Pointer(&in.Portals))
- out.DiscoveryCHAPAuth = in.DiscoveryCHAPAuth
- out.SessionCHAPAuth = in.SessionCHAPAuth
- out.SecretRef = (*v1.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.InitiatorName = (*string)(unsafe.Pointer(in.InitiatorName))
- return nil
-}
-
-// Convert_core_ISCSIPersistentVolumeSource_To_v1_ISCSIPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_ISCSIPersistentVolumeSource_To_v1_ISCSIPersistentVolumeSource(in *core.ISCSIPersistentVolumeSource, out *v1.ISCSIPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ISCSIPersistentVolumeSource_To_v1_ISCSIPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_ISCSIVolumeSource_To_core_ISCSIVolumeSource(in *v1.ISCSIVolumeSource, out *core.ISCSIVolumeSource, s conversion.Scope) error {
- out.TargetPortal = in.TargetPortal
- out.IQN = in.IQN
- out.Lun = in.Lun
- out.ISCSIInterface = in.ISCSIInterface
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.Portals = *(*[]string)(unsafe.Pointer(&in.Portals))
- out.DiscoveryCHAPAuth = in.DiscoveryCHAPAuth
- out.SessionCHAPAuth = in.SessionCHAPAuth
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.InitiatorName = (*string)(unsafe.Pointer(in.InitiatorName))
- return nil
-}
-
-// Convert_v1_ISCSIVolumeSource_To_core_ISCSIVolumeSource is an autogenerated conversion function.
-func Convert_v1_ISCSIVolumeSource_To_core_ISCSIVolumeSource(in *v1.ISCSIVolumeSource, out *core.ISCSIVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ISCSIVolumeSource_To_core_ISCSIVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ISCSIVolumeSource_To_v1_ISCSIVolumeSource(in *core.ISCSIVolumeSource, out *v1.ISCSIVolumeSource, s conversion.Scope) error {
- out.TargetPortal = in.TargetPortal
- out.IQN = in.IQN
- out.Lun = in.Lun
- out.ISCSIInterface = in.ISCSIInterface
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.Portals = *(*[]string)(unsafe.Pointer(&in.Portals))
- out.DiscoveryCHAPAuth = in.DiscoveryCHAPAuth
- out.SessionCHAPAuth = in.SessionCHAPAuth
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.InitiatorName = (*string)(unsafe.Pointer(in.InitiatorName))
- return nil
-}
-
-// Convert_core_ISCSIVolumeSource_To_v1_ISCSIVolumeSource is an autogenerated conversion function.
-func Convert_core_ISCSIVolumeSource_To_v1_ISCSIVolumeSource(in *core.ISCSIVolumeSource, out *v1.ISCSIVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ISCSIVolumeSource_To_v1_ISCSIVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_KeyToPath_To_core_KeyToPath(in *v1.KeyToPath, out *core.KeyToPath, s conversion.Scope) error {
- out.Key = in.Key
- out.Path = in.Path
- out.Mode = (*int32)(unsafe.Pointer(in.Mode))
- return nil
-}
-
-// Convert_v1_KeyToPath_To_core_KeyToPath is an autogenerated conversion function.
-func Convert_v1_KeyToPath_To_core_KeyToPath(in *v1.KeyToPath, out *core.KeyToPath, s conversion.Scope) error {
- return autoConvert_v1_KeyToPath_To_core_KeyToPath(in, out, s)
-}
-
-func autoConvert_core_KeyToPath_To_v1_KeyToPath(in *core.KeyToPath, out *v1.KeyToPath, s conversion.Scope) error {
- out.Key = in.Key
- out.Path = in.Path
- out.Mode = (*int32)(unsafe.Pointer(in.Mode))
- return nil
-}
-
-// Convert_core_KeyToPath_To_v1_KeyToPath is an autogenerated conversion function.
-func Convert_core_KeyToPath_To_v1_KeyToPath(in *core.KeyToPath, out *v1.KeyToPath, s conversion.Scope) error {
- return autoConvert_core_KeyToPath_To_v1_KeyToPath(in, out, s)
-}
-
-func autoConvert_v1_Lifecycle_To_core_Lifecycle(in *v1.Lifecycle, out *core.Lifecycle, s conversion.Scope) error {
- out.PostStart = (*core.Handler)(unsafe.Pointer(in.PostStart))
- out.PreStop = (*core.Handler)(unsafe.Pointer(in.PreStop))
- return nil
-}
-
-// Convert_v1_Lifecycle_To_core_Lifecycle is an autogenerated conversion function.
-func Convert_v1_Lifecycle_To_core_Lifecycle(in *v1.Lifecycle, out *core.Lifecycle, s conversion.Scope) error {
- return autoConvert_v1_Lifecycle_To_core_Lifecycle(in, out, s)
-}
-
-func autoConvert_core_Lifecycle_To_v1_Lifecycle(in *core.Lifecycle, out *v1.Lifecycle, s conversion.Scope) error {
- out.PostStart = (*v1.Handler)(unsafe.Pointer(in.PostStart))
- out.PreStop = (*v1.Handler)(unsafe.Pointer(in.PreStop))
- return nil
-}
-
-// Convert_core_Lifecycle_To_v1_Lifecycle is an autogenerated conversion function.
-func Convert_core_Lifecycle_To_v1_Lifecycle(in *core.Lifecycle, out *v1.Lifecycle, s conversion.Scope) error {
- return autoConvert_core_Lifecycle_To_v1_Lifecycle(in, out, s)
-}
-
-func autoConvert_v1_LimitRange_To_core_LimitRange(in *v1.LimitRange, out *core.LimitRange, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_LimitRangeSpec_To_core_LimitRangeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_LimitRange_To_core_LimitRange is an autogenerated conversion function.
-func Convert_v1_LimitRange_To_core_LimitRange(in *v1.LimitRange, out *core.LimitRange, s conversion.Scope) error {
- return autoConvert_v1_LimitRange_To_core_LimitRange(in, out, s)
-}
-
-func autoConvert_core_LimitRange_To_v1_LimitRange(in *core.LimitRange, out *v1.LimitRange, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_LimitRangeSpec_To_v1_LimitRangeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_LimitRange_To_v1_LimitRange is an autogenerated conversion function.
-func Convert_core_LimitRange_To_v1_LimitRange(in *core.LimitRange, out *v1.LimitRange, s conversion.Scope) error {
- return autoConvert_core_LimitRange_To_v1_LimitRange(in, out, s)
-}
-
-func autoConvert_v1_LimitRangeItem_To_core_LimitRangeItem(in *v1.LimitRangeItem, out *core.LimitRangeItem, s conversion.Scope) error {
- out.Type = core.LimitType(in.Type)
- out.Max = *(*core.ResourceList)(unsafe.Pointer(&in.Max))
- out.Min = *(*core.ResourceList)(unsafe.Pointer(&in.Min))
- out.Default = *(*core.ResourceList)(unsafe.Pointer(&in.Default))
- out.DefaultRequest = *(*core.ResourceList)(unsafe.Pointer(&in.DefaultRequest))
- out.MaxLimitRequestRatio = *(*core.ResourceList)(unsafe.Pointer(&in.MaxLimitRequestRatio))
- return nil
-}
-
-// Convert_v1_LimitRangeItem_To_core_LimitRangeItem is an autogenerated conversion function.
-func Convert_v1_LimitRangeItem_To_core_LimitRangeItem(in *v1.LimitRangeItem, out *core.LimitRangeItem, s conversion.Scope) error {
- return autoConvert_v1_LimitRangeItem_To_core_LimitRangeItem(in, out, s)
-}
-
-func autoConvert_core_LimitRangeItem_To_v1_LimitRangeItem(in *core.LimitRangeItem, out *v1.LimitRangeItem, s conversion.Scope) error {
- out.Type = v1.LimitType(in.Type)
- out.Max = *(*v1.ResourceList)(unsafe.Pointer(&in.Max))
- out.Min = *(*v1.ResourceList)(unsafe.Pointer(&in.Min))
- out.Default = *(*v1.ResourceList)(unsafe.Pointer(&in.Default))
- out.DefaultRequest = *(*v1.ResourceList)(unsafe.Pointer(&in.DefaultRequest))
- out.MaxLimitRequestRatio = *(*v1.ResourceList)(unsafe.Pointer(&in.MaxLimitRequestRatio))
- return nil
-}
-
-// Convert_core_LimitRangeItem_To_v1_LimitRangeItem is an autogenerated conversion function.
-func Convert_core_LimitRangeItem_To_v1_LimitRangeItem(in *core.LimitRangeItem, out *v1.LimitRangeItem, s conversion.Scope) error {
- return autoConvert_core_LimitRangeItem_To_v1_LimitRangeItem(in, out, s)
-}
-
-func autoConvert_v1_LimitRangeList_To_core_LimitRangeList(in *v1.LimitRangeList, out *core.LimitRangeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.LimitRange)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_LimitRangeList_To_core_LimitRangeList is an autogenerated conversion function.
-func Convert_v1_LimitRangeList_To_core_LimitRangeList(in *v1.LimitRangeList, out *core.LimitRangeList, s conversion.Scope) error {
- return autoConvert_v1_LimitRangeList_To_core_LimitRangeList(in, out, s)
-}
-
-func autoConvert_core_LimitRangeList_To_v1_LimitRangeList(in *core.LimitRangeList, out *v1.LimitRangeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.LimitRange)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_LimitRangeList_To_v1_LimitRangeList is an autogenerated conversion function.
-func Convert_core_LimitRangeList_To_v1_LimitRangeList(in *core.LimitRangeList, out *v1.LimitRangeList, s conversion.Scope) error {
- return autoConvert_core_LimitRangeList_To_v1_LimitRangeList(in, out, s)
-}
-
-func autoConvert_v1_LimitRangeSpec_To_core_LimitRangeSpec(in *v1.LimitRangeSpec, out *core.LimitRangeSpec, s conversion.Scope) error {
- out.Limits = *(*[]core.LimitRangeItem)(unsafe.Pointer(&in.Limits))
- return nil
-}
-
-// Convert_v1_LimitRangeSpec_To_core_LimitRangeSpec is an autogenerated conversion function.
-func Convert_v1_LimitRangeSpec_To_core_LimitRangeSpec(in *v1.LimitRangeSpec, out *core.LimitRangeSpec, s conversion.Scope) error {
- return autoConvert_v1_LimitRangeSpec_To_core_LimitRangeSpec(in, out, s)
-}
-
-func autoConvert_core_LimitRangeSpec_To_v1_LimitRangeSpec(in *core.LimitRangeSpec, out *v1.LimitRangeSpec, s conversion.Scope) error {
- out.Limits = *(*[]v1.LimitRangeItem)(unsafe.Pointer(&in.Limits))
- return nil
-}
-
-// Convert_core_LimitRangeSpec_To_v1_LimitRangeSpec is an autogenerated conversion function.
-func Convert_core_LimitRangeSpec_To_v1_LimitRangeSpec(in *core.LimitRangeSpec, out *v1.LimitRangeSpec, s conversion.Scope) error {
- return autoConvert_core_LimitRangeSpec_To_v1_LimitRangeSpec(in, out, s)
-}
-
-func autoConvert_v1_List_To_core_List(in *v1.List, out *core.List, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]runtime.Object, len(*in))
- for i := range *in {
- if err := runtime.Convert_runtime_RawExtension_To_runtime_Object(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_List_To_core_List is an autogenerated conversion function.
-func Convert_v1_List_To_core_List(in *v1.List, out *core.List, s conversion.Scope) error {
- return autoConvert_v1_List_To_core_List(in, out, s)
-}
-
-func autoConvert_core_List_To_v1_List(in *core.List, out *v1.List, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]runtime.RawExtension, len(*in))
- for i := range *in {
- if err := runtime.Convert_runtime_Object_To_runtime_RawExtension(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_List_To_v1_List is an autogenerated conversion function.
-func Convert_core_List_To_v1_List(in *core.List, out *v1.List, s conversion.Scope) error {
- return autoConvert_core_List_To_v1_List(in, out, s)
-}
-
-func autoConvert_v1_ListOptions_To_core_ListOptions(in *v1.ListOptions, out *core.ListOptions, s conversion.Scope) error {
- if err := meta_v1.Convert_string_To_labels_Selector(&in.LabelSelector, &out.LabelSelector, s); err != nil {
- return err
- }
- if err := meta_v1.Convert_string_To_fields_Selector(&in.FieldSelector, &out.FieldSelector, s); err != nil {
- return err
- }
- out.IncludeUninitialized = in.IncludeUninitialized
- out.Watch = in.Watch
- out.ResourceVersion = in.ResourceVersion
- out.TimeoutSeconds = (*int64)(unsafe.Pointer(in.TimeoutSeconds))
- return nil
-}
-
-// Convert_v1_ListOptions_To_core_ListOptions is an autogenerated conversion function.
-func Convert_v1_ListOptions_To_core_ListOptions(in *v1.ListOptions, out *core.ListOptions, s conversion.Scope) error {
- return autoConvert_v1_ListOptions_To_core_ListOptions(in, out, s)
-}
-
-func autoConvert_core_ListOptions_To_v1_ListOptions(in *core.ListOptions, out *v1.ListOptions, s conversion.Scope) error {
- if err := meta_v1.Convert_labels_Selector_To_string(&in.LabelSelector, &out.LabelSelector, s); err != nil {
- return err
- }
- if err := meta_v1.Convert_fields_Selector_To_string(&in.FieldSelector, &out.FieldSelector, s); err != nil {
- return err
- }
- out.IncludeUninitialized = in.IncludeUninitialized
- out.Watch = in.Watch
- out.ResourceVersion = in.ResourceVersion
- out.TimeoutSeconds = (*int64)(unsafe.Pointer(in.TimeoutSeconds))
- return nil
-}
-
-// Convert_core_ListOptions_To_v1_ListOptions is an autogenerated conversion function.
-func Convert_core_ListOptions_To_v1_ListOptions(in *core.ListOptions, out *v1.ListOptions, s conversion.Scope) error {
- return autoConvert_core_ListOptions_To_v1_ListOptions(in, out, s)
-}
-
-func autoConvert_v1_LoadBalancerIngress_To_core_LoadBalancerIngress(in *v1.LoadBalancerIngress, out *core.LoadBalancerIngress, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostname = in.Hostname
- return nil
-}
-
-// Convert_v1_LoadBalancerIngress_To_core_LoadBalancerIngress is an autogenerated conversion function.
-func Convert_v1_LoadBalancerIngress_To_core_LoadBalancerIngress(in *v1.LoadBalancerIngress, out *core.LoadBalancerIngress, s conversion.Scope) error {
- return autoConvert_v1_LoadBalancerIngress_To_core_LoadBalancerIngress(in, out, s)
-}
-
-func autoConvert_core_LoadBalancerIngress_To_v1_LoadBalancerIngress(in *core.LoadBalancerIngress, out *v1.LoadBalancerIngress, s conversion.Scope) error {
- out.IP = in.IP
- out.Hostname = in.Hostname
- return nil
-}
-
-// Convert_core_LoadBalancerIngress_To_v1_LoadBalancerIngress is an autogenerated conversion function.
-func Convert_core_LoadBalancerIngress_To_v1_LoadBalancerIngress(in *core.LoadBalancerIngress, out *v1.LoadBalancerIngress, s conversion.Scope) error {
- return autoConvert_core_LoadBalancerIngress_To_v1_LoadBalancerIngress(in, out, s)
-}
-
-func autoConvert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus(in *v1.LoadBalancerStatus, out *core.LoadBalancerStatus, s conversion.Scope) error {
- out.Ingress = *(*[]core.LoadBalancerIngress)(unsafe.Pointer(&in.Ingress))
- return nil
-}
-
-// Convert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus is an autogenerated conversion function.
-func Convert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus(in *v1.LoadBalancerStatus, out *core.LoadBalancerStatus, s conversion.Scope) error {
- return autoConvert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus(in, out, s)
-}
-
-func autoConvert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus(in *core.LoadBalancerStatus, out *v1.LoadBalancerStatus, s conversion.Scope) error {
- out.Ingress = *(*[]v1.LoadBalancerIngress)(unsafe.Pointer(&in.Ingress))
- return nil
-}
-
-// Convert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus is an autogenerated conversion function.
-func Convert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus(in *core.LoadBalancerStatus, out *v1.LoadBalancerStatus, s conversion.Scope) error {
- return autoConvert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus(in, out, s)
-}
-
-func autoConvert_v1_LocalObjectReference_To_core_LocalObjectReference(in *v1.LocalObjectReference, out *core.LocalObjectReference, s conversion.Scope) error {
- out.Name = in.Name
- return nil
-}
-
-// Convert_v1_LocalObjectReference_To_core_LocalObjectReference is an autogenerated conversion function.
-func Convert_v1_LocalObjectReference_To_core_LocalObjectReference(in *v1.LocalObjectReference, out *core.LocalObjectReference, s conversion.Scope) error {
- return autoConvert_v1_LocalObjectReference_To_core_LocalObjectReference(in, out, s)
-}
-
-func autoConvert_core_LocalObjectReference_To_v1_LocalObjectReference(in *core.LocalObjectReference, out *v1.LocalObjectReference, s conversion.Scope) error {
- out.Name = in.Name
- return nil
-}
-
-// Convert_core_LocalObjectReference_To_v1_LocalObjectReference is an autogenerated conversion function.
-func Convert_core_LocalObjectReference_To_v1_LocalObjectReference(in *core.LocalObjectReference, out *v1.LocalObjectReference, s conversion.Scope) error {
- return autoConvert_core_LocalObjectReference_To_v1_LocalObjectReference(in, out, s)
-}
-
-func autoConvert_v1_LocalVolumeSource_To_core_LocalVolumeSource(in *v1.LocalVolumeSource, out *core.LocalVolumeSource, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_v1_LocalVolumeSource_To_core_LocalVolumeSource is an autogenerated conversion function.
-func Convert_v1_LocalVolumeSource_To_core_LocalVolumeSource(in *v1.LocalVolumeSource, out *core.LocalVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_LocalVolumeSource_To_core_LocalVolumeSource(in, out, s)
-}
-
-func autoConvert_core_LocalVolumeSource_To_v1_LocalVolumeSource(in *core.LocalVolumeSource, out *v1.LocalVolumeSource, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_core_LocalVolumeSource_To_v1_LocalVolumeSource is an autogenerated conversion function.
-func Convert_core_LocalVolumeSource_To_v1_LocalVolumeSource(in *core.LocalVolumeSource, out *v1.LocalVolumeSource, s conversion.Scope) error {
- return autoConvert_core_LocalVolumeSource_To_v1_LocalVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_NFSVolumeSource_To_core_NFSVolumeSource(in *v1.NFSVolumeSource, out *core.NFSVolumeSource, s conversion.Scope) error {
- out.Server = in.Server
- out.Path = in.Path
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_NFSVolumeSource_To_core_NFSVolumeSource is an autogenerated conversion function.
-func Convert_v1_NFSVolumeSource_To_core_NFSVolumeSource(in *v1.NFSVolumeSource, out *core.NFSVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_NFSVolumeSource_To_core_NFSVolumeSource(in, out, s)
-}
-
-func autoConvert_core_NFSVolumeSource_To_v1_NFSVolumeSource(in *core.NFSVolumeSource, out *v1.NFSVolumeSource, s conversion.Scope) error {
- out.Server = in.Server
- out.Path = in.Path
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_NFSVolumeSource_To_v1_NFSVolumeSource is an autogenerated conversion function.
-func Convert_core_NFSVolumeSource_To_v1_NFSVolumeSource(in *core.NFSVolumeSource, out *v1.NFSVolumeSource, s conversion.Scope) error {
- return autoConvert_core_NFSVolumeSource_To_v1_NFSVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Namespace_To_core_Namespace(in *v1.Namespace, out *core.Namespace, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_NamespaceSpec_To_core_NamespaceSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_NamespaceStatus_To_core_NamespaceStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_Namespace_To_core_Namespace is an autogenerated conversion function.
-func Convert_v1_Namespace_To_core_Namespace(in *v1.Namespace, out *core.Namespace, s conversion.Scope) error {
- return autoConvert_v1_Namespace_To_core_Namespace(in, out, s)
-}
-
-func autoConvert_core_Namespace_To_v1_Namespace(in *core.Namespace, out *v1.Namespace, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_NamespaceSpec_To_v1_NamespaceSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_NamespaceStatus_To_v1_NamespaceStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_Namespace_To_v1_Namespace is an autogenerated conversion function.
-func Convert_core_Namespace_To_v1_Namespace(in *core.Namespace, out *v1.Namespace, s conversion.Scope) error {
- return autoConvert_core_Namespace_To_v1_Namespace(in, out, s)
-}
-
-func autoConvert_v1_NamespaceList_To_core_NamespaceList(in *v1.NamespaceList, out *core.NamespaceList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.Namespace)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_NamespaceList_To_core_NamespaceList is an autogenerated conversion function.
-func Convert_v1_NamespaceList_To_core_NamespaceList(in *v1.NamespaceList, out *core.NamespaceList, s conversion.Scope) error {
- return autoConvert_v1_NamespaceList_To_core_NamespaceList(in, out, s)
-}
-
-func autoConvert_core_NamespaceList_To_v1_NamespaceList(in *core.NamespaceList, out *v1.NamespaceList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.Namespace)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_NamespaceList_To_v1_NamespaceList is an autogenerated conversion function.
-func Convert_core_NamespaceList_To_v1_NamespaceList(in *core.NamespaceList, out *v1.NamespaceList, s conversion.Scope) error {
- return autoConvert_core_NamespaceList_To_v1_NamespaceList(in, out, s)
-}
-
-func autoConvert_v1_NamespaceSpec_To_core_NamespaceSpec(in *v1.NamespaceSpec, out *core.NamespaceSpec, s conversion.Scope) error {
- out.Finalizers = *(*[]core.FinalizerName)(unsafe.Pointer(&in.Finalizers))
- return nil
-}
-
-// Convert_v1_NamespaceSpec_To_core_NamespaceSpec is an autogenerated conversion function.
-func Convert_v1_NamespaceSpec_To_core_NamespaceSpec(in *v1.NamespaceSpec, out *core.NamespaceSpec, s conversion.Scope) error {
- return autoConvert_v1_NamespaceSpec_To_core_NamespaceSpec(in, out, s)
-}
-
-func autoConvert_core_NamespaceSpec_To_v1_NamespaceSpec(in *core.NamespaceSpec, out *v1.NamespaceSpec, s conversion.Scope) error {
- out.Finalizers = *(*[]v1.FinalizerName)(unsafe.Pointer(&in.Finalizers))
- return nil
-}
-
-// Convert_core_NamespaceSpec_To_v1_NamespaceSpec is an autogenerated conversion function.
-func Convert_core_NamespaceSpec_To_v1_NamespaceSpec(in *core.NamespaceSpec, out *v1.NamespaceSpec, s conversion.Scope) error {
- return autoConvert_core_NamespaceSpec_To_v1_NamespaceSpec(in, out, s)
-}
-
-func autoConvert_v1_NamespaceStatus_To_core_NamespaceStatus(in *v1.NamespaceStatus, out *core.NamespaceStatus, s conversion.Scope) error {
- out.Phase = core.NamespacePhase(in.Phase)
- return nil
-}
-
-// Convert_v1_NamespaceStatus_To_core_NamespaceStatus is an autogenerated conversion function.
-func Convert_v1_NamespaceStatus_To_core_NamespaceStatus(in *v1.NamespaceStatus, out *core.NamespaceStatus, s conversion.Scope) error {
- return autoConvert_v1_NamespaceStatus_To_core_NamespaceStatus(in, out, s)
-}
-
-func autoConvert_core_NamespaceStatus_To_v1_NamespaceStatus(in *core.NamespaceStatus, out *v1.NamespaceStatus, s conversion.Scope) error {
- out.Phase = v1.NamespacePhase(in.Phase)
- return nil
-}
-
-// Convert_core_NamespaceStatus_To_v1_NamespaceStatus is an autogenerated conversion function.
-func Convert_core_NamespaceStatus_To_v1_NamespaceStatus(in *core.NamespaceStatus, out *v1.NamespaceStatus, s conversion.Scope) error {
- return autoConvert_core_NamespaceStatus_To_v1_NamespaceStatus(in, out, s)
-}
-
-func autoConvert_v1_Node_To_core_Node(in *v1.Node, out *core.Node, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_NodeSpec_To_core_NodeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_NodeStatus_To_core_NodeStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_Node_To_core_Node is an autogenerated conversion function.
-func Convert_v1_Node_To_core_Node(in *v1.Node, out *core.Node, s conversion.Scope) error {
- return autoConvert_v1_Node_To_core_Node(in, out, s)
-}
-
-func autoConvert_core_Node_To_v1_Node(in *core.Node, out *v1.Node, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_NodeSpec_To_v1_NodeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_NodeStatus_To_v1_NodeStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_Node_To_v1_Node is an autogenerated conversion function.
-func Convert_core_Node_To_v1_Node(in *core.Node, out *v1.Node, s conversion.Scope) error {
- return autoConvert_core_Node_To_v1_Node(in, out, s)
-}
-
-func autoConvert_v1_NodeAddress_To_core_NodeAddress(in *v1.NodeAddress, out *core.NodeAddress, s conversion.Scope) error {
- out.Type = core.NodeAddressType(in.Type)
- out.Address = in.Address
- return nil
-}
-
-// Convert_v1_NodeAddress_To_core_NodeAddress is an autogenerated conversion function.
-func Convert_v1_NodeAddress_To_core_NodeAddress(in *v1.NodeAddress, out *core.NodeAddress, s conversion.Scope) error {
- return autoConvert_v1_NodeAddress_To_core_NodeAddress(in, out, s)
-}
-
-func autoConvert_core_NodeAddress_To_v1_NodeAddress(in *core.NodeAddress, out *v1.NodeAddress, s conversion.Scope) error {
- out.Type = v1.NodeAddressType(in.Type)
- out.Address = in.Address
- return nil
-}
-
-// Convert_core_NodeAddress_To_v1_NodeAddress is an autogenerated conversion function.
-func Convert_core_NodeAddress_To_v1_NodeAddress(in *core.NodeAddress, out *v1.NodeAddress, s conversion.Scope) error {
- return autoConvert_core_NodeAddress_To_v1_NodeAddress(in, out, s)
-}
-
-func autoConvert_v1_NodeAffinity_To_core_NodeAffinity(in *v1.NodeAffinity, out *core.NodeAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = (*core.NodeSelector)(unsafe.Pointer(in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]core.PreferredSchedulingTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_v1_NodeAffinity_To_core_NodeAffinity is an autogenerated conversion function.
-func Convert_v1_NodeAffinity_To_core_NodeAffinity(in *v1.NodeAffinity, out *core.NodeAffinity, s conversion.Scope) error {
- return autoConvert_v1_NodeAffinity_To_core_NodeAffinity(in, out, s)
-}
-
-func autoConvert_core_NodeAffinity_To_v1_NodeAffinity(in *core.NodeAffinity, out *v1.NodeAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = (*v1.NodeSelector)(unsafe.Pointer(in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]v1.PreferredSchedulingTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_core_NodeAffinity_To_v1_NodeAffinity is an autogenerated conversion function.
-func Convert_core_NodeAffinity_To_v1_NodeAffinity(in *core.NodeAffinity, out *v1.NodeAffinity, s conversion.Scope) error {
- return autoConvert_core_NodeAffinity_To_v1_NodeAffinity(in, out, s)
-}
-
-func autoConvert_v1_NodeCondition_To_core_NodeCondition(in *v1.NodeCondition, out *core.NodeCondition, s conversion.Scope) error {
- out.Type = core.NodeConditionType(in.Type)
- out.Status = core.ConditionStatus(in.Status)
- out.LastHeartbeatTime = in.LastHeartbeatTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_NodeCondition_To_core_NodeCondition is an autogenerated conversion function.
-func Convert_v1_NodeCondition_To_core_NodeCondition(in *v1.NodeCondition, out *core.NodeCondition, s conversion.Scope) error {
- return autoConvert_v1_NodeCondition_To_core_NodeCondition(in, out, s)
-}
-
-func autoConvert_core_NodeCondition_To_v1_NodeCondition(in *core.NodeCondition, out *v1.NodeCondition, s conversion.Scope) error {
- out.Type = v1.NodeConditionType(in.Type)
- out.Status = v1.ConditionStatus(in.Status)
- out.LastHeartbeatTime = in.LastHeartbeatTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_NodeCondition_To_v1_NodeCondition is an autogenerated conversion function.
-func Convert_core_NodeCondition_To_v1_NodeCondition(in *core.NodeCondition, out *v1.NodeCondition, s conversion.Scope) error {
- return autoConvert_core_NodeCondition_To_v1_NodeCondition(in, out, s)
-}
-
-func autoConvert_v1_NodeConfigSource_To_core_NodeConfigSource(in *v1.NodeConfigSource, out *core.NodeConfigSource, s conversion.Scope) error {
- out.ConfigMapRef = (*core.ObjectReference)(unsafe.Pointer(in.ConfigMapRef))
- return nil
-}
-
-// Convert_v1_NodeConfigSource_To_core_NodeConfigSource is an autogenerated conversion function.
-func Convert_v1_NodeConfigSource_To_core_NodeConfigSource(in *v1.NodeConfigSource, out *core.NodeConfigSource, s conversion.Scope) error {
- return autoConvert_v1_NodeConfigSource_To_core_NodeConfigSource(in, out, s)
-}
-
-func autoConvert_core_NodeConfigSource_To_v1_NodeConfigSource(in *core.NodeConfigSource, out *v1.NodeConfigSource, s conversion.Scope) error {
- out.ConfigMapRef = (*v1.ObjectReference)(unsafe.Pointer(in.ConfigMapRef))
- return nil
-}
-
-// Convert_core_NodeConfigSource_To_v1_NodeConfigSource is an autogenerated conversion function.
-func Convert_core_NodeConfigSource_To_v1_NodeConfigSource(in *core.NodeConfigSource, out *v1.NodeConfigSource, s conversion.Scope) error {
- return autoConvert_core_NodeConfigSource_To_v1_NodeConfigSource(in, out, s)
-}
-
-func autoConvert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints(in *v1.NodeDaemonEndpoints, out *core.NodeDaemonEndpoints, s conversion.Scope) error {
- if err := Convert_v1_DaemonEndpoint_To_core_DaemonEndpoint(&in.KubeletEndpoint, &out.KubeletEndpoint, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints is an autogenerated conversion function.
-func Convert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints(in *v1.NodeDaemonEndpoints, out *core.NodeDaemonEndpoints, s conversion.Scope) error {
- return autoConvert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints(in, out, s)
-}
-
-func autoConvert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints(in *core.NodeDaemonEndpoints, out *v1.NodeDaemonEndpoints, s conversion.Scope) error {
- if err := Convert_core_DaemonEndpoint_To_v1_DaemonEndpoint(&in.KubeletEndpoint, &out.KubeletEndpoint, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints is an autogenerated conversion function.
-func Convert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints(in *core.NodeDaemonEndpoints, out *v1.NodeDaemonEndpoints, s conversion.Scope) error {
- return autoConvert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints(in, out, s)
-}
-
-func autoConvert_v1_NodeList_To_core_NodeList(in *v1.NodeList, out *core.NodeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.Node)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_NodeList_To_core_NodeList is an autogenerated conversion function.
-func Convert_v1_NodeList_To_core_NodeList(in *v1.NodeList, out *core.NodeList, s conversion.Scope) error {
- return autoConvert_v1_NodeList_To_core_NodeList(in, out, s)
-}
-
-func autoConvert_core_NodeList_To_v1_NodeList(in *core.NodeList, out *v1.NodeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.Node)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_NodeList_To_v1_NodeList is an autogenerated conversion function.
-func Convert_core_NodeList_To_v1_NodeList(in *core.NodeList, out *v1.NodeList, s conversion.Scope) error {
- return autoConvert_core_NodeList_To_v1_NodeList(in, out, s)
-}
-
-func autoConvert_v1_NodeProxyOptions_To_core_NodeProxyOptions(in *v1.NodeProxyOptions, out *core.NodeProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_v1_NodeProxyOptions_To_core_NodeProxyOptions is an autogenerated conversion function.
-func Convert_v1_NodeProxyOptions_To_core_NodeProxyOptions(in *v1.NodeProxyOptions, out *core.NodeProxyOptions, s conversion.Scope) error {
- return autoConvert_v1_NodeProxyOptions_To_core_NodeProxyOptions(in, out, s)
-}
-
-func autoConvert_core_NodeProxyOptions_To_v1_NodeProxyOptions(in *core.NodeProxyOptions, out *v1.NodeProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_core_NodeProxyOptions_To_v1_NodeProxyOptions is an autogenerated conversion function.
-func Convert_core_NodeProxyOptions_To_v1_NodeProxyOptions(in *core.NodeProxyOptions, out *v1.NodeProxyOptions, s conversion.Scope) error {
- return autoConvert_core_NodeProxyOptions_To_v1_NodeProxyOptions(in, out, s)
-}
-
-func autoConvert_v1_NodeResources_To_core_NodeResources(in *v1.NodeResources, out *core.NodeResources, s conversion.Scope) error {
- out.Capacity = *(*core.ResourceList)(unsafe.Pointer(&in.Capacity))
- return nil
-}
-
-// Convert_v1_NodeResources_To_core_NodeResources is an autogenerated conversion function.
-func Convert_v1_NodeResources_To_core_NodeResources(in *v1.NodeResources, out *core.NodeResources, s conversion.Scope) error {
- return autoConvert_v1_NodeResources_To_core_NodeResources(in, out, s)
-}
-
-func autoConvert_core_NodeResources_To_v1_NodeResources(in *core.NodeResources, out *v1.NodeResources, s conversion.Scope) error {
- out.Capacity = *(*v1.ResourceList)(unsafe.Pointer(&in.Capacity))
- return nil
-}
-
-// Convert_core_NodeResources_To_v1_NodeResources is an autogenerated conversion function.
-func Convert_core_NodeResources_To_v1_NodeResources(in *core.NodeResources, out *v1.NodeResources, s conversion.Scope) error {
- return autoConvert_core_NodeResources_To_v1_NodeResources(in, out, s)
-}
-
-func autoConvert_v1_NodeSelector_To_core_NodeSelector(in *v1.NodeSelector, out *core.NodeSelector, s conversion.Scope) error {
- out.NodeSelectorTerms = *(*[]core.NodeSelectorTerm)(unsafe.Pointer(&in.NodeSelectorTerms))
- return nil
-}
-
-// Convert_v1_NodeSelector_To_core_NodeSelector is an autogenerated conversion function.
-func Convert_v1_NodeSelector_To_core_NodeSelector(in *v1.NodeSelector, out *core.NodeSelector, s conversion.Scope) error {
- return autoConvert_v1_NodeSelector_To_core_NodeSelector(in, out, s)
-}
-
-func autoConvert_core_NodeSelector_To_v1_NodeSelector(in *core.NodeSelector, out *v1.NodeSelector, s conversion.Scope) error {
- out.NodeSelectorTerms = *(*[]v1.NodeSelectorTerm)(unsafe.Pointer(&in.NodeSelectorTerms))
- return nil
-}
-
-// Convert_core_NodeSelector_To_v1_NodeSelector is an autogenerated conversion function.
-func Convert_core_NodeSelector_To_v1_NodeSelector(in *core.NodeSelector, out *v1.NodeSelector, s conversion.Scope) error {
- return autoConvert_core_NodeSelector_To_v1_NodeSelector(in, out, s)
-}
-
-func autoConvert_v1_NodeSelectorRequirement_To_core_NodeSelectorRequirement(in *v1.NodeSelectorRequirement, out *core.NodeSelectorRequirement, s conversion.Scope) error {
- out.Key = in.Key
- out.Operator = core.NodeSelectorOperator(in.Operator)
- out.Values = *(*[]string)(unsafe.Pointer(&in.Values))
- return nil
-}
-
-// Convert_v1_NodeSelectorRequirement_To_core_NodeSelectorRequirement is an autogenerated conversion function.
-func Convert_v1_NodeSelectorRequirement_To_core_NodeSelectorRequirement(in *v1.NodeSelectorRequirement, out *core.NodeSelectorRequirement, s conversion.Scope) error {
- return autoConvert_v1_NodeSelectorRequirement_To_core_NodeSelectorRequirement(in, out, s)
-}
-
-func autoConvert_core_NodeSelectorRequirement_To_v1_NodeSelectorRequirement(in *core.NodeSelectorRequirement, out *v1.NodeSelectorRequirement, s conversion.Scope) error {
- out.Key = in.Key
- out.Operator = v1.NodeSelectorOperator(in.Operator)
- out.Values = *(*[]string)(unsafe.Pointer(&in.Values))
- return nil
-}
-
-// Convert_core_NodeSelectorRequirement_To_v1_NodeSelectorRequirement is an autogenerated conversion function.
-func Convert_core_NodeSelectorRequirement_To_v1_NodeSelectorRequirement(in *core.NodeSelectorRequirement, out *v1.NodeSelectorRequirement, s conversion.Scope) error {
- return autoConvert_core_NodeSelectorRequirement_To_v1_NodeSelectorRequirement(in, out, s)
-}
-
-func autoConvert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm(in *v1.NodeSelectorTerm, out *core.NodeSelectorTerm, s conversion.Scope) error {
- out.MatchExpressions = *(*[]core.NodeSelectorRequirement)(unsafe.Pointer(&in.MatchExpressions))
- return nil
-}
-
-// Convert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm is an autogenerated conversion function.
-func Convert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm(in *v1.NodeSelectorTerm, out *core.NodeSelectorTerm, s conversion.Scope) error {
- return autoConvert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm(in, out, s)
-}
-
-func autoConvert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm(in *core.NodeSelectorTerm, out *v1.NodeSelectorTerm, s conversion.Scope) error {
- out.MatchExpressions = *(*[]v1.NodeSelectorRequirement)(unsafe.Pointer(&in.MatchExpressions))
- return nil
-}
-
-// Convert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm is an autogenerated conversion function.
-func Convert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm(in *core.NodeSelectorTerm, out *v1.NodeSelectorTerm, s conversion.Scope) error {
- return autoConvert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm(in, out, s)
-}
-
-func autoConvert_v1_NodeSpec_To_core_NodeSpec(in *v1.NodeSpec, out *core.NodeSpec, s conversion.Scope) error {
- out.PodCIDR = in.PodCIDR
- out.ExternalID = in.ExternalID
- out.ProviderID = in.ProviderID
- out.Unschedulable = in.Unschedulable
- out.Taints = *(*[]core.Taint)(unsafe.Pointer(&in.Taints))
- out.ConfigSource = (*core.NodeConfigSource)(unsafe.Pointer(in.ConfigSource))
- return nil
-}
-
-// Convert_v1_NodeSpec_To_core_NodeSpec is an autogenerated conversion function.
-func Convert_v1_NodeSpec_To_core_NodeSpec(in *v1.NodeSpec, out *core.NodeSpec, s conversion.Scope) error {
- return autoConvert_v1_NodeSpec_To_core_NodeSpec(in, out, s)
-}
-
-func autoConvert_core_NodeSpec_To_v1_NodeSpec(in *core.NodeSpec, out *v1.NodeSpec, s conversion.Scope) error {
- out.PodCIDR = in.PodCIDR
- out.ExternalID = in.ExternalID
- out.ProviderID = in.ProviderID
- out.Unschedulable = in.Unschedulable
- out.Taints = *(*[]v1.Taint)(unsafe.Pointer(&in.Taints))
- out.ConfigSource = (*v1.NodeConfigSource)(unsafe.Pointer(in.ConfigSource))
- return nil
-}
-
-// Convert_core_NodeSpec_To_v1_NodeSpec is an autogenerated conversion function.
-func Convert_core_NodeSpec_To_v1_NodeSpec(in *core.NodeSpec, out *v1.NodeSpec, s conversion.Scope) error {
- return autoConvert_core_NodeSpec_To_v1_NodeSpec(in, out, s)
-}
-
-func autoConvert_v1_NodeStatus_To_core_NodeStatus(in *v1.NodeStatus, out *core.NodeStatus, s conversion.Scope) error {
- out.Capacity = *(*core.ResourceList)(unsafe.Pointer(&in.Capacity))
- out.Allocatable = *(*core.ResourceList)(unsafe.Pointer(&in.Allocatable))
- out.Phase = core.NodePhase(in.Phase)
- out.Conditions = *(*[]core.NodeCondition)(unsafe.Pointer(&in.Conditions))
- out.Addresses = *(*[]core.NodeAddress)(unsafe.Pointer(&in.Addresses))
- if err := Convert_v1_NodeDaemonEndpoints_To_core_NodeDaemonEndpoints(&in.DaemonEndpoints, &out.DaemonEndpoints, s); err != nil {
- return err
- }
- if err := Convert_v1_NodeSystemInfo_To_core_NodeSystemInfo(&in.NodeInfo, &out.NodeInfo, s); err != nil {
- return err
- }
- out.Images = *(*[]core.ContainerImage)(unsafe.Pointer(&in.Images))
- out.VolumesInUse = *(*[]core.UniqueVolumeName)(unsafe.Pointer(&in.VolumesInUse))
- out.VolumesAttached = *(*[]core.AttachedVolume)(unsafe.Pointer(&in.VolumesAttached))
- return nil
-}
-
-// Convert_v1_NodeStatus_To_core_NodeStatus is an autogenerated conversion function.
-func Convert_v1_NodeStatus_To_core_NodeStatus(in *v1.NodeStatus, out *core.NodeStatus, s conversion.Scope) error {
- return autoConvert_v1_NodeStatus_To_core_NodeStatus(in, out, s)
-}
-
-func autoConvert_core_NodeStatus_To_v1_NodeStatus(in *core.NodeStatus, out *v1.NodeStatus, s conversion.Scope) error {
- out.Capacity = *(*v1.ResourceList)(unsafe.Pointer(&in.Capacity))
- out.Allocatable = *(*v1.ResourceList)(unsafe.Pointer(&in.Allocatable))
- out.Phase = v1.NodePhase(in.Phase)
- out.Conditions = *(*[]v1.NodeCondition)(unsafe.Pointer(&in.Conditions))
- out.Addresses = *(*[]v1.NodeAddress)(unsafe.Pointer(&in.Addresses))
- if err := Convert_core_NodeDaemonEndpoints_To_v1_NodeDaemonEndpoints(&in.DaemonEndpoints, &out.DaemonEndpoints, s); err != nil {
- return err
- }
- if err := Convert_core_NodeSystemInfo_To_v1_NodeSystemInfo(&in.NodeInfo, &out.NodeInfo, s); err != nil {
- return err
- }
- out.Images = *(*[]v1.ContainerImage)(unsafe.Pointer(&in.Images))
- out.VolumesInUse = *(*[]v1.UniqueVolumeName)(unsafe.Pointer(&in.VolumesInUse))
- out.VolumesAttached = *(*[]v1.AttachedVolume)(unsafe.Pointer(&in.VolumesAttached))
- return nil
-}
-
-// Convert_core_NodeStatus_To_v1_NodeStatus is an autogenerated conversion function.
-func Convert_core_NodeStatus_To_v1_NodeStatus(in *core.NodeStatus, out *v1.NodeStatus, s conversion.Scope) error {
- return autoConvert_core_NodeStatus_To_v1_NodeStatus(in, out, s)
-}
-
-func autoConvert_v1_NodeSystemInfo_To_core_NodeSystemInfo(in *v1.NodeSystemInfo, out *core.NodeSystemInfo, s conversion.Scope) error {
- out.MachineID = in.MachineID
- out.SystemUUID = in.SystemUUID
- out.BootID = in.BootID
- out.KernelVersion = in.KernelVersion
- out.OSImage = in.OSImage
- out.ContainerRuntimeVersion = in.ContainerRuntimeVersion
- out.KubeletVersion = in.KubeletVersion
- out.KubeProxyVersion = in.KubeProxyVersion
- out.OperatingSystem = in.OperatingSystem
- out.Architecture = in.Architecture
- return nil
-}
-
-// Convert_v1_NodeSystemInfo_To_core_NodeSystemInfo is an autogenerated conversion function.
-func Convert_v1_NodeSystemInfo_To_core_NodeSystemInfo(in *v1.NodeSystemInfo, out *core.NodeSystemInfo, s conversion.Scope) error {
- return autoConvert_v1_NodeSystemInfo_To_core_NodeSystemInfo(in, out, s)
-}
-
-func autoConvert_core_NodeSystemInfo_To_v1_NodeSystemInfo(in *core.NodeSystemInfo, out *v1.NodeSystemInfo, s conversion.Scope) error {
- out.MachineID = in.MachineID
- out.SystemUUID = in.SystemUUID
- out.BootID = in.BootID
- out.KernelVersion = in.KernelVersion
- out.OSImage = in.OSImage
- out.ContainerRuntimeVersion = in.ContainerRuntimeVersion
- out.KubeletVersion = in.KubeletVersion
- out.KubeProxyVersion = in.KubeProxyVersion
- out.OperatingSystem = in.OperatingSystem
- out.Architecture = in.Architecture
- return nil
-}
-
-// Convert_core_NodeSystemInfo_To_v1_NodeSystemInfo is an autogenerated conversion function.
-func Convert_core_NodeSystemInfo_To_v1_NodeSystemInfo(in *core.NodeSystemInfo, out *v1.NodeSystemInfo, s conversion.Scope) error {
- return autoConvert_core_NodeSystemInfo_To_v1_NodeSystemInfo(in, out, s)
-}
-
-func autoConvert_v1_ObjectFieldSelector_To_core_ObjectFieldSelector(in *v1.ObjectFieldSelector, out *core.ObjectFieldSelector, s conversion.Scope) error {
- out.APIVersion = in.APIVersion
- out.FieldPath = in.FieldPath
- return nil
-}
-
-// Convert_v1_ObjectFieldSelector_To_core_ObjectFieldSelector is an autogenerated conversion function.
-func Convert_v1_ObjectFieldSelector_To_core_ObjectFieldSelector(in *v1.ObjectFieldSelector, out *core.ObjectFieldSelector, s conversion.Scope) error {
- return autoConvert_v1_ObjectFieldSelector_To_core_ObjectFieldSelector(in, out, s)
-}
-
-func autoConvert_core_ObjectFieldSelector_To_v1_ObjectFieldSelector(in *core.ObjectFieldSelector, out *v1.ObjectFieldSelector, s conversion.Scope) error {
- out.APIVersion = in.APIVersion
- out.FieldPath = in.FieldPath
- return nil
-}
-
-// Convert_core_ObjectFieldSelector_To_v1_ObjectFieldSelector is an autogenerated conversion function.
-func Convert_core_ObjectFieldSelector_To_v1_ObjectFieldSelector(in *core.ObjectFieldSelector, out *v1.ObjectFieldSelector, s conversion.Scope) error {
- return autoConvert_core_ObjectFieldSelector_To_v1_ObjectFieldSelector(in, out, s)
-}
-
-func autoConvert_v1_ObjectMeta_To_core_ObjectMeta(in *v1.ObjectMeta, out *core.ObjectMeta, s conversion.Scope) error {
- out.Name = in.Name
- out.GenerateName = in.GenerateName
- out.Namespace = in.Namespace
- out.SelfLink = in.SelfLink
- out.UID = types.UID(in.UID)
- out.ResourceVersion = in.ResourceVersion
- out.Generation = in.Generation
- out.CreationTimestamp = in.CreationTimestamp
- out.DeletionTimestamp = (*meta_v1.Time)(unsafe.Pointer(in.DeletionTimestamp))
- out.DeletionGracePeriodSeconds = (*int64)(unsafe.Pointer(in.DeletionGracePeriodSeconds))
- out.Labels = *(*map[string]string)(unsafe.Pointer(&in.Labels))
- out.Annotations = *(*map[string]string)(unsafe.Pointer(&in.Annotations))
- out.OwnerReferences = *(*[]meta_v1.OwnerReference)(unsafe.Pointer(&in.OwnerReferences))
- out.Initializers = (*meta_v1.Initializers)(unsafe.Pointer(in.Initializers))
- out.Finalizers = *(*[]string)(unsafe.Pointer(&in.Finalizers))
- out.ClusterName = in.ClusterName
- return nil
-}
-
-// Convert_v1_ObjectMeta_To_core_ObjectMeta is an autogenerated conversion function.
-func Convert_v1_ObjectMeta_To_core_ObjectMeta(in *v1.ObjectMeta, out *core.ObjectMeta, s conversion.Scope) error {
- return autoConvert_v1_ObjectMeta_To_core_ObjectMeta(in, out, s)
-}
-
-func autoConvert_core_ObjectMeta_To_v1_ObjectMeta(in *core.ObjectMeta, out *v1.ObjectMeta, s conversion.Scope) error {
- out.Name = in.Name
- out.GenerateName = in.GenerateName
- out.Namespace = in.Namespace
- out.SelfLink = in.SelfLink
- out.UID = types.UID(in.UID)
- out.ResourceVersion = in.ResourceVersion
- out.Generation = in.Generation
- out.CreationTimestamp = in.CreationTimestamp
- out.DeletionTimestamp = (*meta_v1.Time)(unsafe.Pointer(in.DeletionTimestamp))
- out.DeletionGracePeriodSeconds = (*int64)(unsafe.Pointer(in.DeletionGracePeriodSeconds))
- out.Labels = *(*map[string]string)(unsafe.Pointer(&in.Labels))
- out.Annotations = *(*map[string]string)(unsafe.Pointer(&in.Annotations))
- out.OwnerReferences = *(*[]meta_v1.OwnerReference)(unsafe.Pointer(&in.OwnerReferences))
- out.Initializers = (*meta_v1.Initializers)(unsafe.Pointer(in.Initializers))
- out.Finalizers = *(*[]string)(unsafe.Pointer(&in.Finalizers))
- out.ClusterName = in.ClusterName
- return nil
-}
-
-// Convert_core_ObjectMeta_To_v1_ObjectMeta is an autogenerated conversion function.
-func Convert_core_ObjectMeta_To_v1_ObjectMeta(in *core.ObjectMeta, out *v1.ObjectMeta, s conversion.Scope) error {
- return autoConvert_core_ObjectMeta_To_v1_ObjectMeta(in, out, s)
-}
-
-func autoConvert_v1_ObjectReference_To_core_ObjectReference(in *v1.ObjectReference, out *core.ObjectReference, s conversion.Scope) error {
- out.Kind = in.Kind
- out.Namespace = in.Namespace
- out.Name = in.Name
- out.UID = types.UID(in.UID)
- out.APIVersion = in.APIVersion
- out.ResourceVersion = in.ResourceVersion
- out.FieldPath = in.FieldPath
- return nil
-}
-
-// Convert_v1_ObjectReference_To_core_ObjectReference is an autogenerated conversion function.
-func Convert_v1_ObjectReference_To_core_ObjectReference(in *v1.ObjectReference, out *core.ObjectReference, s conversion.Scope) error {
- return autoConvert_v1_ObjectReference_To_core_ObjectReference(in, out, s)
-}
-
-func autoConvert_core_ObjectReference_To_v1_ObjectReference(in *core.ObjectReference, out *v1.ObjectReference, s conversion.Scope) error {
- out.Kind = in.Kind
- out.Namespace = in.Namespace
- out.Name = in.Name
- out.UID = types.UID(in.UID)
- out.APIVersion = in.APIVersion
- out.ResourceVersion = in.ResourceVersion
- out.FieldPath = in.FieldPath
- return nil
-}
-
-// Convert_core_ObjectReference_To_v1_ObjectReference is an autogenerated conversion function.
-func Convert_core_ObjectReference_To_v1_ObjectReference(in *core.ObjectReference, out *v1.ObjectReference, s conversion.Scope) error {
- return autoConvert_core_ObjectReference_To_v1_ObjectReference(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolume_To_core_PersistentVolume(in *v1.PersistentVolume, out *core.PersistentVolume, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_PersistentVolume_To_core_PersistentVolume is an autogenerated conversion function.
-func Convert_v1_PersistentVolume_To_core_PersistentVolume(in *v1.PersistentVolume, out *core.PersistentVolume, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolume_To_core_PersistentVolume(in, out, s)
-}
-
-func autoConvert_core_PersistentVolume_To_v1_PersistentVolume(in *core.PersistentVolume, out *v1.PersistentVolume, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_PersistentVolume_To_v1_PersistentVolume is an autogenerated conversion function.
-func Convert_core_PersistentVolume_To_v1_PersistentVolume(in *core.PersistentVolume, out *v1.PersistentVolume, s conversion.Scope) error {
- return autoConvert_core_PersistentVolume_To_v1_PersistentVolume(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaim_To_core_PersistentVolumeClaim(in *v1.PersistentVolumeClaim, out *core.PersistentVolumeClaim, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaim_To_core_PersistentVolumeClaim is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaim_To_core_PersistentVolumeClaim(in *v1.PersistentVolumeClaim, out *core.PersistentVolumeClaim, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaim_To_core_PersistentVolumeClaim(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaim_To_v1_PersistentVolumeClaim(in *core.PersistentVolumeClaim, out *v1.PersistentVolumeClaim, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_PersistentVolumeClaim_To_v1_PersistentVolumeClaim is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaim_To_v1_PersistentVolumeClaim(in *core.PersistentVolumeClaim, out *v1.PersistentVolumeClaim, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaim_To_v1_PersistentVolumeClaim(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaimCondition_To_core_PersistentVolumeClaimCondition(in *v1.PersistentVolumeClaimCondition, out *core.PersistentVolumeClaimCondition, s conversion.Scope) error {
- out.Type = core.PersistentVolumeClaimConditionType(in.Type)
- out.Status = core.ConditionStatus(in.Status)
- out.LastProbeTime = in.LastProbeTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaimCondition_To_core_PersistentVolumeClaimCondition is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaimCondition_To_core_PersistentVolumeClaimCondition(in *v1.PersistentVolumeClaimCondition, out *core.PersistentVolumeClaimCondition, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaimCondition_To_core_PersistentVolumeClaimCondition(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaimCondition_To_v1_PersistentVolumeClaimCondition(in *core.PersistentVolumeClaimCondition, out *v1.PersistentVolumeClaimCondition, s conversion.Scope) error {
- out.Type = v1.PersistentVolumeClaimConditionType(in.Type)
- out.Status = v1.ConditionStatus(in.Status)
- out.LastProbeTime = in.LastProbeTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_PersistentVolumeClaimCondition_To_v1_PersistentVolumeClaimCondition is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaimCondition_To_v1_PersistentVolumeClaimCondition(in *core.PersistentVolumeClaimCondition, out *v1.PersistentVolumeClaimCondition, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaimCondition_To_v1_PersistentVolumeClaimCondition(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaimList_To_core_PersistentVolumeClaimList(in *v1.PersistentVolumeClaimList, out *core.PersistentVolumeClaimList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.PersistentVolumeClaim)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaimList_To_core_PersistentVolumeClaimList is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaimList_To_core_PersistentVolumeClaimList(in *v1.PersistentVolumeClaimList, out *core.PersistentVolumeClaimList, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaimList_To_core_PersistentVolumeClaimList(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaimList_To_v1_PersistentVolumeClaimList(in *core.PersistentVolumeClaimList, out *v1.PersistentVolumeClaimList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.PersistentVolumeClaim)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_PersistentVolumeClaimList_To_v1_PersistentVolumeClaimList is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaimList_To_v1_PersistentVolumeClaimList(in *core.PersistentVolumeClaimList, out *v1.PersistentVolumeClaimList, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaimList_To_v1_PersistentVolumeClaimList(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec(in *v1.PersistentVolumeClaimSpec, out *core.PersistentVolumeClaimSpec, s conversion.Scope) error {
- out.AccessModes = *(*[]core.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.Selector = (*meta_v1.LabelSelector)(unsafe.Pointer(in.Selector))
- if err := Convert_v1_ResourceRequirements_To_core_ResourceRequirements(&in.Resources, &out.Resources, s); err != nil {
- return err
- }
- out.VolumeName = in.VolumeName
- out.StorageClassName = (*string)(unsafe.Pointer(in.StorageClassName))
- out.VolumeMode = (*core.PersistentVolumeMode)(unsafe.Pointer(in.VolumeMode))
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec(in *v1.PersistentVolumeClaimSpec, out *core.PersistentVolumeClaimSpec, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaimSpec_To_core_PersistentVolumeClaimSpec(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec(in *core.PersistentVolumeClaimSpec, out *v1.PersistentVolumeClaimSpec, s conversion.Scope) error {
- out.AccessModes = *(*[]v1.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.Selector = (*meta_v1.LabelSelector)(unsafe.Pointer(in.Selector))
- if err := Convert_core_ResourceRequirements_To_v1_ResourceRequirements(&in.Resources, &out.Resources, s); err != nil {
- return err
- }
- out.VolumeName = in.VolumeName
- out.StorageClassName = (*string)(unsafe.Pointer(in.StorageClassName))
- out.VolumeMode = (*v1.PersistentVolumeMode)(unsafe.Pointer(in.VolumeMode))
- return nil
-}
-
-// Convert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec(in *core.PersistentVolumeClaimSpec, out *v1.PersistentVolumeClaimSpec, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaimSpec_To_v1_PersistentVolumeClaimSpec(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus(in *v1.PersistentVolumeClaimStatus, out *core.PersistentVolumeClaimStatus, s conversion.Scope) error {
- out.Phase = core.PersistentVolumeClaimPhase(in.Phase)
- out.AccessModes = *(*[]core.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.Capacity = *(*core.ResourceList)(unsafe.Pointer(&in.Capacity))
- out.Conditions = *(*[]core.PersistentVolumeClaimCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus(in *v1.PersistentVolumeClaimStatus, out *core.PersistentVolumeClaimStatus, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaimStatus_To_core_PersistentVolumeClaimStatus(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus(in *core.PersistentVolumeClaimStatus, out *v1.PersistentVolumeClaimStatus, s conversion.Scope) error {
- out.Phase = v1.PersistentVolumeClaimPhase(in.Phase)
- out.AccessModes = *(*[]v1.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.Capacity = *(*v1.ResourceList)(unsafe.Pointer(&in.Capacity))
- out.Conditions = *(*[]v1.PersistentVolumeClaimCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus(in *core.PersistentVolumeClaimStatus, out *v1.PersistentVolumeClaimStatus, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaimStatus_To_v1_PersistentVolumeClaimStatus(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeClaimVolumeSource_To_core_PersistentVolumeClaimVolumeSource(in *v1.PersistentVolumeClaimVolumeSource, out *core.PersistentVolumeClaimVolumeSource, s conversion.Scope) error {
- out.ClaimName = in.ClaimName
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_PersistentVolumeClaimVolumeSource_To_core_PersistentVolumeClaimVolumeSource is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeClaimVolumeSource_To_core_PersistentVolumeClaimVolumeSource(in *v1.PersistentVolumeClaimVolumeSource, out *core.PersistentVolumeClaimVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeClaimVolumeSource_To_core_PersistentVolumeClaimVolumeSource(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeClaimVolumeSource_To_v1_PersistentVolumeClaimVolumeSource(in *core.PersistentVolumeClaimVolumeSource, out *v1.PersistentVolumeClaimVolumeSource, s conversion.Scope) error {
- out.ClaimName = in.ClaimName
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_PersistentVolumeClaimVolumeSource_To_v1_PersistentVolumeClaimVolumeSource is an autogenerated conversion function.
-func Convert_core_PersistentVolumeClaimVolumeSource_To_v1_PersistentVolumeClaimVolumeSource(in *core.PersistentVolumeClaimVolumeSource, out *v1.PersistentVolumeClaimVolumeSource, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeClaimVolumeSource_To_v1_PersistentVolumeClaimVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeList_To_core_PersistentVolumeList(in *v1.PersistentVolumeList, out *core.PersistentVolumeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.PersistentVolume, len(*in))
- for i := range *in {
- if err := Convert_v1_PersistentVolume_To_core_PersistentVolume(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_PersistentVolumeList_To_core_PersistentVolumeList is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeList_To_core_PersistentVolumeList(in *v1.PersistentVolumeList, out *core.PersistentVolumeList, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeList_To_core_PersistentVolumeList(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeList_To_v1_PersistentVolumeList(in *core.PersistentVolumeList, out *v1.PersistentVolumeList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.PersistentVolume, len(*in))
- for i := range *in {
- if err := Convert_core_PersistentVolume_To_v1_PersistentVolume(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_PersistentVolumeList_To_v1_PersistentVolumeList is an autogenerated conversion function.
-func Convert_core_PersistentVolumeList_To_v1_PersistentVolumeList(in *core.PersistentVolumeList, out *v1.PersistentVolumeList, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeList_To_v1_PersistentVolumeList(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource(in *v1.PersistentVolumeSource, out *core.PersistentVolumeSource, s conversion.Scope) error {
- out.GCEPersistentDisk = (*core.GCEPersistentDiskVolumeSource)(unsafe.Pointer(in.GCEPersistentDisk))
- out.AWSElasticBlockStore = (*core.AWSElasticBlockStoreVolumeSource)(unsafe.Pointer(in.AWSElasticBlockStore))
- out.HostPath = (*core.HostPathVolumeSource)(unsafe.Pointer(in.HostPath))
- out.Glusterfs = (*core.GlusterfsVolumeSource)(unsafe.Pointer(in.Glusterfs))
- out.NFS = (*core.NFSVolumeSource)(unsafe.Pointer(in.NFS))
- out.RBD = (*core.RBDPersistentVolumeSource)(unsafe.Pointer(in.RBD))
- out.ISCSI = (*core.ISCSIPersistentVolumeSource)(unsafe.Pointer(in.ISCSI))
- out.Cinder = (*core.CinderVolumeSource)(unsafe.Pointer(in.Cinder))
- out.CephFS = (*core.CephFSPersistentVolumeSource)(unsafe.Pointer(in.CephFS))
- out.FC = (*core.FCVolumeSource)(unsafe.Pointer(in.FC))
- out.Flocker = (*core.FlockerVolumeSource)(unsafe.Pointer(in.Flocker))
- out.FlexVolume = (*core.FlexPersistentVolumeSource)(unsafe.Pointer(in.FlexVolume))
- out.AzureFile = (*core.AzureFilePersistentVolumeSource)(unsafe.Pointer(in.AzureFile))
- out.VsphereVolume = (*core.VsphereVirtualDiskVolumeSource)(unsafe.Pointer(in.VsphereVolume))
- out.Quobyte = (*core.QuobyteVolumeSource)(unsafe.Pointer(in.Quobyte))
- out.AzureDisk = (*core.AzureDiskVolumeSource)(unsafe.Pointer(in.AzureDisk))
- out.PhotonPersistentDisk = (*core.PhotonPersistentDiskVolumeSource)(unsafe.Pointer(in.PhotonPersistentDisk))
- out.PortworxVolume = (*core.PortworxVolumeSource)(unsafe.Pointer(in.PortworxVolume))
- out.ScaleIO = (*core.ScaleIOPersistentVolumeSource)(unsafe.Pointer(in.ScaleIO))
- out.Local = (*core.LocalVolumeSource)(unsafe.Pointer(in.Local))
- out.StorageOS = (*core.StorageOSPersistentVolumeSource)(unsafe.Pointer(in.StorageOS))
- out.CSI = (*core.CSIPersistentVolumeSource)(unsafe.Pointer(in.CSI))
- return nil
-}
-
-// Convert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource(in *v1.PersistentVolumeSource, out *core.PersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource(in *core.PersistentVolumeSource, out *v1.PersistentVolumeSource, s conversion.Scope) error {
- out.GCEPersistentDisk = (*v1.GCEPersistentDiskVolumeSource)(unsafe.Pointer(in.GCEPersistentDisk))
- out.AWSElasticBlockStore = (*v1.AWSElasticBlockStoreVolumeSource)(unsafe.Pointer(in.AWSElasticBlockStore))
- out.HostPath = (*v1.HostPathVolumeSource)(unsafe.Pointer(in.HostPath))
- out.Glusterfs = (*v1.GlusterfsVolumeSource)(unsafe.Pointer(in.Glusterfs))
- out.NFS = (*v1.NFSVolumeSource)(unsafe.Pointer(in.NFS))
- out.RBD = (*v1.RBDPersistentVolumeSource)(unsafe.Pointer(in.RBD))
- out.Quobyte = (*v1.QuobyteVolumeSource)(unsafe.Pointer(in.Quobyte))
- out.ISCSI = (*v1.ISCSIPersistentVolumeSource)(unsafe.Pointer(in.ISCSI))
- out.FlexVolume = (*v1.FlexPersistentVolumeSource)(unsafe.Pointer(in.FlexVolume))
- out.Cinder = (*v1.CinderVolumeSource)(unsafe.Pointer(in.Cinder))
- out.CephFS = (*v1.CephFSPersistentVolumeSource)(unsafe.Pointer(in.CephFS))
- out.FC = (*v1.FCVolumeSource)(unsafe.Pointer(in.FC))
- out.Flocker = (*v1.FlockerVolumeSource)(unsafe.Pointer(in.Flocker))
- out.AzureFile = (*v1.AzureFilePersistentVolumeSource)(unsafe.Pointer(in.AzureFile))
- out.VsphereVolume = (*v1.VsphereVirtualDiskVolumeSource)(unsafe.Pointer(in.VsphereVolume))
- out.AzureDisk = (*v1.AzureDiskVolumeSource)(unsafe.Pointer(in.AzureDisk))
- out.PhotonPersistentDisk = (*v1.PhotonPersistentDiskVolumeSource)(unsafe.Pointer(in.PhotonPersistentDisk))
- out.PortworxVolume = (*v1.PortworxVolumeSource)(unsafe.Pointer(in.PortworxVolume))
- out.ScaleIO = (*v1.ScaleIOPersistentVolumeSource)(unsafe.Pointer(in.ScaleIO))
- out.Local = (*v1.LocalVolumeSource)(unsafe.Pointer(in.Local))
- out.StorageOS = (*v1.StorageOSPersistentVolumeSource)(unsafe.Pointer(in.StorageOS))
- out.CSI = (*v1.CSIPersistentVolumeSource)(unsafe.Pointer(in.CSI))
- return nil
-}
-
-// Convert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource(in *core.PersistentVolumeSource, out *v1.PersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec(in *v1.PersistentVolumeSpec, out *core.PersistentVolumeSpec, s conversion.Scope) error {
- out.Capacity = *(*core.ResourceList)(unsafe.Pointer(&in.Capacity))
- if err := Convert_v1_PersistentVolumeSource_To_core_PersistentVolumeSource(&in.PersistentVolumeSource, &out.PersistentVolumeSource, s); err != nil {
- return err
- }
- out.AccessModes = *(*[]core.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.ClaimRef = (*core.ObjectReference)(unsafe.Pointer(in.ClaimRef))
- out.PersistentVolumeReclaimPolicy = core.PersistentVolumeReclaimPolicy(in.PersistentVolumeReclaimPolicy)
- out.StorageClassName = in.StorageClassName
- out.MountOptions = *(*[]string)(unsafe.Pointer(&in.MountOptions))
- out.VolumeMode = (*core.PersistentVolumeMode)(unsafe.Pointer(in.VolumeMode))
- out.NodeAffinity = (*core.VolumeNodeAffinity)(unsafe.Pointer(in.NodeAffinity))
- return nil
-}
-
-// Convert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec(in *v1.PersistentVolumeSpec, out *core.PersistentVolumeSpec, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeSpec_To_core_PersistentVolumeSpec(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec(in *core.PersistentVolumeSpec, out *v1.PersistentVolumeSpec, s conversion.Scope) error {
- out.Capacity = *(*v1.ResourceList)(unsafe.Pointer(&in.Capacity))
- if err := Convert_core_PersistentVolumeSource_To_v1_PersistentVolumeSource(&in.PersistentVolumeSource, &out.PersistentVolumeSource, s); err != nil {
- return err
- }
- out.AccessModes = *(*[]v1.PersistentVolumeAccessMode)(unsafe.Pointer(&in.AccessModes))
- out.ClaimRef = (*v1.ObjectReference)(unsafe.Pointer(in.ClaimRef))
- out.PersistentVolumeReclaimPolicy = v1.PersistentVolumeReclaimPolicy(in.PersistentVolumeReclaimPolicy)
- out.StorageClassName = in.StorageClassName
- out.MountOptions = *(*[]string)(unsafe.Pointer(&in.MountOptions))
- out.VolumeMode = (*v1.PersistentVolumeMode)(unsafe.Pointer(in.VolumeMode))
- out.NodeAffinity = (*v1.VolumeNodeAffinity)(unsafe.Pointer(in.NodeAffinity))
- return nil
-}
-
-// Convert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec is an autogenerated conversion function.
-func Convert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec(in *core.PersistentVolumeSpec, out *v1.PersistentVolumeSpec, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeSpec_To_v1_PersistentVolumeSpec(in, out, s)
-}
-
-func autoConvert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus(in *v1.PersistentVolumeStatus, out *core.PersistentVolumeStatus, s conversion.Scope) error {
- out.Phase = core.PersistentVolumePhase(in.Phase)
- out.Message = in.Message
- out.Reason = in.Reason
- return nil
-}
-
-// Convert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus is an autogenerated conversion function.
-func Convert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus(in *v1.PersistentVolumeStatus, out *core.PersistentVolumeStatus, s conversion.Scope) error {
- return autoConvert_v1_PersistentVolumeStatus_To_core_PersistentVolumeStatus(in, out, s)
-}
-
-func autoConvert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus(in *core.PersistentVolumeStatus, out *v1.PersistentVolumeStatus, s conversion.Scope) error {
- out.Phase = v1.PersistentVolumePhase(in.Phase)
- out.Message = in.Message
- out.Reason = in.Reason
- return nil
-}
-
-// Convert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus is an autogenerated conversion function.
-func Convert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus(in *core.PersistentVolumeStatus, out *v1.PersistentVolumeStatus, s conversion.Scope) error {
- return autoConvert_core_PersistentVolumeStatus_To_v1_PersistentVolumeStatus(in, out, s)
-}
-
-func autoConvert_v1_PhotonPersistentDiskVolumeSource_To_core_PhotonPersistentDiskVolumeSource(in *v1.PhotonPersistentDiskVolumeSource, out *core.PhotonPersistentDiskVolumeSource, s conversion.Scope) error {
- out.PdID = in.PdID
- out.FSType = in.FSType
- return nil
-}
-
-// Convert_v1_PhotonPersistentDiskVolumeSource_To_core_PhotonPersistentDiskVolumeSource is an autogenerated conversion function.
-func Convert_v1_PhotonPersistentDiskVolumeSource_To_core_PhotonPersistentDiskVolumeSource(in *v1.PhotonPersistentDiskVolumeSource, out *core.PhotonPersistentDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_PhotonPersistentDiskVolumeSource_To_core_PhotonPersistentDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_core_PhotonPersistentDiskVolumeSource_To_v1_PhotonPersistentDiskVolumeSource(in *core.PhotonPersistentDiskVolumeSource, out *v1.PhotonPersistentDiskVolumeSource, s conversion.Scope) error {
- out.PdID = in.PdID
- out.FSType = in.FSType
- return nil
-}
-
-// Convert_core_PhotonPersistentDiskVolumeSource_To_v1_PhotonPersistentDiskVolumeSource is an autogenerated conversion function.
-func Convert_core_PhotonPersistentDiskVolumeSource_To_v1_PhotonPersistentDiskVolumeSource(in *core.PhotonPersistentDiskVolumeSource, out *v1.PhotonPersistentDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_core_PhotonPersistentDiskVolumeSource_To_v1_PhotonPersistentDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Pod_To_core_Pod(in *v1.Pod, out *core.Pod, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PodSpec_To_core_PodSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_PodStatus_To_core_PodStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-func autoConvert_core_Pod_To_v1_Pod(in *core.Pod, out *v1.Pod, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PodSpec_To_v1_PodSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_PodStatus_To_v1_PodStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-func autoConvert_v1_PodAffinity_To_core_PodAffinity(in *v1.PodAffinity, out *core.PodAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = *(*[]core.PodAffinityTerm)(unsafe.Pointer(&in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]core.WeightedPodAffinityTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_v1_PodAffinity_To_core_PodAffinity is an autogenerated conversion function.
-func Convert_v1_PodAffinity_To_core_PodAffinity(in *v1.PodAffinity, out *core.PodAffinity, s conversion.Scope) error {
- return autoConvert_v1_PodAffinity_To_core_PodAffinity(in, out, s)
-}
-
-func autoConvert_core_PodAffinity_To_v1_PodAffinity(in *core.PodAffinity, out *v1.PodAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = *(*[]v1.PodAffinityTerm)(unsafe.Pointer(&in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]v1.WeightedPodAffinityTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_core_PodAffinity_To_v1_PodAffinity is an autogenerated conversion function.
-func Convert_core_PodAffinity_To_v1_PodAffinity(in *core.PodAffinity, out *v1.PodAffinity, s conversion.Scope) error {
- return autoConvert_core_PodAffinity_To_v1_PodAffinity(in, out, s)
-}
-
-func autoConvert_v1_PodAffinityTerm_To_core_PodAffinityTerm(in *v1.PodAffinityTerm, out *core.PodAffinityTerm, s conversion.Scope) error {
- out.LabelSelector = (*meta_v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
- out.Namespaces = *(*[]string)(unsafe.Pointer(&in.Namespaces))
- out.TopologyKey = in.TopologyKey
- return nil
-}
-
-// Convert_v1_PodAffinityTerm_To_core_PodAffinityTerm is an autogenerated conversion function.
-func Convert_v1_PodAffinityTerm_To_core_PodAffinityTerm(in *v1.PodAffinityTerm, out *core.PodAffinityTerm, s conversion.Scope) error {
- return autoConvert_v1_PodAffinityTerm_To_core_PodAffinityTerm(in, out, s)
-}
-
-func autoConvert_core_PodAffinityTerm_To_v1_PodAffinityTerm(in *core.PodAffinityTerm, out *v1.PodAffinityTerm, s conversion.Scope) error {
- out.LabelSelector = (*meta_v1.LabelSelector)(unsafe.Pointer(in.LabelSelector))
- out.Namespaces = *(*[]string)(unsafe.Pointer(&in.Namespaces))
- out.TopologyKey = in.TopologyKey
- return nil
-}
-
-// Convert_core_PodAffinityTerm_To_v1_PodAffinityTerm is an autogenerated conversion function.
-func Convert_core_PodAffinityTerm_To_v1_PodAffinityTerm(in *core.PodAffinityTerm, out *v1.PodAffinityTerm, s conversion.Scope) error {
- return autoConvert_core_PodAffinityTerm_To_v1_PodAffinityTerm(in, out, s)
-}
-
-func autoConvert_v1_PodAntiAffinity_To_core_PodAntiAffinity(in *v1.PodAntiAffinity, out *core.PodAntiAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = *(*[]core.PodAffinityTerm)(unsafe.Pointer(&in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]core.WeightedPodAffinityTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_v1_PodAntiAffinity_To_core_PodAntiAffinity is an autogenerated conversion function.
-func Convert_v1_PodAntiAffinity_To_core_PodAntiAffinity(in *v1.PodAntiAffinity, out *core.PodAntiAffinity, s conversion.Scope) error {
- return autoConvert_v1_PodAntiAffinity_To_core_PodAntiAffinity(in, out, s)
-}
-
-func autoConvert_core_PodAntiAffinity_To_v1_PodAntiAffinity(in *core.PodAntiAffinity, out *v1.PodAntiAffinity, s conversion.Scope) error {
- out.RequiredDuringSchedulingIgnoredDuringExecution = *(*[]v1.PodAffinityTerm)(unsafe.Pointer(&in.RequiredDuringSchedulingIgnoredDuringExecution))
- out.PreferredDuringSchedulingIgnoredDuringExecution = *(*[]v1.WeightedPodAffinityTerm)(unsafe.Pointer(&in.PreferredDuringSchedulingIgnoredDuringExecution))
- return nil
-}
-
-// Convert_core_PodAntiAffinity_To_v1_PodAntiAffinity is an autogenerated conversion function.
-func Convert_core_PodAntiAffinity_To_v1_PodAntiAffinity(in *core.PodAntiAffinity, out *v1.PodAntiAffinity, s conversion.Scope) error {
- return autoConvert_core_PodAntiAffinity_To_v1_PodAntiAffinity(in, out, s)
-}
-
-func autoConvert_v1_PodAttachOptions_To_core_PodAttachOptions(in *v1.PodAttachOptions, out *core.PodAttachOptions, s conversion.Scope) error {
- out.Stdin = in.Stdin
- out.Stdout = in.Stdout
- out.Stderr = in.Stderr
- out.TTY = in.TTY
- out.Container = in.Container
- return nil
-}
-
-// Convert_v1_PodAttachOptions_To_core_PodAttachOptions is an autogenerated conversion function.
-func Convert_v1_PodAttachOptions_To_core_PodAttachOptions(in *v1.PodAttachOptions, out *core.PodAttachOptions, s conversion.Scope) error {
- return autoConvert_v1_PodAttachOptions_To_core_PodAttachOptions(in, out, s)
-}
-
-func autoConvert_core_PodAttachOptions_To_v1_PodAttachOptions(in *core.PodAttachOptions, out *v1.PodAttachOptions, s conversion.Scope) error {
- out.Stdin = in.Stdin
- out.Stdout = in.Stdout
- out.Stderr = in.Stderr
- out.TTY = in.TTY
- out.Container = in.Container
- return nil
-}
-
-// Convert_core_PodAttachOptions_To_v1_PodAttachOptions is an autogenerated conversion function.
-func Convert_core_PodAttachOptions_To_v1_PodAttachOptions(in *core.PodAttachOptions, out *v1.PodAttachOptions, s conversion.Scope) error {
- return autoConvert_core_PodAttachOptions_To_v1_PodAttachOptions(in, out, s)
-}
-
-func autoConvert_v1_PodCondition_To_core_PodCondition(in *v1.PodCondition, out *core.PodCondition, s conversion.Scope) error {
- out.Type = core.PodConditionType(in.Type)
- out.Status = core.ConditionStatus(in.Status)
- out.LastProbeTime = in.LastProbeTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_PodCondition_To_core_PodCondition is an autogenerated conversion function.
-func Convert_v1_PodCondition_To_core_PodCondition(in *v1.PodCondition, out *core.PodCondition, s conversion.Scope) error {
- return autoConvert_v1_PodCondition_To_core_PodCondition(in, out, s)
-}
-
-func autoConvert_core_PodCondition_To_v1_PodCondition(in *core.PodCondition, out *v1.PodCondition, s conversion.Scope) error {
- out.Type = v1.PodConditionType(in.Type)
- out.Status = v1.ConditionStatus(in.Status)
- out.LastProbeTime = in.LastProbeTime
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_PodCondition_To_v1_PodCondition is an autogenerated conversion function.
-func Convert_core_PodCondition_To_v1_PodCondition(in *core.PodCondition, out *v1.PodCondition, s conversion.Scope) error {
- return autoConvert_core_PodCondition_To_v1_PodCondition(in, out, s)
-}
-
-func autoConvert_v1_PodDNSConfig_To_core_PodDNSConfig(in *v1.PodDNSConfig, out *core.PodDNSConfig, s conversion.Scope) error {
- out.Nameservers = *(*[]string)(unsafe.Pointer(&in.Nameservers))
- out.Searches = *(*[]string)(unsafe.Pointer(&in.Searches))
- out.Options = *(*[]core.PodDNSConfigOption)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_v1_PodDNSConfig_To_core_PodDNSConfig is an autogenerated conversion function.
-func Convert_v1_PodDNSConfig_To_core_PodDNSConfig(in *v1.PodDNSConfig, out *core.PodDNSConfig, s conversion.Scope) error {
- return autoConvert_v1_PodDNSConfig_To_core_PodDNSConfig(in, out, s)
-}
-
-func autoConvert_core_PodDNSConfig_To_v1_PodDNSConfig(in *core.PodDNSConfig, out *v1.PodDNSConfig, s conversion.Scope) error {
- out.Nameservers = *(*[]string)(unsafe.Pointer(&in.Nameservers))
- out.Searches = *(*[]string)(unsafe.Pointer(&in.Searches))
- out.Options = *(*[]v1.PodDNSConfigOption)(unsafe.Pointer(&in.Options))
- return nil
-}
-
-// Convert_core_PodDNSConfig_To_v1_PodDNSConfig is an autogenerated conversion function.
-func Convert_core_PodDNSConfig_To_v1_PodDNSConfig(in *core.PodDNSConfig, out *v1.PodDNSConfig, s conversion.Scope) error {
- return autoConvert_core_PodDNSConfig_To_v1_PodDNSConfig(in, out, s)
-}
-
-func autoConvert_v1_PodDNSConfigOption_To_core_PodDNSConfigOption(in *v1.PodDNSConfigOption, out *core.PodDNSConfigOption, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = (*string)(unsafe.Pointer(in.Value))
- return nil
-}
-
-// Convert_v1_PodDNSConfigOption_To_core_PodDNSConfigOption is an autogenerated conversion function.
-func Convert_v1_PodDNSConfigOption_To_core_PodDNSConfigOption(in *v1.PodDNSConfigOption, out *core.PodDNSConfigOption, s conversion.Scope) error {
- return autoConvert_v1_PodDNSConfigOption_To_core_PodDNSConfigOption(in, out, s)
-}
-
-func autoConvert_core_PodDNSConfigOption_To_v1_PodDNSConfigOption(in *core.PodDNSConfigOption, out *v1.PodDNSConfigOption, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = (*string)(unsafe.Pointer(in.Value))
- return nil
-}
-
-// Convert_core_PodDNSConfigOption_To_v1_PodDNSConfigOption is an autogenerated conversion function.
-func Convert_core_PodDNSConfigOption_To_v1_PodDNSConfigOption(in *core.PodDNSConfigOption, out *v1.PodDNSConfigOption, s conversion.Scope) error {
- return autoConvert_core_PodDNSConfigOption_To_v1_PodDNSConfigOption(in, out, s)
-}
-
-func autoConvert_v1_PodExecOptions_To_core_PodExecOptions(in *v1.PodExecOptions, out *core.PodExecOptions, s conversion.Scope) error {
- out.Stdin = in.Stdin
- out.Stdout = in.Stdout
- out.Stderr = in.Stderr
- out.TTY = in.TTY
- out.Container = in.Container
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- return nil
-}
-
-// Convert_v1_PodExecOptions_To_core_PodExecOptions is an autogenerated conversion function.
-func Convert_v1_PodExecOptions_To_core_PodExecOptions(in *v1.PodExecOptions, out *core.PodExecOptions, s conversion.Scope) error {
- return autoConvert_v1_PodExecOptions_To_core_PodExecOptions(in, out, s)
-}
-
-func autoConvert_core_PodExecOptions_To_v1_PodExecOptions(in *core.PodExecOptions, out *v1.PodExecOptions, s conversion.Scope) error {
- out.Stdin = in.Stdin
- out.Stdout = in.Stdout
- out.Stderr = in.Stderr
- out.TTY = in.TTY
- out.Container = in.Container
- out.Command = *(*[]string)(unsafe.Pointer(&in.Command))
- return nil
-}
-
-// Convert_core_PodExecOptions_To_v1_PodExecOptions is an autogenerated conversion function.
-func Convert_core_PodExecOptions_To_v1_PodExecOptions(in *core.PodExecOptions, out *v1.PodExecOptions, s conversion.Scope) error {
- return autoConvert_core_PodExecOptions_To_v1_PodExecOptions(in, out, s)
-}
-
-func autoConvert_v1_PodList_To_core_PodList(in *v1.PodList, out *core.PodList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.Pod, len(*in))
- for i := range *in {
- if err := Convert_v1_Pod_To_core_Pod(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_PodList_To_core_PodList is an autogenerated conversion function.
-func Convert_v1_PodList_To_core_PodList(in *v1.PodList, out *core.PodList, s conversion.Scope) error {
- return autoConvert_v1_PodList_To_core_PodList(in, out, s)
-}
-
-func autoConvert_core_PodList_To_v1_PodList(in *core.PodList, out *v1.PodList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.Pod, len(*in))
- for i := range *in {
- if err := Convert_core_Pod_To_v1_Pod(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_PodList_To_v1_PodList is an autogenerated conversion function.
-func Convert_core_PodList_To_v1_PodList(in *core.PodList, out *v1.PodList, s conversion.Scope) error {
- return autoConvert_core_PodList_To_v1_PodList(in, out, s)
-}
-
-func autoConvert_v1_PodLogOptions_To_core_PodLogOptions(in *v1.PodLogOptions, out *core.PodLogOptions, s conversion.Scope) error {
- out.Container = in.Container
- out.Follow = in.Follow
- out.Previous = in.Previous
- out.SinceSeconds = (*int64)(unsafe.Pointer(in.SinceSeconds))
- out.SinceTime = (*meta_v1.Time)(unsafe.Pointer(in.SinceTime))
- out.Timestamps = in.Timestamps
- out.TailLines = (*int64)(unsafe.Pointer(in.TailLines))
- out.LimitBytes = (*int64)(unsafe.Pointer(in.LimitBytes))
- return nil
-}
-
-// Convert_v1_PodLogOptions_To_core_PodLogOptions is an autogenerated conversion function.
-func Convert_v1_PodLogOptions_To_core_PodLogOptions(in *v1.PodLogOptions, out *core.PodLogOptions, s conversion.Scope) error {
- return autoConvert_v1_PodLogOptions_To_core_PodLogOptions(in, out, s)
-}
-
-func autoConvert_core_PodLogOptions_To_v1_PodLogOptions(in *core.PodLogOptions, out *v1.PodLogOptions, s conversion.Scope) error {
- out.Container = in.Container
- out.Follow = in.Follow
- out.Previous = in.Previous
- out.SinceSeconds = (*int64)(unsafe.Pointer(in.SinceSeconds))
- out.SinceTime = (*meta_v1.Time)(unsafe.Pointer(in.SinceTime))
- out.Timestamps = in.Timestamps
- out.TailLines = (*int64)(unsafe.Pointer(in.TailLines))
- out.LimitBytes = (*int64)(unsafe.Pointer(in.LimitBytes))
- return nil
-}
-
-// Convert_core_PodLogOptions_To_v1_PodLogOptions is an autogenerated conversion function.
-func Convert_core_PodLogOptions_To_v1_PodLogOptions(in *core.PodLogOptions, out *v1.PodLogOptions, s conversion.Scope) error {
- return autoConvert_core_PodLogOptions_To_v1_PodLogOptions(in, out, s)
-}
-
-func autoConvert_v1_PodPortForwardOptions_To_core_PodPortForwardOptions(in *v1.PodPortForwardOptions, out *core.PodPortForwardOptions, s conversion.Scope) error {
- out.Ports = *(*[]int32)(unsafe.Pointer(&in.Ports))
- return nil
-}
-
-// Convert_v1_PodPortForwardOptions_To_core_PodPortForwardOptions is an autogenerated conversion function.
-func Convert_v1_PodPortForwardOptions_To_core_PodPortForwardOptions(in *v1.PodPortForwardOptions, out *core.PodPortForwardOptions, s conversion.Scope) error {
- return autoConvert_v1_PodPortForwardOptions_To_core_PodPortForwardOptions(in, out, s)
-}
-
-func autoConvert_core_PodPortForwardOptions_To_v1_PodPortForwardOptions(in *core.PodPortForwardOptions, out *v1.PodPortForwardOptions, s conversion.Scope) error {
- out.Ports = *(*[]int32)(unsafe.Pointer(&in.Ports))
- return nil
-}
-
-// Convert_core_PodPortForwardOptions_To_v1_PodPortForwardOptions is an autogenerated conversion function.
-func Convert_core_PodPortForwardOptions_To_v1_PodPortForwardOptions(in *core.PodPortForwardOptions, out *v1.PodPortForwardOptions, s conversion.Scope) error {
- return autoConvert_core_PodPortForwardOptions_To_v1_PodPortForwardOptions(in, out, s)
-}
-
-func autoConvert_v1_PodProxyOptions_To_core_PodProxyOptions(in *v1.PodProxyOptions, out *core.PodProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_v1_PodProxyOptions_To_core_PodProxyOptions is an autogenerated conversion function.
-func Convert_v1_PodProxyOptions_To_core_PodProxyOptions(in *v1.PodProxyOptions, out *core.PodProxyOptions, s conversion.Scope) error {
- return autoConvert_v1_PodProxyOptions_To_core_PodProxyOptions(in, out, s)
-}
-
-func autoConvert_core_PodProxyOptions_To_v1_PodProxyOptions(in *core.PodProxyOptions, out *v1.PodProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_core_PodProxyOptions_To_v1_PodProxyOptions is an autogenerated conversion function.
-func Convert_core_PodProxyOptions_To_v1_PodProxyOptions(in *core.PodProxyOptions, out *v1.PodProxyOptions, s conversion.Scope) error {
- return autoConvert_core_PodProxyOptions_To_v1_PodProxyOptions(in, out, s)
-}
-
-func autoConvert_v1_PodSecurityContext_To_core_PodSecurityContext(in *v1.PodSecurityContext, out *core.PodSecurityContext, s conversion.Scope) error {
- out.SELinuxOptions = (*core.SELinuxOptions)(unsafe.Pointer(in.SELinuxOptions))
- out.RunAsUser = (*int64)(unsafe.Pointer(in.RunAsUser))
- out.RunAsNonRoot = (*bool)(unsafe.Pointer(in.RunAsNonRoot))
- out.SupplementalGroups = *(*[]int64)(unsafe.Pointer(&in.SupplementalGroups))
- out.FSGroup = (*int64)(unsafe.Pointer(in.FSGroup))
- return nil
-}
-
-func autoConvert_core_PodSecurityContext_To_v1_PodSecurityContext(in *core.PodSecurityContext, out *v1.PodSecurityContext, s conversion.Scope) error {
- // INFO: in.HostNetwork opted out of conversion generation
- // INFO: in.HostPID opted out of conversion generation
- // INFO: in.HostIPC opted out of conversion generation
- // INFO: in.ShareProcessNamespace opted out of conversion generation
- out.SELinuxOptions = (*v1.SELinuxOptions)(unsafe.Pointer(in.SELinuxOptions))
- out.RunAsUser = (*int64)(unsafe.Pointer(in.RunAsUser))
- out.RunAsNonRoot = (*bool)(unsafe.Pointer(in.RunAsNonRoot))
- out.SupplementalGroups = *(*[]int64)(unsafe.Pointer(&in.SupplementalGroups))
- out.FSGroup = (*int64)(unsafe.Pointer(in.FSGroup))
- return nil
-}
-
-func autoConvert_v1_PodSignature_To_core_PodSignature(in *v1.PodSignature, out *core.PodSignature, s conversion.Scope) error {
- out.PodController = (*meta_v1.OwnerReference)(unsafe.Pointer(in.PodController))
- return nil
-}
-
-// Convert_v1_PodSignature_To_core_PodSignature is an autogenerated conversion function.
-func Convert_v1_PodSignature_To_core_PodSignature(in *v1.PodSignature, out *core.PodSignature, s conversion.Scope) error {
- return autoConvert_v1_PodSignature_To_core_PodSignature(in, out, s)
-}
-
-func autoConvert_core_PodSignature_To_v1_PodSignature(in *core.PodSignature, out *v1.PodSignature, s conversion.Scope) error {
- out.PodController = (*meta_v1.OwnerReference)(unsafe.Pointer(in.PodController))
- return nil
-}
-
-// Convert_core_PodSignature_To_v1_PodSignature is an autogenerated conversion function.
-func Convert_core_PodSignature_To_v1_PodSignature(in *core.PodSignature, out *v1.PodSignature, s conversion.Scope) error {
- return autoConvert_core_PodSignature_To_v1_PodSignature(in, out, s)
-}
-
-func autoConvert_v1_PodSpec_To_core_PodSpec(in *v1.PodSpec, out *core.PodSpec, s conversion.Scope) error {
- if in.Volumes != nil {
- in, out := &in.Volumes, &out.Volumes
- *out = make([]core.Volume, len(*in))
- for i := range *in {
- if err := Convert_v1_Volume_To_core_Volume(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Volumes = nil
- }
- if in.InitContainers != nil {
- in, out := &in.InitContainers, &out.InitContainers
- *out = make([]core.Container, len(*in))
- for i := range *in {
- if err := Convert_v1_Container_To_core_Container(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.InitContainers = nil
- }
- if in.Containers != nil {
- in, out := &in.Containers, &out.Containers
- *out = make([]core.Container, len(*in))
- for i := range *in {
- if err := Convert_v1_Container_To_core_Container(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Containers = nil
- }
- out.RestartPolicy = core.RestartPolicy(in.RestartPolicy)
- out.TerminationGracePeriodSeconds = (*int64)(unsafe.Pointer(in.TerminationGracePeriodSeconds))
- out.ActiveDeadlineSeconds = (*int64)(unsafe.Pointer(in.ActiveDeadlineSeconds))
- out.DNSPolicy = core.DNSPolicy(in.DNSPolicy)
- out.NodeSelector = *(*map[string]string)(unsafe.Pointer(&in.NodeSelector))
- out.ServiceAccountName = in.ServiceAccountName
- // INFO: in.DeprecatedServiceAccount opted out of conversion generation
- out.AutomountServiceAccountToken = (*bool)(unsafe.Pointer(in.AutomountServiceAccountToken))
- out.NodeName = in.NodeName
- // INFO: in.HostNetwork opted out of conversion generation
- // INFO: in.HostPID opted out of conversion generation
- // INFO: in.HostIPC opted out of conversion generation
- // INFO: in.ShareProcessNamespace opted out of conversion generation
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- *out = new(core.PodSecurityContext)
- if err := Convert_v1_PodSecurityContext_To_core_PodSecurityContext(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.SecurityContext = nil
- }
- out.ImagePullSecrets = *(*[]core.LocalObjectReference)(unsafe.Pointer(&in.ImagePullSecrets))
- out.Hostname = in.Hostname
- out.Subdomain = in.Subdomain
- out.Affinity = (*core.Affinity)(unsafe.Pointer(in.Affinity))
- out.SchedulerName = in.SchedulerName
- out.Tolerations = *(*[]core.Toleration)(unsafe.Pointer(&in.Tolerations))
- out.HostAliases = *(*[]core.HostAlias)(unsafe.Pointer(&in.HostAliases))
- out.PriorityClassName = in.PriorityClassName
- out.Priority = (*int32)(unsafe.Pointer(in.Priority))
- out.DNSConfig = (*core.PodDNSConfig)(unsafe.Pointer(in.DNSConfig))
- return nil
-}
-
-func autoConvert_core_PodSpec_To_v1_PodSpec(in *core.PodSpec, out *v1.PodSpec, s conversion.Scope) error {
- if in.Volumes != nil {
- in, out := &in.Volumes, &out.Volumes
- *out = make([]v1.Volume, len(*in))
- for i := range *in {
- if err := Convert_core_Volume_To_v1_Volume(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Volumes = nil
- }
- if in.InitContainers != nil {
- in, out := &in.InitContainers, &out.InitContainers
- *out = make([]v1.Container, len(*in))
- for i := range *in {
- if err := Convert_core_Container_To_v1_Container(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.InitContainers = nil
- }
- if in.Containers != nil {
- in, out := &in.Containers, &out.Containers
- *out = make([]v1.Container, len(*in))
- for i := range *in {
- if err := Convert_core_Container_To_v1_Container(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Containers = nil
- }
- out.RestartPolicy = v1.RestartPolicy(in.RestartPolicy)
- out.TerminationGracePeriodSeconds = (*int64)(unsafe.Pointer(in.TerminationGracePeriodSeconds))
- out.ActiveDeadlineSeconds = (*int64)(unsafe.Pointer(in.ActiveDeadlineSeconds))
- out.DNSPolicy = v1.DNSPolicy(in.DNSPolicy)
- out.NodeSelector = *(*map[string]string)(unsafe.Pointer(&in.NodeSelector))
- out.ServiceAccountName = in.ServiceAccountName
- out.AutomountServiceAccountToken = (*bool)(unsafe.Pointer(in.AutomountServiceAccountToken))
- out.NodeName = in.NodeName
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- *out = new(v1.PodSecurityContext)
- if err := Convert_core_PodSecurityContext_To_v1_PodSecurityContext(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.SecurityContext = nil
- }
- out.ImagePullSecrets = *(*[]v1.LocalObjectReference)(unsafe.Pointer(&in.ImagePullSecrets))
- out.Hostname = in.Hostname
- out.Subdomain = in.Subdomain
- out.Affinity = (*v1.Affinity)(unsafe.Pointer(in.Affinity))
- out.SchedulerName = in.SchedulerName
- out.Tolerations = *(*[]v1.Toleration)(unsafe.Pointer(&in.Tolerations))
- out.HostAliases = *(*[]v1.HostAlias)(unsafe.Pointer(&in.HostAliases))
- out.PriorityClassName = in.PriorityClassName
- out.Priority = (*int32)(unsafe.Pointer(in.Priority))
- out.DNSConfig = (*v1.PodDNSConfig)(unsafe.Pointer(in.DNSConfig))
- return nil
-}
-
-func autoConvert_v1_PodStatus_To_core_PodStatus(in *v1.PodStatus, out *core.PodStatus, s conversion.Scope) error {
- out.Phase = core.PodPhase(in.Phase)
- out.Conditions = *(*[]core.PodCondition)(unsafe.Pointer(&in.Conditions))
- out.Message = in.Message
- out.Reason = in.Reason
- out.NominatedNodeName = in.NominatedNodeName
- out.HostIP = in.HostIP
- out.PodIP = in.PodIP
- out.StartTime = (*meta_v1.Time)(unsafe.Pointer(in.StartTime))
- out.InitContainerStatuses = *(*[]core.ContainerStatus)(unsafe.Pointer(&in.InitContainerStatuses))
- out.ContainerStatuses = *(*[]core.ContainerStatus)(unsafe.Pointer(&in.ContainerStatuses))
- out.QOSClass = core.PodQOSClass(in.QOSClass)
- return nil
-}
-
-// Convert_v1_PodStatus_To_core_PodStatus is an autogenerated conversion function.
-func Convert_v1_PodStatus_To_core_PodStatus(in *v1.PodStatus, out *core.PodStatus, s conversion.Scope) error {
- return autoConvert_v1_PodStatus_To_core_PodStatus(in, out, s)
-}
-
-func autoConvert_core_PodStatus_To_v1_PodStatus(in *core.PodStatus, out *v1.PodStatus, s conversion.Scope) error {
- out.Phase = v1.PodPhase(in.Phase)
- out.Conditions = *(*[]v1.PodCondition)(unsafe.Pointer(&in.Conditions))
- out.Message = in.Message
- out.Reason = in.Reason
- out.NominatedNodeName = in.NominatedNodeName
- out.HostIP = in.HostIP
- out.PodIP = in.PodIP
- out.StartTime = (*meta_v1.Time)(unsafe.Pointer(in.StartTime))
- out.QOSClass = v1.PodQOSClass(in.QOSClass)
- out.InitContainerStatuses = *(*[]v1.ContainerStatus)(unsafe.Pointer(&in.InitContainerStatuses))
- out.ContainerStatuses = *(*[]v1.ContainerStatus)(unsafe.Pointer(&in.ContainerStatuses))
- return nil
-}
-
-// Convert_core_PodStatus_To_v1_PodStatus is an autogenerated conversion function.
-func Convert_core_PodStatus_To_v1_PodStatus(in *core.PodStatus, out *v1.PodStatus, s conversion.Scope) error {
- return autoConvert_core_PodStatus_To_v1_PodStatus(in, out, s)
-}
-
-func autoConvert_v1_PodStatusResult_To_core_PodStatusResult(in *v1.PodStatusResult, out *core.PodStatusResult, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PodStatus_To_core_PodStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_PodStatusResult_To_core_PodStatusResult is an autogenerated conversion function.
-func Convert_v1_PodStatusResult_To_core_PodStatusResult(in *v1.PodStatusResult, out *core.PodStatusResult, s conversion.Scope) error {
- return autoConvert_v1_PodStatusResult_To_core_PodStatusResult(in, out, s)
-}
-
-func autoConvert_core_PodStatusResult_To_v1_PodStatusResult(in *core.PodStatusResult, out *v1.PodStatusResult, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PodStatus_To_v1_PodStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_PodStatusResult_To_v1_PodStatusResult is an autogenerated conversion function.
-func Convert_core_PodStatusResult_To_v1_PodStatusResult(in *core.PodStatusResult, out *v1.PodStatusResult, s conversion.Scope) error {
- return autoConvert_core_PodStatusResult_To_v1_PodStatusResult(in, out, s)
-}
-
-func autoConvert_v1_PodTemplate_To_core_PodTemplate(in *v1.PodTemplate, out *core.PodTemplate, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec(&in.Template, &out.Template, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_PodTemplate_To_core_PodTemplate is an autogenerated conversion function.
-func Convert_v1_PodTemplate_To_core_PodTemplate(in *v1.PodTemplate, out *core.PodTemplate, s conversion.Scope) error {
- return autoConvert_v1_PodTemplate_To_core_PodTemplate(in, out, s)
-}
-
-func autoConvert_core_PodTemplate_To_v1_PodTemplate(in *core.PodTemplate, out *v1.PodTemplate, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec(&in.Template, &out.Template, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_PodTemplate_To_v1_PodTemplate is an autogenerated conversion function.
-func Convert_core_PodTemplate_To_v1_PodTemplate(in *core.PodTemplate, out *v1.PodTemplate, s conversion.Scope) error {
- return autoConvert_core_PodTemplate_To_v1_PodTemplate(in, out, s)
-}
-
-func autoConvert_v1_PodTemplateList_To_core_PodTemplateList(in *v1.PodTemplateList, out *core.PodTemplateList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.PodTemplate, len(*in))
- for i := range *in {
- if err := Convert_v1_PodTemplate_To_core_PodTemplate(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_PodTemplateList_To_core_PodTemplateList is an autogenerated conversion function.
-func Convert_v1_PodTemplateList_To_core_PodTemplateList(in *v1.PodTemplateList, out *core.PodTemplateList, s conversion.Scope) error {
- return autoConvert_v1_PodTemplateList_To_core_PodTemplateList(in, out, s)
-}
-
-func autoConvert_core_PodTemplateList_To_v1_PodTemplateList(in *core.PodTemplateList, out *v1.PodTemplateList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.PodTemplate, len(*in))
- for i := range *in {
- if err := Convert_core_PodTemplate_To_v1_PodTemplate(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_PodTemplateList_To_v1_PodTemplateList is an autogenerated conversion function.
-func Convert_core_PodTemplateList_To_v1_PodTemplateList(in *core.PodTemplateList, out *v1.PodTemplateList, s conversion.Scope) error {
- return autoConvert_core_PodTemplateList_To_v1_PodTemplateList(in, out, s)
-}
-
-func autoConvert_v1_PodTemplateSpec_To_core_PodTemplateSpec(in *v1.PodTemplateSpec, out *core.PodTemplateSpec, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_PodSpec_To_core_PodSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- return nil
-}
-
-func autoConvert_core_PodTemplateSpec_To_v1_PodTemplateSpec(in *core.PodTemplateSpec, out *v1.PodTemplateSpec, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_PodSpec_To_v1_PodSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- return nil
-}
-
-func autoConvert_v1_PortworxVolumeSource_To_core_PortworxVolumeSource(in *v1.PortworxVolumeSource, out *core.PortworxVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_PortworxVolumeSource_To_core_PortworxVolumeSource is an autogenerated conversion function.
-func Convert_v1_PortworxVolumeSource_To_core_PortworxVolumeSource(in *v1.PortworxVolumeSource, out *core.PortworxVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_PortworxVolumeSource_To_core_PortworxVolumeSource(in, out, s)
-}
-
-func autoConvert_core_PortworxVolumeSource_To_v1_PortworxVolumeSource(in *core.PortworxVolumeSource, out *v1.PortworxVolumeSource, s conversion.Scope) error {
- out.VolumeID = in.VolumeID
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_PortworxVolumeSource_To_v1_PortworxVolumeSource is an autogenerated conversion function.
-func Convert_core_PortworxVolumeSource_To_v1_PortworxVolumeSource(in *core.PortworxVolumeSource, out *v1.PortworxVolumeSource, s conversion.Scope) error {
- return autoConvert_core_PortworxVolumeSource_To_v1_PortworxVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Preconditions_To_core_Preconditions(in *v1.Preconditions, out *core.Preconditions, s conversion.Scope) error {
- out.UID = (*types.UID)(unsafe.Pointer(in.UID))
- return nil
-}
-
-// Convert_v1_Preconditions_To_core_Preconditions is an autogenerated conversion function.
-func Convert_v1_Preconditions_To_core_Preconditions(in *v1.Preconditions, out *core.Preconditions, s conversion.Scope) error {
- return autoConvert_v1_Preconditions_To_core_Preconditions(in, out, s)
-}
-
-func autoConvert_core_Preconditions_To_v1_Preconditions(in *core.Preconditions, out *v1.Preconditions, s conversion.Scope) error {
- out.UID = (*types.UID)(unsafe.Pointer(in.UID))
- return nil
-}
-
-// Convert_core_Preconditions_To_v1_Preconditions is an autogenerated conversion function.
-func Convert_core_Preconditions_To_v1_Preconditions(in *core.Preconditions, out *v1.Preconditions, s conversion.Scope) error {
- return autoConvert_core_Preconditions_To_v1_Preconditions(in, out, s)
-}
-
-func autoConvert_v1_PreferAvoidPodsEntry_To_core_PreferAvoidPodsEntry(in *v1.PreferAvoidPodsEntry, out *core.PreferAvoidPodsEntry, s conversion.Scope) error {
- if err := Convert_v1_PodSignature_To_core_PodSignature(&in.PodSignature, &out.PodSignature, s); err != nil {
- return err
- }
- out.EvictionTime = in.EvictionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_PreferAvoidPodsEntry_To_core_PreferAvoidPodsEntry is an autogenerated conversion function.
-func Convert_v1_PreferAvoidPodsEntry_To_core_PreferAvoidPodsEntry(in *v1.PreferAvoidPodsEntry, out *core.PreferAvoidPodsEntry, s conversion.Scope) error {
- return autoConvert_v1_PreferAvoidPodsEntry_To_core_PreferAvoidPodsEntry(in, out, s)
-}
-
-func autoConvert_core_PreferAvoidPodsEntry_To_v1_PreferAvoidPodsEntry(in *core.PreferAvoidPodsEntry, out *v1.PreferAvoidPodsEntry, s conversion.Scope) error {
- if err := Convert_core_PodSignature_To_v1_PodSignature(&in.PodSignature, &out.PodSignature, s); err != nil {
- return err
- }
- out.EvictionTime = in.EvictionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_PreferAvoidPodsEntry_To_v1_PreferAvoidPodsEntry is an autogenerated conversion function.
-func Convert_core_PreferAvoidPodsEntry_To_v1_PreferAvoidPodsEntry(in *core.PreferAvoidPodsEntry, out *v1.PreferAvoidPodsEntry, s conversion.Scope) error {
- return autoConvert_core_PreferAvoidPodsEntry_To_v1_PreferAvoidPodsEntry(in, out, s)
-}
-
-func autoConvert_v1_PreferredSchedulingTerm_To_core_PreferredSchedulingTerm(in *v1.PreferredSchedulingTerm, out *core.PreferredSchedulingTerm, s conversion.Scope) error {
- out.Weight = in.Weight
- if err := Convert_v1_NodeSelectorTerm_To_core_NodeSelectorTerm(&in.Preference, &out.Preference, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_PreferredSchedulingTerm_To_core_PreferredSchedulingTerm is an autogenerated conversion function.
-func Convert_v1_PreferredSchedulingTerm_To_core_PreferredSchedulingTerm(in *v1.PreferredSchedulingTerm, out *core.PreferredSchedulingTerm, s conversion.Scope) error {
- return autoConvert_v1_PreferredSchedulingTerm_To_core_PreferredSchedulingTerm(in, out, s)
-}
-
-func autoConvert_core_PreferredSchedulingTerm_To_v1_PreferredSchedulingTerm(in *core.PreferredSchedulingTerm, out *v1.PreferredSchedulingTerm, s conversion.Scope) error {
- out.Weight = in.Weight
- if err := Convert_core_NodeSelectorTerm_To_v1_NodeSelectorTerm(&in.Preference, &out.Preference, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_PreferredSchedulingTerm_To_v1_PreferredSchedulingTerm is an autogenerated conversion function.
-func Convert_core_PreferredSchedulingTerm_To_v1_PreferredSchedulingTerm(in *core.PreferredSchedulingTerm, out *v1.PreferredSchedulingTerm, s conversion.Scope) error {
- return autoConvert_core_PreferredSchedulingTerm_To_v1_PreferredSchedulingTerm(in, out, s)
-}
-
-func autoConvert_v1_Probe_To_core_Probe(in *v1.Probe, out *core.Probe, s conversion.Scope) error {
- if err := Convert_v1_Handler_To_core_Handler(&in.Handler, &out.Handler, s); err != nil {
- return err
- }
- out.InitialDelaySeconds = in.InitialDelaySeconds
- out.TimeoutSeconds = in.TimeoutSeconds
- out.PeriodSeconds = in.PeriodSeconds
- out.SuccessThreshold = in.SuccessThreshold
- out.FailureThreshold = in.FailureThreshold
- return nil
-}
-
-// Convert_v1_Probe_To_core_Probe is an autogenerated conversion function.
-func Convert_v1_Probe_To_core_Probe(in *v1.Probe, out *core.Probe, s conversion.Scope) error {
- return autoConvert_v1_Probe_To_core_Probe(in, out, s)
-}
-
-func autoConvert_core_Probe_To_v1_Probe(in *core.Probe, out *v1.Probe, s conversion.Scope) error {
- if err := Convert_core_Handler_To_v1_Handler(&in.Handler, &out.Handler, s); err != nil {
- return err
- }
- out.InitialDelaySeconds = in.InitialDelaySeconds
- out.TimeoutSeconds = in.TimeoutSeconds
- out.PeriodSeconds = in.PeriodSeconds
- out.SuccessThreshold = in.SuccessThreshold
- out.FailureThreshold = in.FailureThreshold
- return nil
-}
-
-// Convert_core_Probe_To_v1_Probe is an autogenerated conversion function.
-func Convert_core_Probe_To_v1_Probe(in *core.Probe, out *v1.Probe, s conversion.Scope) error {
- return autoConvert_core_Probe_To_v1_Probe(in, out, s)
-}
-
-func autoConvert_v1_ProjectedVolumeSource_To_core_ProjectedVolumeSource(in *v1.ProjectedVolumeSource, out *core.ProjectedVolumeSource, s conversion.Scope) error {
- out.Sources = *(*[]core.VolumeProjection)(unsafe.Pointer(&in.Sources))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- return nil
-}
-
-// Convert_v1_ProjectedVolumeSource_To_core_ProjectedVolumeSource is an autogenerated conversion function.
-func Convert_v1_ProjectedVolumeSource_To_core_ProjectedVolumeSource(in *v1.ProjectedVolumeSource, out *core.ProjectedVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ProjectedVolumeSource_To_core_ProjectedVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ProjectedVolumeSource_To_v1_ProjectedVolumeSource(in *core.ProjectedVolumeSource, out *v1.ProjectedVolumeSource, s conversion.Scope) error {
- out.Sources = *(*[]v1.VolumeProjection)(unsafe.Pointer(&in.Sources))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- return nil
-}
-
-// Convert_core_ProjectedVolumeSource_To_v1_ProjectedVolumeSource is an autogenerated conversion function.
-func Convert_core_ProjectedVolumeSource_To_v1_ProjectedVolumeSource(in *core.ProjectedVolumeSource, out *v1.ProjectedVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ProjectedVolumeSource_To_v1_ProjectedVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_QuobyteVolumeSource_To_core_QuobyteVolumeSource(in *v1.QuobyteVolumeSource, out *core.QuobyteVolumeSource, s conversion.Scope) error {
- out.Registry = in.Registry
- out.Volume = in.Volume
- out.ReadOnly = in.ReadOnly
- out.User = in.User
- out.Group = in.Group
- return nil
-}
-
-// Convert_v1_QuobyteVolumeSource_To_core_QuobyteVolumeSource is an autogenerated conversion function.
-func Convert_v1_QuobyteVolumeSource_To_core_QuobyteVolumeSource(in *v1.QuobyteVolumeSource, out *core.QuobyteVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_QuobyteVolumeSource_To_core_QuobyteVolumeSource(in, out, s)
-}
-
-func autoConvert_core_QuobyteVolumeSource_To_v1_QuobyteVolumeSource(in *core.QuobyteVolumeSource, out *v1.QuobyteVolumeSource, s conversion.Scope) error {
- out.Registry = in.Registry
- out.Volume = in.Volume
- out.ReadOnly = in.ReadOnly
- out.User = in.User
- out.Group = in.Group
- return nil
-}
-
-// Convert_core_QuobyteVolumeSource_To_v1_QuobyteVolumeSource is an autogenerated conversion function.
-func Convert_core_QuobyteVolumeSource_To_v1_QuobyteVolumeSource(in *core.QuobyteVolumeSource, out *v1.QuobyteVolumeSource, s conversion.Scope) error {
- return autoConvert_core_QuobyteVolumeSource_To_v1_QuobyteVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_RBDPersistentVolumeSource_To_core_RBDPersistentVolumeSource(in *v1.RBDPersistentVolumeSource, out *core.RBDPersistentVolumeSource, s conversion.Scope) error {
- out.CephMonitors = *(*[]string)(unsafe.Pointer(&in.CephMonitors))
- out.RBDImage = in.RBDImage
- out.FSType = in.FSType
- out.RBDPool = in.RBDPool
- out.RadosUser = in.RadosUser
- out.Keyring = in.Keyring
- out.SecretRef = (*core.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_RBDPersistentVolumeSource_To_core_RBDPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_RBDPersistentVolumeSource_To_core_RBDPersistentVolumeSource(in *v1.RBDPersistentVolumeSource, out *core.RBDPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_RBDPersistentVolumeSource_To_core_RBDPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_RBDPersistentVolumeSource_To_v1_RBDPersistentVolumeSource(in *core.RBDPersistentVolumeSource, out *v1.RBDPersistentVolumeSource, s conversion.Scope) error {
- out.CephMonitors = *(*[]string)(unsafe.Pointer(&in.CephMonitors))
- out.RBDImage = in.RBDImage
- out.FSType = in.FSType
- out.RBDPool = in.RBDPool
- out.RadosUser = in.RadosUser
- out.Keyring = in.Keyring
- out.SecretRef = (*v1.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_RBDPersistentVolumeSource_To_v1_RBDPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_RBDPersistentVolumeSource_To_v1_RBDPersistentVolumeSource(in *core.RBDPersistentVolumeSource, out *v1.RBDPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_RBDPersistentVolumeSource_To_v1_RBDPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_RBDVolumeSource_To_core_RBDVolumeSource(in *v1.RBDVolumeSource, out *core.RBDVolumeSource, s conversion.Scope) error {
- out.CephMonitors = *(*[]string)(unsafe.Pointer(&in.CephMonitors))
- out.RBDImage = in.RBDImage
- out.FSType = in.FSType
- out.RBDPool = in.RBDPool
- out.RadosUser = in.RadosUser
- out.Keyring = in.Keyring
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_RBDVolumeSource_To_core_RBDVolumeSource is an autogenerated conversion function.
-func Convert_v1_RBDVolumeSource_To_core_RBDVolumeSource(in *v1.RBDVolumeSource, out *core.RBDVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_RBDVolumeSource_To_core_RBDVolumeSource(in, out, s)
-}
-
-func autoConvert_core_RBDVolumeSource_To_v1_RBDVolumeSource(in *core.RBDVolumeSource, out *v1.RBDVolumeSource, s conversion.Scope) error {
- out.CephMonitors = *(*[]string)(unsafe.Pointer(&in.CephMonitors))
- out.RBDImage = in.RBDImage
- out.FSType = in.FSType
- out.RBDPool = in.RBDPool
- out.RadosUser = in.RadosUser
- out.Keyring = in.Keyring
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_RBDVolumeSource_To_v1_RBDVolumeSource is an autogenerated conversion function.
-func Convert_core_RBDVolumeSource_To_v1_RBDVolumeSource(in *core.RBDVolumeSource, out *v1.RBDVolumeSource, s conversion.Scope) error {
- return autoConvert_core_RBDVolumeSource_To_v1_RBDVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_RangeAllocation_To_core_RangeAllocation(in *v1.RangeAllocation, out *core.RangeAllocation, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Range = in.Range
- out.Data = *(*[]byte)(unsafe.Pointer(&in.Data))
- return nil
-}
-
-// Convert_v1_RangeAllocation_To_core_RangeAllocation is an autogenerated conversion function.
-func Convert_v1_RangeAllocation_To_core_RangeAllocation(in *v1.RangeAllocation, out *core.RangeAllocation, s conversion.Scope) error {
- return autoConvert_v1_RangeAllocation_To_core_RangeAllocation(in, out, s)
-}
-
-func autoConvert_core_RangeAllocation_To_v1_RangeAllocation(in *core.RangeAllocation, out *v1.RangeAllocation, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Range = in.Range
- out.Data = *(*[]byte)(unsafe.Pointer(&in.Data))
- return nil
-}
-
-// Convert_core_RangeAllocation_To_v1_RangeAllocation is an autogenerated conversion function.
-func Convert_core_RangeAllocation_To_v1_RangeAllocation(in *core.RangeAllocation, out *v1.RangeAllocation, s conversion.Scope) error {
- return autoConvert_core_RangeAllocation_To_v1_RangeAllocation(in, out, s)
-}
-
-func autoConvert_v1_ReplicationController_To_core_ReplicationController(in *v1.ReplicationController, out *core.ReplicationController, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ReplicationControllerSpec_To_core_ReplicationControllerSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_ReplicationController_To_core_ReplicationController is an autogenerated conversion function.
-func Convert_v1_ReplicationController_To_core_ReplicationController(in *v1.ReplicationController, out *core.ReplicationController, s conversion.Scope) error {
- return autoConvert_v1_ReplicationController_To_core_ReplicationController(in, out, s)
-}
-
-func autoConvert_core_ReplicationController_To_v1_ReplicationController(in *core.ReplicationController, out *v1.ReplicationController, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_ReplicationControllerSpec_To_v1_ReplicationControllerSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_ReplicationController_To_v1_ReplicationController is an autogenerated conversion function.
-func Convert_core_ReplicationController_To_v1_ReplicationController(in *core.ReplicationController, out *v1.ReplicationController, s conversion.Scope) error {
- return autoConvert_core_ReplicationController_To_v1_ReplicationController(in, out, s)
-}
-
-func autoConvert_v1_ReplicationControllerCondition_To_core_ReplicationControllerCondition(in *v1.ReplicationControllerCondition, out *core.ReplicationControllerCondition, s conversion.Scope) error {
- out.Type = core.ReplicationControllerConditionType(in.Type)
- out.Status = core.ConditionStatus(in.Status)
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_v1_ReplicationControllerCondition_To_core_ReplicationControllerCondition is an autogenerated conversion function.
-func Convert_v1_ReplicationControllerCondition_To_core_ReplicationControllerCondition(in *v1.ReplicationControllerCondition, out *core.ReplicationControllerCondition, s conversion.Scope) error {
- return autoConvert_v1_ReplicationControllerCondition_To_core_ReplicationControllerCondition(in, out, s)
-}
-
-func autoConvert_core_ReplicationControllerCondition_To_v1_ReplicationControllerCondition(in *core.ReplicationControllerCondition, out *v1.ReplicationControllerCondition, s conversion.Scope) error {
- out.Type = v1.ReplicationControllerConditionType(in.Type)
- out.Status = v1.ConditionStatus(in.Status)
- out.LastTransitionTime = in.LastTransitionTime
- out.Reason = in.Reason
- out.Message = in.Message
- return nil
-}
-
-// Convert_core_ReplicationControllerCondition_To_v1_ReplicationControllerCondition is an autogenerated conversion function.
-func Convert_core_ReplicationControllerCondition_To_v1_ReplicationControllerCondition(in *core.ReplicationControllerCondition, out *v1.ReplicationControllerCondition, s conversion.Scope) error {
- return autoConvert_core_ReplicationControllerCondition_To_v1_ReplicationControllerCondition(in, out, s)
-}
-
-func autoConvert_v1_ReplicationControllerList_To_core_ReplicationControllerList(in *v1.ReplicationControllerList, out *core.ReplicationControllerList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.ReplicationController, len(*in))
- for i := range *in {
- if err := Convert_v1_ReplicationController_To_core_ReplicationController(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_ReplicationControllerList_To_core_ReplicationControllerList is an autogenerated conversion function.
-func Convert_v1_ReplicationControllerList_To_core_ReplicationControllerList(in *v1.ReplicationControllerList, out *core.ReplicationControllerList, s conversion.Scope) error {
- return autoConvert_v1_ReplicationControllerList_To_core_ReplicationControllerList(in, out, s)
-}
-
-func autoConvert_core_ReplicationControllerList_To_v1_ReplicationControllerList(in *core.ReplicationControllerList, out *v1.ReplicationControllerList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.ReplicationController, len(*in))
- for i := range *in {
- if err := Convert_core_ReplicationController_To_v1_ReplicationController(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_ReplicationControllerList_To_v1_ReplicationControllerList is an autogenerated conversion function.
-func Convert_core_ReplicationControllerList_To_v1_ReplicationControllerList(in *core.ReplicationControllerList, out *v1.ReplicationControllerList, s conversion.Scope) error {
- return autoConvert_core_ReplicationControllerList_To_v1_ReplicationControllerList(in, out, s)
-}
-
-func autoConvert_v1_ReplicationControllerSpec_To_core_ReplicationControllerSpec(in *v1.ReplicationControllerSpec, out *core.ReplicationControllerSpec, s conversion.Scope) error {
- if err := meta_v1.Convert_Pointer_int32_To_int32(&in.Replicas, &out.Replicas, s); err != nil {
- return err
- }
- out.MinReadySeconds = in.MinReadySeconds
- out.Selector = *(*map[string]string)(unsafe.Pointer(&in.Selector))
- if in.Template != nil {
- in, out := &in.Template, &out.Template
- *out = new(core.PodTemplateSpec)
- if err := Convert_v1_PodTemplateSpec_To_core_PodTemplateSpec(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.Template = nil
- }
- return nil
-}
-
-func autoConvert_core_ReplicationControllerSpec_To_v1_ReplicationControllerSpec(in *core.ReplicationControllerSpec, out *v1.ReplicationControllerSpec, s conversion.Scope) error {
- if err := meta_v1.Convert_int32_To_Pointer_int32(&in.Replicas, &out.Replicas, s); err != nil {
- return err
- }
- out.MinReadySeconds = in.MinReadySeconds
- out.Selector = *(*map[string]string)(unsafe.Pointer(&in.Selector))
- if in.Template != nil {
- in, out := &in.Template, &out.Template
- *out = new(v1.PodTemplateSpec)
- if err := Convert_core_PodTemplateSpec_To_v1_PodTemplateSpec(*in, *out, s); err != nil {
- return err
- }
- } else {
- out.Template = nil
- }
- return nil
-}
-
-func autoConvert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus(in *v1.ReplicationControllerStatus, out *core.ReplicationControllerStatus, s conversion.Scope) error {
- out.Replicas = in.Replicas
- out.FullyLabeledReplicas = in.FullyLabeledReplicas
- out.ReadyReplicas = in.ReadyReplicas
- out.AvailableReplicas = in.AvailableReplicas
- out.ObservedGeneration = in.ObservedGeneration
- out.Conditions = *(*[]core.ReplicationControllerCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus is an autogenerated conversion function.
-func Convert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus(in *v1.ReplicationControllerStatus, out *core.ReplicationControllerStatus, s conversion.Scope) error {
- return autoConvert_v1_ReplicationControllerStatus_To_core_ReplicationControllerStatus(in, out, s)
-}
-
-func autoConvert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus(in *core.ReplicationControllerStatus, out *v1.ReplicationControllerStatus, s conversion.Scope) error {
- out.Replicas = in.Replicas
- out.FullyLabeledReplicas = in.FullyLabeledReplicas
- out.ReadyReplicas = in.ReadyReplicas
- out.AvailableReplicas = in.AvailableReplicas
- out.ObservedGeneration = in.ObservedGeneration
- out.Conditions = *(*[]v1.ReplicationControllerCondition)(unsafe.Pointer(&in.Conditions))
- return nil
-}
-
-// Convert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus is an autogenerated conversion function.
-func Convert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus(in *core.ReplicationControllerStatus, out *v1.ReplicationControllerStatus, s conversion.Scope) error {
- return autoConvert_core_ReplicationControllerStatus_To_v1_ReplicationControllerStatus(in, out, s)
-}
-
-func autoConvert_v1_ResourceFieldSelector_To_core_ResourceFieldSelector(in *v1.ResourceFieldSelector, out *core.ResourceFieldSelector, s conversion.Scope) error {
- out.ContainerName = in.ContainerName
- out.Resource = in.Resource
- out.Divisor = in.Divisor
- return nil
-}
-
-// Convert_v1_ResourceFieldSelector_To_core_ResourceFieldSelector is an autogenerated conversion function.
-func Convert_v1_ResourceFieldSelector_To_core_ResourceFieldSelector(in *v1.ResourceFieldSelector, out *core.ResourceFieldSelector, s conversion.Scope) error {
- return autoConvert_v1_ResourceFieldSelector_To_core_ResourceFieldSelector(in, out, s)
-}
-
-func autoConvert_core_ResourceFieldSelector_To_v1_ResourceFieldSelector(in *core.ResourceFieldSelector, out *v1.ResourceFieldSelector, s conversion.Scope) error {
- out.ContainerName = in.ContainerName
- out.Resource = in.Resource
- out.Divisor = in.Divisor
- return nil
-}
-
-// Convert_core_ResourceFieldSelector_To_v1_ResourceFieldSelector is an autogenerated conversion function.
-func Convert_core_ResourceFieldSelector_To_v1_ResourceFieldSelector(in *core.ResourceFieldSelector, out *v1.ResourceFieldSelector, s conversion.Scope) error {
- return autoConvert_core_ResourceFieldSelector_To_v1_ResourceFieldSelector(in, out, s)
-}
-
-func autoConvert_v1_ResourceQuota_To_core_ResourceQuota(in *v1.ResourceQuota, out *core.ResourceQuota, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_ResourceQuota_To_core_ResourceQuota is an autogenerated conversion function.
-func Convert_v1_ResourceQuota_To_core_ResourceQuota(in *v1.ResourceQuota, out *core.ResourceQuota, s conversion.Scope) error {
- return autoConvert_v1_ResourceQuota_To_core_ResourceQuota(in, out, s)
-}
-
-func autoConvert_core_ResourceQuota_To_v1_ResourceQuota(in *core.ResourceQuota, out *v1.ResourceQuota, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_ResourceQuota_To_v1_ResourceQuota is an autogenerated conversion function.
-func Convert_core_ResourceQuota_To_v1_ResourceQuota(in *core.ResourceQuota, out *v1.ResourceQuota, s conversion.Scope) error {
- return autoConvert_core_ResourceQuota_To_v1_ResourceQuota(in, out, s)
-}
-
-func autoConvert_v1_ResourceQuotaList_To_core_ResourceQuotaList(in *v1.ResourceQuotaList, out *core.ResourceQuotaList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.ResourceQuota)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_ResourceQuotaList_To_core_ResourceQuotaList is an autogenerated conversion function.
-func Convert_v1_ResourceQuotaList_To_core_ResourceQuotaList(in *v1.ResourceQuotaList, out *core.ResourceQuotaList, s conversion.Scope) error {
- return autoConvert_v1_ResourceQuotaList_To_core_ResourceQuotaList(in, out, s)
-}
-
-func autoConvert_core_ResourceQuotaList_To_v1_ResourceQuotaList(in *core.ResourceQuotaList, out *v1.ResourceQuotaList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.ResourceQuota)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_ResourceQuotaList_To_v1_ResourceQuotaList is an autogenerated conversion function.
-func Convert_core_ResourceQuotaList_To_v1_ResourceQuotaList(in *core.ResourceQuotaList, out *v1.ResourceQuotaList, s conversion.Scope) error {
- return autoConvert_core_ResourceQuotaList_To_v1_ResourceQuotaList(in, out, s)
-}
-
-func autoConvert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec(in *v1.ResourceQuotaSpec, out *core.ResourceQuotaSpec, s conversion.Scope) error {
- out.Hard = *(*core.ResourceList)(unsafe.Pointer(&in.Hard))
- out.Scopes = *(*[]core.ResourceQuotaScope)(unsafe.Pointer(&in.Scopes))
- return nil
-}
-
-// Convert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec is an autogenerated conversion function.
-func Convert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec(in *v1.ResourceQuotaSpec, out *core.ResourceQuotaSpec, s conversion.Scope) error {
- return autoConvert_v1_ResourceQuotaSpec_To_core_ResourceQuotaSpec(in, out, s)
-}
-
-func autoConvert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec(in *core.ResourceQuotaSpec, out *v1.ResourceQuotaSpec, s conversion.Scope) error {
- out.Hard = *(*v1.ResourceList)(unsafe.Pointer(&in.Hard))
- out.Scopes = *(*[]v1.ResourceQuotaScope)(unsafe.Pointer(&in.Scopes))
- return nil
-}
-
-// Convert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec is an autogenerated conversion function.
-func Convert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec(in *core.ResourceQuotaSpec, out *v1.ResourceQuotaSpec, s conversion.Scope) error {
- return autoConvert_core_ResourceQuotaSpec_To_v1_ResourceQuotaSpec(in, out, s)
-}
-
-func autoConvert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus(in *v1.ResourceQuotaStatus, out *core.ResourceQuotaStatus, s conversion.Scope) error {
- out.Hard = *(*core.ResourceList)(unsafe.Pointer(&in.Hard))
- out.Used = *(*core.ResourceList)(unsafe.Pointer(&in.Used))
- return nil
-}
-
-// Convert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus is an autogenerated conversion function.
-func Convert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus(in *v1.ResourceQuotaStatus, out *core.ResourceQuotaStatus, s conversion.Scope) error {
- return autoConvert_v1_ResourceQuotaStatus_To_core_ResourceQuotaStatus(in, out, s)
-}
-
-func autoConvert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus(in *core.ResourceQuotaStatus, out *v1.ResourceQuotaStatus, s conversion.Scope) error {
- out.Hard = *(*v1.ResourceList)(unsafe.Pointer(&in.Hard))
- out.Used = *(*v1.ResourceList)(unsafe.Pointer(&in.Used))
- return nil
-}
-
-// Convert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus is an autogenerated conversion function.
-func Convert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus(in *core.ResourceQuotaStatus, out *v1.ResourceQuotaStatus, s conversion.Scope) error {
- return autoConvert_core_ResourceQuotaStatus_To_v1_ResourceQuotaStatus(in, out, s)
-}
-
-func autoConvert_v1_ResourceRequirements_To_core_ResourceRequirements(in *v1.ResourceRequirements, out *core.ResourceRequirements, s conversion.Scope) error {
- out.Limits = *(*core.ResourceList)(unsafe.Pointer(&in.Limits))
- out.Requests = *(*core.ResourceList)(unsafe.Pointer(&in.Requests))
- return nil
-}
-
-// Convert_v1_ResourceRequirements_To_core_ResourceRequirements is an autogenerated conversion function.
-func Convert_v1_ResourceRequirements_To_core_ResourceRequirements(in *v1.ResourceRequirements, out *core.ResourceRequirements, s conversion.Scope) error {
- return autoConvert_v1_ResourceRequirements_To_core_ResourceRequirements(in, out, s)
-}
-
-func autoConvert_core_ResourceRequirements_To_v1_ResourceRequirements(in *core.ResourceRequirements, out *v1.ResourceRequirements, s conversion.Scope) error {
- out.Limits = *(*v1.ResourceList)(unsafe.Pointer(&in.Limits))
- out.Requests = *(*v1.ResourceList)(unsafe.Pointer(&in.Requests))
- return nil
-}
-
-// Convert_core_ResourceRequirements_To_v1_ResourceRequirements is an autogenerated conversion function.
-func Convert_core_ResourceRequirements_To_v1_ResourceRequirements(in *core.ResourceRequirements, out *v1.ResourceRequirements, s conversion.Scope) error {
- return autoConvert_core_ResourceRequirements_To_v1_ResourceRequirements(in, out, s)
-}
-
-func autoConvert_v1_SELinuxOptions_To_core_SELinuxOptions(in *v1.SELinuxOptions, out *core.SELinuxOptions, s conversion.Scope) error {
- out.User = in.User
- out.Role = in.Role
- out.Type = in.Type
- out.Level = in.Level
- return nil
-}
-
-// Convert_v1_SELinuxOptions_To_core_SELinuxOptions is an autogenerated conversion function.
-func Convert_v1_SELinuxOptions_To_core_SELinuxOptions(in *v1.SELinuxOptions, out *core.SELinuxOptions, s conversion.Scope) error {
- return autoConvert_v1_SELinuxOptions_To_core_SELinuxOptions(in, out, s)
-}
-
-func autoConvert_core_SELinuxOptions_To_v1_SELinuxOptions(in *core.SELinuxOptions, out *v1.SELinuxOptions, s conversion.Scope) error {
- out.User = in.User
- out.Role = in.Role
- out.Type = in.Type
- out.Level = in.Level
- return nil
-}
-
-// Convert_core_SELinuxOptions_To_v1_SELinuxOptions is an autogenerated conversion function.
-func Convert_core_SELinuxOptions_To_v1_SELinuxOptions(in *core.SELinuxOptions, out *v1.SELinuxOptions, s conversion.Scope) error {
- return autoConvert_core_SELinuxOptions_To_v1_SELinuxOptions(in, out, s)
-}
-
-func autoConvert_v1_ScaleIOPersistentVolumeSource_To_core_ScaleIOPersistentVolumeSource(in *v1.ScaleIOPersistentVolumeSource, out *core.ScaleIOPersistentVolumeSource, s conversion.Scope) error {
- out.Gateway = in.Gateway
- out.System = in.System
- out.SecretRef = (*core.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.SSLEnabled = in.SSLEnabled
- out.ProtectionDomain = in.ProtectionDomain
- out.StoragePool = in.StoragePool
- out.StorageMode = in.StorageMode
- out.VolumeName = in.VolumeName
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_ScaleIOPersistentVolumeSource_To_core_ScaleIOPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_ScaleIOPersistentVolumeSource_To_core_ScaleIOPersistentVolumeSource(in *v1.ScaleIOPersistentVolumeSource, out *core.ScaleIOPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ScaleIOPersistentVolumeSource_To_core_ScaleIOPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ScaleIOPersistentVolumeSource_To_v1_ScaleIOPersistentVolumeSource(in *core.ScaleIOPersistentVolumeSource, out *v1.ScaleIOPersistentVolumeSource, s conversion.Scope) error {
- out.Gateway = in.Gateway
- out.System = in.System
- out.SecretRef = (*v1.SecretReference)(unsafe.Pointer(in.SecretRef))
- out.SSLEnabled = in.SSLEnabled
- out.ProtectionDomain = in.ProtectionDomain
- out.StoragePool = in.StoragePool
- out.StorageMode = in.StorageMode
- out.VolumeName = in.VolumeName
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_ScaleIOPersistentVolumeSource_To_v1_ScaleIOPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_ScaleIOPersistentVolumeSource_To_v1_ScaleIOPersistentVolumeSource(in *core.ScaleIOPersistentVolumeSource, out *v1.ScaleIOPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ScaleIOPersistentVolumeSource_To_v1_ScaleIOPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_ScaleIOVolumeSource_To_core_ScaleIOVolumeSource(in *v1.ScaleIOVolumeSource, out *core.ScaleIOVolumeSource, s conversion.Scope) error {
- out.Gateway = in.Gateway
- out.System = in.System
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.SSLEnabled = in.SSLEnabled
- out.ProtectionDomain = in.ProtectionDomain
- out.StoragePool = in.StoragePool
- out.StorageMode = in.StorageMode
- out.VolumeName = in.VolumeName
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_v1_ScaleIOVolumeSource_To_core_ScaleIOVolumeSource is an autogenerated conversion function.
-func Convert_v1_ScaleIOVolumeSource_To_core_ScaleIOVolumeSource(in *v1.ScaleIOVolumeSource, out *core.ScaleIOVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_ScaleIOVolumeSource_To_core_ScaleIOVolumeSource(in, out, s)
-}
-
-func autoConvert_core_ScaleIOVolumeSource_To_v1_ScaleIOVolumeSource(in *core.ScaleIOVolumeSource, out *v1.ScaleIOVolumeSource, s conversion.Scope) error {
- out.Gateway = in.Gateway
- out.System = in.System
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- out.SSLEnabled = in.SSLEnabled
- out.ProtectionDomain = in.ProtectionDomain
- out.StoragePool = in.StoragePool
- out.StorageMode = in.StorageMode
- out.VolumeName = in.VolumeName
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- return nil
-}
-
-// Convert_core_ScaleIOVolumeSource_To_v1_ScaleIOVolumeSource is an autogenerated conversion function.
-func Convert_core_ScaleIOVolumeSource_To_v1_ScaleIOVolumeSource(in *core.ScaleIOVolumeSource, out *v1.ScaleIOVolumeSource, s conversion.Scope) error {
- return autoConvert_core_ScaleIOVolumeSource_To_v1_ScaleIOVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Secret_To_core_Secret(in *v1.Secret, out *core.Secret, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Data = *(*map[string][]byte)(unsafe.Pointer(&in.Data))
- // INFO: in.StringData opted out of conversion generation
- out.Type = core.SecretType(in.Type)
- return nil
-}
-
-func autoConvert_core_Secret_To_v1_Secret(in *core.Secret, out *v1.Secret, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Data = *(*map[string][]byte)(unsafe.Pointer(&in.Data))
- out.Type = v1.SecretType(in.Type)
- return nil
-}
-
-// Convert_core_Secret_To_v1_Secret is an autogenerated conversion function.
-func Convert_core_Secret_To_v1_Secret(in *core.Secret, out *v1.Secret, s conversion.Scope) error {
- return autoConvert_core_Secret_To_v1_Secret(in, out, s)
-}
-
-func autoConvert_v1_SecretEnvSource_To_core_SecretEnvSource(in *v1.SecretEnvSource, out *core.SecretEnvSource, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_SecretEnvSource_To_core_SecretEnvSource is an autogenerated conversion function.
-func Convert_v1_SecretEnvSource_To_core_SecretEnvSource(in *v1.SecretEnvSource, out *core.SecretEnvSource, s conversion.Scope) error {
- return autoConvert_v1_SecretEnvSource_To_core_SecretEnvSource(in, out, s)
-}
-
-func autoConvert_core_SecretEnvSource_To_v1_SecretEnvSource(in *core.SecretEnvSource, out *v1.SecretEnvSource, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_SecretEnvSource_To_v1_SecretEnvSource is an autogenerated conversion function.
-func Convert_core_SecretEnvSource_To_v1_SecretEnvSource(in *core.SecretEnvSource, out *v1.SecretEnvSource, s conversion.Scope) error {
- return autoConvert_core_SecretEnvSource_To_v1_SecretEnvSource(in, out, s)
-}
-
-func autoConvert_v1_SecretKeySelector_To_core_SecretKeySelector(in *v1.SecretKeySelector, out *core.SecretKeySelector, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Key = in.Key
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_SecretKeySelector_To_core_SecretKeySelector is an autogenerated conversion function.
-func Convert_v1_SecretKeySelector_To_core_SecretKeySelector(in *v1.SecretKeySelector, out *core.SecretKeySelector, s conversion.Scope) error {
- return autoConvert_v1_SecretKeySelector_To_core_SecretKeySelector(in, out, s)
-}
-
-func autoConvert_core_SecretKeySelector_To_v1_SecretKeySelector(in *core.SecretKeySelector, out *v1.SecretKeySelector, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Key = in.Key
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_SecretKeySelector_To_v1_SecretKeySelector is an autogenerated conversion function.
-func Convert_core_SecretKeySelector_To_v1_SecretKeySelector(in *core.SecretKeySelector, out *v1.SecretKeySelector, s conversion.Scope) error {
- return autoConvert_core_SecretKeySelector_To_v1_SecretKeySelector(in, out, s)
-}
-
-func autoConvert_v1_SecretList_To_core_SecretList(in *v1.SecretList, out *core.SecretList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.Secret, len(*in))
- for i := range *in {
- if err := Convert_v1_Secret_To_core_Secret(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_SecretList_To_core_SecretList is an autogenerated conversion function.
-func Convert_v1_SecretList_To_core_SecretList(in *v1.SecretList, out *core.SecretList, s conversion.Scope) error {
- return autoConvert_v1_SecretList_To_core_SecretList(in, out, s)
-}
-
-func autoConvert_core_SecretList_To_v1_SecretList(in *core.SecretList, out *v1.SecretList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.Secret, len(*in))
- for i := range *in {
- if err := Convert_core_Secret_To_v1_Secret(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_SecretList_To_v1_SecretList is an autogenerated conversion function.
-func Convert_core_SecretList_To_v1_SecretList(in *core.SecretList, out *v1.SecretList, s conversion.Scope) error {
- return autoConvert_core_SecretList_To_v1_SecretList(in, out, s)
-}
-
-func autoConvert_v1_SecretProjection_To_core_SecretProjection(in *v1.SecretProjection, out *core.SecretProjection, s conversion.Scope) error {
- if err := Convert_v1_LocalObjectReference_To_core_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]core.KeyToPath)(unsafe.Pointer(&in.Items))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_SecretProjection_To_core_SecretProjection is an autogenerated conversion function.
-func Convert_v1_SecretProjection_To_core_SecretProjection(in *v1.SecretProjection, out *core.SecretProjection, s conversion.Scope) error {
- return autoConvert_v1_SecretProjection_To_core_SecretProjection(in, out, s)
-}
-
-func autoConvert_core_SecretProjection_To_v1_SecretProjection(in *core.SecretProjection, out *v1.SecretProjection, s conversion.Scope) error {
- if err := Convert_core_LocalObjectReference_To_v1_LocalObjectReference(&in.LocalObjectReference, &out.LocalObjectReference, s); err != nil {
- return err
- }
- out.Items = *(*[]v1.KeyToPath)(unsafe.Pointer(&in.Items))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_SecretProjection_To_v1_SecretProjection is an autogenerated conversion function.
-func Convert_core_SecretProjection_To_v1_SecretProjection(in *core.SecretProjection, out *v1.SecretProjection, s conversion.Scope) error {
- return autoConvert_core_SecretProjection_To_v1_SecretProjection(in, out, s)
-}
-
-func autoConvert_v1_SecretReference_To_core_SecretReference(in *v1.SecretReference, out *core.SecretReference, s conversion.Scope) error {
- out.Name = in.Name
- out.Namespace = in.Namespace
- return nil
-}
-
-// Convert_v1_SecretReference_To_core_SecretReference is an autogenerated conversion function.
-func Convert_v1_SecretReference_To_core_SecretReference(in *v1.SecretReference, out *core.SecretReference, s conversion.Scope) error {
- return autoConvert_v1_SecretReference_To_core_SecretReference(in, out, s)
-}
-
-func autoConvert_core_SecretReference_To_v1_SecretReference(in *core.SecretReference, out *v1.SecretReference, s conversion.Scope) error {
- out.Name = in.Name
- out.Namespace = in.Namespace
- return nil
-}
-
-// Convert_core_SecretReference_To_v1_SecretReference is an autogenerated conversion function.
-func Convert_core_SecretReference_To_v1_SecretReference(in *core.SecretReference, out *v1.SecretReference, s conversion.Scope) error {
- return autoConvert_core_SecretReference_To_v1_SecretReference(in, out, s)
-}
-
-func autoConvert_v1_SecretVolumeSource_To_core_SecretVolumeSource(in *v1.SecretVolumeSource, out *core.SecretVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.Items = *(*[]core.KeyToPath)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_v1_SecretVolumeSource_To_core_SecretVolumeSource is an autogenerated conversion function.
-func Convert_v1_SecretVolumeSource_To_core_SecretVolumeSource(in *v1.SecretVolumeSource, out *core.SecretVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_SecretVolumeSource_To_core_SecretVolumeSource(in, out, s)
-}
-
-func autoConvert_core_SecretVolumeSource_To_v1_SecretVolumeSource(in *core.SecretVolumeSource, out *v1.SecretVolumeSource, s conversion.Scope) error {
- out.SecretName = in.SecretName
- out.Items = *(*[]v1.KeyToPath)(unsafe.Pointer(&in.Items))
- out.DefaultMode = (*int32)(unsafe.Pointer(in.DefaultMode))
- out.Optional = (*bool)(unsafe.Pointer(in.Optional))
- return nil
-}
-
-// Convert_core_SecretVolumeSource_To_v1_SecretVolumeSource is an autogenerated conversion function.
-func Convert_core_SecretVolumeSource_To_v1_SecretVolumeSource(in *core.SecretVolumeSource, out *v1.SecretVolumeSource, s conversion.Scope) error {
- return autoConvert_core_SecretVolumeSource_To_v1_SecretVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_SecurityContext_To_core_SecurityContext(in *v1.SecurityContext, out *core.SecurityContext, s conversion.Scope) error {
- out.Capabilities = (*core.Capabilities)(unsafe.Pointer(in.Capabilities))
- out.Privileged = (*bool)(unsafe.Pointer(in.Privileged))
- out.SELinuxOptions = (*core.SELinuxOptions)(unsafe.Pointer(in.SELinuxOptions))
- out.RunAsUser = (*int64)(unsafe.Pointer(in.RunAsUser))
- out.RunAsNonRoot = (*bool)(unsafe.Pointer(in.RunAsNonRoot))
- out.ReadOnlyRootFilesystem = (*bool)(unsafe.Pointer(in.ReadOnlyRootFilesystem))
- out.AllowPrivilegeEscalation = (*bool)(unsafe.Pointer(in.AllowPrivilegeEscalation))
- return nil
-}
-
-// Convert_v1_SecurityContext_To_core_SecurityContext is an autogenerated conversion function.
-func Convert_v1_SecurityContext_To_core_SecurityContext(in *v1.SecurityContext, out *core.SecurityContext, s conversion.Scope) error {
- return autoConvert_v1_SecurityContext_To_core_SecurityContext(in, out, s)
-}
-
-func autoConvert_core_SecurityContext_To_v1_SecurityContext(in *core.SecurityContext, out *v1.SecurityContext, s conversion.Scope) error {
- out.Capabilities = (*v1.Capabilities)(unsafe.Pointer(in.Capabilities))
- out.Privileged = (*bool)(unsafe.Pointer(in.Privileged))
- out.SELinuxOptions = (*v1.SELinuxOptions)(unsafe.Pointer(in.SELinuxOptions))
- out.RunAsUser = (*int64)(unsafe.Pointer(in.RunAsUser))
- out.RunAsNonRoot = (*bool)(unsafe.Pointer(in.RunAsNonRoot))
- out.ReadOnlyRootFilesystem = (*bool)(unsafe.Pointer(in.ReadOnlyRootFilesystem))
- out.AllowPrivilegeEscalation = (*bool)(unsafe.Pointer(in.AllowPrivilegeEscalation))
- return nil
-}
-
-func autoConvert_v1_SerializedReference_To_core_SerializedReference(in *v1.SerializedReference, out *core.SerializedReference, s conversion.Scope) error {
- if err := Convert_v1_ObjectReference_To_core_ObjectReference(&in.Reference, &out.Reference, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_SerializedReference_To_core_SerializedReference is an autogenerated conversion function.
-func Convert_v1_SerializedReference_To_core_SerializedReference(in *v1.SerializedReference, out *core.SerializedReference, s conversion.Scope) error {
- return autoConvert_v1_SerializedReference_To_core_SerializedReference(in, out, s)
-}
-
-func autoConvert_core_SerializedReference_To_v1_SerializedReference(in *core.SerializedReference, out *v1.SerializedReference, s conversion.Scope) error {
- if err := Convert_core_ObjectReference_To_v1_ObjectReference(&in.Reference, &out.Reference, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_SerializedReference_To_v1_SerializedReference is an autogenerated conversion function.
-func Convert_core_SerializedReference_To_v1_SerializedReference(in *core.SerializedReference, out *v1.SerializedReference, s conversion.Scope) error {
- return autoConvert_core_SerializedReference_To_v1_SerializedReference(in, out, s)
-}
-
-func autoConvert_v1_Service_To_core_Service(in *v1.Service, out *core.Service, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_v1_ServiceSpec_To_core_ServiceSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_v1_ServiceStatus_To_core_ServiceStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_Service_To_core_Service is an autogenerated conversion function.
-func Convert_v1_Service_To_core_Service(in *v1.Service, out *core.Service, s conversion.Scope) error {
- return autoConvert_v1_Service_To_core_Service(in, out, s)
-}
-
-func autoConvert_core_Service_To_v1_Service(in *core.Service, out *v1.Service, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- if err := Convert_core_ServiceSpec_To_v1_ServiceSpec(&in.Spec, &out.Spec, s); err != nil {
- return err
- }
- if err := Convert_core_ServiceStatus_To_v1_ServiceStatus(&in.Status, &out.Status, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_Service_To_v1_Service is an autogenerated conversion function.
-func Convert_core_Service_To_v1_Service(in *core.Service, out *v1.Service, s conversion.Scope) error {
- return autoConvert_core_Service_To_v1_Service(in, out, s)
-}
-
-func autoConvert_v1_ServiceAccount_To_core_ServiceAccount(in *v1.ServiceAccount, out *core.ServiceAccount, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Secrets = *(*[]core.ObjectReference)(unsafe.Pointer(&in.Secrets))
- out.ImagePullSecrets = *(*[]core.LocalObjectReference)(unsafe.Pointer(&in.ImagePullSecrets))
- out.AutomountServiceAccountToken = (*bool)(unsafe.Pointer(in.AutomountServiceAccountToken))
- return nil
-}
-
-// Convert_v1_ServiceAccount_To_core_ServiceAccount is an autogenerated conversion function.
-func Convert_v1_ServiceAccount_To_core_ServiceAccount(in *v1.ServiceAccount, out *core.ServiceAccount, s conversion.Scope) error {
- return autoConvert_v1_ServiceAccount_To_core_ServiceAccount(in, out, s)
-}
-
-func autoConvert_core_ServiceAccount_To_v1_ServiceAccount(in *core.ServiceAccount, out *v1.ServiceAccount, s conversion.Scope) error {
- out.ObjectMeta = in.ObjectMeta
- out.Secrets = *(*[]v1.ObjectReference)(unsafe.Pointer(&in.Secrets))
- out.ImagePullSecrets = *(*[]v1.LocalObjectReference)(unsafe.Pointer(&in.ImagePullSecrets))
- out.AutomountServiceAccountToken = (*bool)(unsafe.Pointer(in.AutomountServiceAccountToken))
- return nil
-}
-
-// Convert_core_ServiceAccount_To_v1_ServiceAccount is an autogenerated conversion function.
-func Convert_core_ServiceAccount_To_v1_ServiceAccount(in *core.ServiceAccount, out *v1.ServiceAccount, s conversion.Scope) error {
- return autoConvert_core_ServiceAccount_To_v1_ServiceAccount(in, out, s)
-}
-
-func autoConvert_v1_ServiceAccountList_To_core_ServiceAccountList(in *v1.ServiceAccountList, out *core.ServiceAccountList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]core.ServiceAccount)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_v1_ServiceAccountList_To_core_ServiceAccountList is an autogenerated conversion function.
-func Convert_v1_ServiceAccountList_To_core_ServiceAccountList(in *v1.ServiceAccountList, out *core.ServiceAccountList, s conversion.Scope) error {
- return autoConvert_v1_ServiceAccountList_To_core_ServiceAccountList(in, out, s)
-}
-
-func autoConvert_core_ServiceAccountList_To_v1_ServiceAccountList(in *core.ServiceAccountList, out *v1.ServiceAccountList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- out.Items = *(*[]v1.ServiceAccount)(unsafe.Pointer(&in.Items))
- return nil
-}
-
-// Convert_core_ServiceAccountList_To_v1_ServiceAccountList is an autogenerated conversion function.
-func Convert_core_ServiceAccountList_To_v1_ServiceAccountList(in *core.ServiceAccountList, out *v1.ServiceAccountList, s conversion.Scope) error {
- return autoConvert_core_ServiceAccountList_To_v1_ServiceAccountList(in, out, s)
-}
-
-func autoConvert_v1_ServiceList_To_core_ServiceList(in *v1.ServiceList, out *core.ServiceList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]core.Service, len(*in))
- for i := range *in {
- if err := Convert_v1_Service_To_core_Service(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_v1_ServiceList_To_core_ServiceList is an autogenerated conversion function.
-func Convert_v1_ServiceList_To_core_ServiceList(in *v1.ServiceList, out *core.ServiceList, s conversion.Scope) error {
- return autoConvert_v1_ServiceList_To_core_ServiceList(in, out, s)
-}
-
-func autoConvert_core_ServiceList_To_v1_ServiceList(in *core.ServiceList, out *v1.ServiceList, s conversion.Scope) error {
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]v1.Service, len(*in))
- for i := range *in {
- if err := Convert_core_Service_To_v1_Service(&(*in)[i], &(*out)[i], s); err != nil {
- return err
- }
- }
- } else {
- out.Items = nil
- }
- return nil
-}
-
-// Convert_core_ServiceList_To_v1_ServiceList is an autogenerated conversion function.
-func Convert_core_ServiceList_To_v1_ServiceList(in *core.ServiceList, out *v1.ServiceList, s conversion.Scope) error {
- return autoConvert_core_ServiceList_To_v1_ServiceList(in, out, s)
-}
-
-func autoConvert_v1_ServicePort_To_core_ServicePort(in *v1.ServicePort, out *core.ServicePort, s conversion.Scope) error {
- out.Name = in.Name
- out.Protocol = core.Protocol(in.Protocol)
- out.Port = in.Port
- out.TargetPort = in.TargetPort
- out.NodePort = in.NodePort
- return nil
-}
-
-// Convert_v1_ServicePort_To_core_ServicePort is an autogenerated conversion function.
-func Convert_v1_ServicePort_To_core_ServicePort(in *v1.ServicePort, out *core.ServicePort, s conversion.Scope) error {
- return autoConvert_v1_ServicePort_To_core_ServicePort(in, out, s)
-}
-
-func autoConvert_core_ServicePort_To_v1_ServicePort(in *core.ServicePort, out *v1.ServicePort, s conversion.Scope) error {
- out.Name = in.Name
- out.Protocol = v1.Protocol(in.Protocol)
- out.Port = in.Port
- out.TargetPort = in.TargetPort
- out.NodePort = in.NodePort
- return nil
-}
-
-// Convert_core_ServicePort_To_v1_ServicePort is an autogenerated conversion function.
-func Convert_core_ServicePort_To_v1_ServicePort(in *core.ServicePort, out *v1.ServicePort, s conversion.Scope) error {
- return autoConvert_core_ServicePort_To_v1_ServicePort(in, out, s)
-}
-
-func autoConvert_v1_ServiceProxyOptions_To_core_ServiceProxyOptions(in *v1.ServiceProxyOptions, out *core.ServiceProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_v1_ServiceProxyOptions_To_core_ServiceProxyOptions is an autogenerated conversion function.
-func Convert_v1_ServiceProxyOptions_To_core_ServiceProxyOptions(in *v1.ServiceProxyOptions, out *core.ServiceProxyOptions, s conversion.Scope) error {
- return autoConvert_v1_ServiceProxyOptions_To_core_ServiceProxyOptions(in, out, s)
-}
-
-func autoConvert_core_ServiceProxyOptions_To_v1_ServiceProxyOptions(in *core.ServiceProxyOptions, out *v1.ServiceProxyOptions, s conversion.Scope) error {
- out.Path = in.Path
- return nil
-}
-
-// Convert_core_ServiceProxyOptions_To_v1_ServiceProxyOptions is an autogenerated conversion function.
-func Convert_core_ServiceProxyOptions_To_v1_ServiceProxyOptions(in *core.ServiceProxyOptions, out *v1.ServiceProxyOptions, s conversion.Scope) error {
- return autoConvert_core_ServiceProxyOptions_To_v1_ServiceProxyOptions(in, out, s)
-}
-
-func autoConvert_v1_ServiceSpec_To_core_ServiceSpec(in *v1.ServiceSpec, out *core.ServiceSpec, s conversion.Scope) error {
- out.Ports = *(*[]core.ServicePort)(unsafe.Pointer(&in.Ports))
- out.Selector = *(*map[string]string)(unsafe.Pointer(&in.Selector))
- out.ClusterIP = in.ClusterIP
- out.Type = core.ServiceType(in.Type)
- out.ExternalIPs = *(*[]string)(unsafe.Pointer(&in.ExternalIPs))
- out.SessionAffinity = core.ServiceAffinity(in.SessionAffinity)
- out.LoadBalancerIP = in.LoadBalancerIP
- out.LoadBalancerSourceRanges = *(*[]string)(unsafe.Pointer(&in.LoadBalancerSourceRanges))
- out.ExternalName = in.ExternalName
- out.ExternalTrafficPolicy = core.ServiceExternalTrafficPolicyType(in.ExternalTrafficPolicy)
- out.HealthCheckNodePort = in.HealthCheckNodePort
- out.PublishNotReadyAddresses = in.PublishNotReadyAddresses
- out.SessionAffinityConfig = (*core.SessionAffinityConfig)(unsafe.Pointer(in.SessionAffinityConfig))
- return nil
-}
-
-// Convert_v1_ServiceSpec_To_core_ServiceSpec is an autogenerated conversion function.
-func Convert_v1_ServiceSpec_To_core_ServiceSpec(in *v1.ServiceSpec, out *core.ServiceSpec, s conversion.Scope) error {
- return autoConvert_v1_ServiceSpec_To_core_ServiceSpec(in, out, s)
-}
-
-func autoConvert_core_ServiceSpec_To_v1_ServiceSpec(in *core.ServiceSpec, out *v1.ServiceSpec, s conversion.Scope) error {
- out.Type = v1.ServiceType(in.Type)
- out.Ports = *(*[]v1.ServicePort)(unsafe.Pointer(&in.Ports))
- out.Selector = *(*map[string]string)(unsafe.Pointer(&in.Selector))
- out.ClusterIP = in.ClusterIP
- out.ExternalName = in.ExternalName
- out.ExternalIPs = *(*[]string)(unsafe.Pointer(&in.ExternalIPs))
- out.LoadBalancerIP = in.LoadBalancerIP
- out.SessionAffinity = v1.ServiceAffinity(in.SessionAffinity)
- out.SessionAffinityConfig = (*v1.SessionAffinityConfig)(unsafe.Pointer(in.SessionAffinityConfig))
- out.LoadBalancerSourceRanges = *(*[]string)(unsafe.Pointer(&in.LoadBalancerSourceRanges))
- out.ExternalTrafficPolicy = v1.ServiceExternalTrafficPolicyType(in.ExternalTrafficPolicy)
- out.HealthCheckNodePort = in.HealthCheckNodePort
- out.PublishNotReadyAddresses = in.PublishNotReadyAddresses
- return nil
-}
-
-// Convert_core_ServiceSpec_To_v1_ServiceSpec is an autogenerated conversion function.
-func Convert_core_ServiceSpec_To_v1_ServiceSpec(in *core.ServiceSpec, out *v1.ServiceSpec, s conversion.Scope) error {
- return autoConvert_core_ServiceSpec_To_v1_ServiceSpec(in, out, s)
-}
-
-func autoConvert_v1_ServiceStatus_To_core_ServiceStatus(in *v1.ServiceStatus, out *core.ServiceStatus, s conversion.Scope) error {
- if err := Convert_v1_LoadBalancerStatus_To_core_LoadBalancerStatus(&in.LoadBalancer, &out.LoadBalancer, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_ServiceStatus_To_core_ServiceStatus is an autogenerated conversion function.
-func Convert_v1_ServiceStatus_To_core_ServiceStatus(in *v1.ServiceStatus, out *core.ServiceStatus, s conversion.Scope) error {
- return autoConvert_v1_ServiceStatus_To_core_ServiceStatus(in, out, s)
-}
-
-func autoConvert_core_ServiceStatus_To_v1_ServiceStatus(in *core.ServiceStatus, out *v1.ServiceStatus, s conversion.Scope) error {
- if err := Convert_core_LoadBalancerStatus_To_v1_LoadBalancerStatus(&in.LoadBalancer, &out.LoadBalancer, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_ServiceStatus_To_v1_ServiceStatus is an autogenerated conversion function.
-func Convert_core_ServiceStatus_To_v1_ServiceStatus(in *core.ServiceStatus, out *v1.ServiceStatus, s conversion.Scope) error {
- return autoConvert_core_ServiceStatus_To_v1_ServiceStatus(in, out, s)
-}
-
-func autoConvert_v1_SessionAffinityConfig_To_core_SessionAffinityConfig(in *v1.SessionAffinityConfig, out *core.SessionAffinityConfig, s conversion.Scope) error {
- out.ClientIP = (*core.ClientIPConfig)(unsafe.Pointer(in.ClientIP))
- return nil
-}
-
-// Convert_v1_SessionAffinityConfig_To_core_SessionAffinityConfig is an autogenerated conversion function.
-func Convert_v1_SessionAffinityConfig_To_core_SessionAffinityConfig(in *v1.SessionAffinityConfig, out *core.SessionAffinityConfig, s conversion.Scope) error {
- return autoConvert_v1_SessionAffinityConfig_To_core_SessionAffinityConfig(in, out, s)
-}
-
-func autoConvert_core_SessionAffinityConfig_To_v1_SessionAffinityConfig(in *core.SessionAffinityConfig, out *v1.SessionAffinityConfig, s conversion.Scope) error {
- out.ClientIP = (*v1.ClientIPConfig)(unsafe.Pointer(in.ClientIP))
- return nil
-}
-
-// Convert_core_SessionAffinityConfig_To_v1_SessionAffinityConfig is an autogenerated conversion function.
-func Convert_core_SessionAffinityConfig_To_v1_SessionAffinityConfig(in *core.SessionAffinityConfig, out *v1.SessionAffinityConfig, s conversion.Scope) error {
- return autoConvert_core_SessionAffinityConfig_To_v1_SessionAffinityConfig(in, out, s)
-}
-
-func autoConvert_v1_StorageOSPersistentVolumeSource_To_core_StorageOSPersistentVolumeSource(in *v1.StorageOSPersistentVolumeSource, out *core.StorageOSPersistentVolumeSource, s conversion.Scope) error {
- out.VolumeName = in.VolumeName
- out.VolumeNamespace = in.VolumeNamespace
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.SecretRef = (*core.ObjectReference)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_v1_StorageOSPersistentVolumeSource_To_core_StorageOSPersistentVolumeSource is an autogenerated conversion function.
-func Convert_v1_StorageOSPersistentVolumeSource_To_core_StorageOSPersistentVolumeSource(in *v1.StorageOSPersistentVolumeSource, out *core.StorageOSPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_StorageOSPersistentVolumeSource_To_core_StorageOSPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_core_StorageOSPersistentVolumeSource_To_v1_StorageOSPersistentVolumeSource(in *core.StorageOSPersistentVolumeSource, out *v1.StorageOSPersistentVolumeSource, s conversion.Scope) error {
- out.VolumeName = in.VolumeName
- out.VolumeNamespace = in.VolumeNamespace
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.SecretRef = (*v1.ObjectReference)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_core_StorageOSPersistentVolumeSource_To_v1_StorageOSPersistentVolumeSource is an autogenerated conversion function.
-func Convert_core_StorageOSPersistentVolumeSource_To_v1_StorageOSPersistentVolumeSource(in *core.StorageOSPersistentVolumeSource, out *v1.StorageOSPersistentVolumeSource, s conversion.Scope) error {
- return autoConvert_core_StorageOSPersistentVolumeSource_To_v1_StorageOSPersistentVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_StorageOSVolumeSource_To_core_StorageOSVolumeSource(in *v1.StorageOSVolumeSource, out *core.StorageOSVolumeSource, s conversion.Scope) error {
- out.VolumeName = in.VolumeName
- out.VolumeNamespace = in.VolumeNamespace
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.SecretRef = (*core.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_v1_StorageOSVolumeSource_To_core_StorageOSVolumeSource is an autogenerated conversion function.
-func Convert_v1_StorageOSVolumeSource_To_core_StorageOSVolumeSource(in *v1.StorageOSVolumeSource, out *core.StorageOSVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_StorageOSVolumeSource_To_core_StorageOSVolumeSource(in, out, s)
-}
-
-func autoConvert_core_StorageOSVolumeSource_To_v1_StorageOSVolumeSource(in *core.StorageOSVolumeSource, out *v1.StorageOSVolumeSource, s conversion.Scope) error {
- out.VolumeName = in.VolumeName
- out.VolumeNamespace = in.VolumeNamespace
- out.FSType = in.FSType
- out.ReadOnly = in.ReadOnly
- out.SecretRef = (*v1.LocalObjectReference)(unsafe.Pointer(in.SecretRef))
- return nil
-}
-
-// Convert_core_StorageOSVolumeSource_To_v1_StorageOSVolumeSource is an autogenerated conversion function.
-func Convert_core_StorageOSVolumeSource_To_v1_StorageOSVolumeSource(in *core.StorageOSVolumeSource, out *v1.StorageOSVolumeSource, s conversion.Scope) error {
- return autoConvert_core_StorageOSVolumeSource_To_v1_StorageOSVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_Sysctl_To_core_Sysctl(in *v1.Sysctl, out *core.Sysctl, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- return nil
-}
-
-// Convert_v1_Sysctl_To_core_Sysctl is an autogenerated conversion function.
-func Convert_v1_Sysctl_To_core_Sysctl(in *v1.Sysctl, out *core.Sysctl, s conversion.Scope) error {
- return autoConvert_v1_Sysctl_To_core_Sysctl(in, out, s)
-}
-
-func autoConvert_core_Sysctl_To_v1_Sysctl(in *core.Sysctl, out *v1.Sysctl, s conversion.Scope) error {
- out.Name = in.Name
- out.Value = in.Value
- return nil
-}
-
-// Convert_core_Sysctl_To_v1_Sysctl is an autogenerated conversion function.
-func Convert_core_Sysctl_To_v1_Sysctl(in *core.Sysctl, out *v1.Sysctl, s conversion.Scope) error {
- return autoConvert_core_Sysctl_To_v1_Sysctl(in, out, s)
-}
-
-func autoConvert_v1_TCPSocketAction_To_core_TCPSocketAction(in *v1.TCPSocketAction, out *core.TCPSocketAction, s conversion.Scope) error {
- out.Port = in.Port
- out.Host = in.Host
- return nil
-}
-
-// Convert_v1_TCPSocketAction_To_core_TCPSocketAction is an autogenerated conversion function.
-func Convert_v1_TCPSocketAction_To_core_TCPSocketAction(in *v1.TCPSocketAction, out *core.TCPSocketAction, s conversion.Scope) error {
- return autoConvert_v1_TCPSocketAction_To_core_TCPSocketAction(in, out, s)
-}
-
-func autoConvert_core_TCPSocketAction_To_v1_TCPSocketAction(in *core.TCPSocketAction, out *v1.TCPSocketAction, s conversion.Scope) error {
- out.Port = in.Port
- out.Host = in.Host
- return nil
-}
-
-// Convert_core_TCPSocketAction_To_v1_TCPSocketAction is an autogenerated conversion function.
-func Convert_core_TCPSocketAction_To_v1_TCPSocketAction(in *core.TCPSocketAction, out *v1.TCPSocketAction, s conversion.Scope) error {
- return autoConvert_core_TCPSocketAction_To_v1_TCPSocketAction(in, out, s)
-}
-
-func autoConvert_v1_Taint_To_core_Taint(in *v1.Taint, out *core.Taint, s conversion.Scope) error {
- out.Key = in.Key
- out.Value = in.Value
- out.Effect = core.TaintEffect(in.Effect)
- out.TimeAdded = (*meta_v1.Time)(unsafe.Pointer(in.TimeAdded))
- return nil
-}
-
-// Convert_v1_Taint_To_core_Taint is an autogenerated conversion function.
-func Convert_v1_Taint_To_core_Taint(in *v1.Taint, out *core.Taint, s conversion.Scope) error {
- return autoConvert_v1_Taint_To_core_Taint(in, out, s)
-}
-
-func autoConvert_core_Taint_To_v1_Taint(in *core.Taint, out *v1.Taint, s conversion.Scope) error {
- out.Key = in.Key
- out.Value = in.Value
- out.Effect = v1.TaintEffect(in.Effect)
- out.TimeAdded = (*meta_v1.Time)(unsafe.Pointer(in.TimeAdded))
- return nil
-}
-
-// Convert_core_Taint_To_v1_Taint is an autogenerated conversion function.
-func Convert_core_Taint_To_v1_Taint(in *core.Taint, out *v1.Taint, s conversion.Scope) error {
- return autoConvert_core_Taint_To_v1_Taint(in, out, s)
-}
-
-func autoConvert_v1_Toleration_To_core_Toleration(in *v1.Toleration, out *core.Toleration, s conversion.Scope) error {
- out.Key = in.Key
- out.Operator = core.TolerationOperator(in.Operator)
- out.Value = in.Value
- out.Effect = core.TaintEffect(in.Effect)
- out.TolerationSeconds = (*int64)(unsafe.Pointer(in.TolerationSeconds))
- return nil
-}
-
-// Convert_v1_Toleration_To_core_Toleration is an autogenerated conversion function.
-func Convert_v1_Toleration_To_core_Toleration(in *v1.Toleration, out *core.Toleration, s conversion.Scope) error {
- return autoConvert_v1_Toleration_To_core_Toleration(in, out, s)
-}
-
-func autoConvert_core_Toleration_To_v1_Toleration(in *core.Toleration, out *v1.Toleration, s conversion.Scope) error {
- out.Key = in.Key
- out.Operator = v1.TolerationOperator(in.Operator)
- out.Value = in.Value
- out.Effect = v1.TaintEffect(in.Effect)
- out.TolerationSeconds = (*int64)(unsafe.Pointer(in.TolerationSeconds))
- return nil
-}
-
-// Convert_core_Toleration_To_v1_Toleration is an autogenerated conversion function.
-func Convert_core_Toleration_To_v1_Toleration(in *core.Toleration, out *v1.Toleration, s conversion.Scope) error {
- return autoConvert_core_Toleration_To_v1_Toleration(in, out, s)
-}
-
-func autoConvert_v1_Volume_To_core_Volume(in *v1.Volume, out *core.Volume, s conversion.Scope) error {
- out.Name = in.Name
- if err := Convert_v1_VolumeSource_To_core_VolumeSource(&in.VolumeSource, &out.VolumeSource, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_Volume_To_core_Volume is an autogenerated conversion function.
-func Convert_v1_Volume_To_core_Volume(in *v1.Volume, out *core.Volume, s conversion.Scope) error {
- return autoConvert_v1_Volume_To_core_Volume(in, out, s)
-}
-
-func autoConvert_core_Volume_To_v1_Volume(in *core.Volume, out *v1.Volume, s conversion.Scope) error {
- out.Name = in.Name
- if err := Convert_core_VolumeSource_To_v1_VolumeSource(&in.VolumeSource, &out.VolumeSource, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_Volume_To_v1_Volume is an autogenerated conversion function.
-func Convert_core_Volume_To_v1_Volume(in *core.Volume, out *v1.Volume, s conversion.Scope) error {
- return autoConvert_core_Volume_To_v1_Volume(in, out, s)
-}
-
-func autoConvert_v1_VolumeDevice_To_core_VolumeDevice(in *v1.VolumeDevice, out *core.VolumeDevice, s conversion.Scope) error {
- out.Name = in.Name
- out.DevicePath = in.DevicePath
- return nil
-}
-
-// Convert_v1_VolumeDevice_To_core_VolumeDevice is an autogenerated conversion function.
-func Convert_v1_VolumeDevice_To_core_VolumeDevice(in *v1.VolumeDevice, out *core.VolumeDevice, s conversion.Scope) error {
- return autoConvert_v1_VolumeDevice_To_core_VolumeDevice(in, out, s)
-}
-
-func autoConvert_core_VolumeDevice_To_v1_VolumeDevice(in *core.VolumeDevice, out *v1.VolumeDevice, s conversion.Scope) error {
- out.Name = in.Name
- out.DevicePath = in.DevicePath
- return nil
-}
-
-// Convert_core_VolumeDevice_To_v1_VolumeDevice is an autogenerated conversion function.
-func Convert_core_VolumeDevice_To_v1_VolumeDevice(in *core.VolumeDevice, out *v1.VolumeDevice, s conversion.Scope) error {
- return autoConvert_core_VolumeDevice_To_v1_VolumeDevice(in, out, s)
-}
-
-func autoConvert_v1_VolumeMount_To_core_VolumeMount(in *v1.VolumeMount, out *core.VolumeMount, s conversion.Scope) error {
- out.Name = in.Name
- out.ReadOnly = in.ReadOnly
- out.MountPath = in.MountPath
- out.SubPath = in.SubPath
- out.MountPropagation = (*core.MountPropagationMode)(unsafe.Pointer(in.MountPropagation))
- return nil
-}
-
-// Convert_v1_VolumeMount_To_core_VolumeMount is an autogenerated conversion function.
-func Convert_v1_VolumeMount_To_core_VolumeMount(in *v1.VolumeMount, out *core.VolumeMount, s conversion.Scope) error {
- return autoConvert_v1_VolumeMount_To_core_VolumeMount(in, out, s)
-}
-
-func autoConvert_core_VolumeMount_To_v1_VolumeMount(in *core.VolumeMount, out *v1.VolumeMount, s conversion.Scope) error {
- out.Name = in.Name
- out.ReadOnly = in.ReadOnly
- out.MountPath = in.MountPath
- out.SubPath = in.SubPath
- out.MountPropagation = (*v1.MountPropagationMode)(unsafe.Pointer(in.MountPropagation))
- return nil
-}
-
-// Convert_core_VolumeMount_To_v1_VolumeMount is an autogenerated conversion function.
-func Convert_core_VolumeMount_To_v1_VolumeMount(in *core.VolumeMount, out *v1.VolumeMount, s conversion.Scope) error {
- return autoConvert_core_VolumeMount_To_v1_VolumeMount(in, out, s)
-}
-
-func autoConvert_v1_VolumeNodeAffinity_To_core_VolumeNodeAffinity(in *v1.VolumeNodeAffinity, out *core.VolumeNodeAffinity, s conversion.Scope) error {
- out.Required = (*core.NodeSelector)(unsafe.Pointer(in.Required))
- return nil
-}
-
-// Convert_v1_VolumeNodeAffinity_To_core_VolumeNodeAffinity is an autogenerated conversion function.
-func Convert_v1_VolumeNodeAffinity_To_core_VolumeNodeAffinity(in *v1.VolumeNodeAffinity, out *core.VolumeNodeAffinity, s conversion.Scope) error {
- return autoConvert_v1_VolumeNodeAffinity_To_core_VolumeNodeAffinity(in, out, s)
-}
-
-func autoConvert_core_VolumeNodeAffinity_To_v1_VolumeNodeAffinity(in *core.VolumeNodeAffinity, out *v1.VolumeNodeAffinity, s conversion.Scope) error {
- out.Required = (*v1.NodeSelector)(unsafe.Pointer(in.Required))
- return nil
-}
-
-// Convert_core_VolumeNodeAffinity_To_v1_VolumeNodeAffinity is an autogenerated conversion function.
-func Convert_core_VolumeNodeAffinity_To_v1_VolumeNodeAffinity(in *core.VolumeNodeAffinity, out *v1.VolumeNodeAffinity, s conversion.Scope) error {
- return autoConvert_core_VolumeNodeAffinity_To_v1_VolumeNodeAffinity(in, out, s)
-}
-
-func autoConvert_v1_VolumeProjection_To_core_VolumeProjection(in *v1.VolumeProjection, out *core.VolumeProjection, s conversion.Scope) error {
- out.Secret = (*core.SecretProjection)(unsafe.Pointer(in.Secret))
- out.DownwardAPI = (*core.DownwardAPIProjection)(unsafe.Pointer(in.DownwardAPI))
- out.ConfigMap = (*core.ConfigMapProjection)(unsafe.Pointer(in.ConfigMap))
- return nil
-}
-
-// Convert_v1_VolumeProjection_To_core_VolumeProjection is an autogenerated conversion function.
-func Convert_v1_VolumeProjection_To_core_VolumeProjection(in *v1.VolumeProjection, out *core.VolumeProjection, s conversion.Scope) error {
- return autoConvert_v1_VolumeProjection_To_core_VolumeProjection(in, out, s)
-}
-
-func autoConvert_core_VolumeProjection_To_v1_VolumeProjection(in *core.VolumeProjection, out *v1.VolumeProjection, s conversion.Scope) error {
- out.Secret = (*v1.SecretProjection)(unsafe.Pointer(in.Secret))
- out.DownwardAPI = (*v1.DownwardAPIProjection)(unsafe.Pointer(in.DownwardAPI))
- out.ConfigMap = (*v1.ConfigMapProjection)(unsafe.Pointer(in.ConfigMap))
- return nil
-}
-
-// Convert_core_VolumeProjection_To_v1_VolumeProjection is an autogenerated conversion function.
-func Convert_core_VolumeProjection_To_v1_VolumeProjection(in *core.VolumeProjection, out *v1.VolumeProjection, s conversion.Scope) error {
- return autoConvert_core_VolumeProjection_To_v1_VolumeProjection(in, out, s)
-}
-
-func autoConvert_v1_VolumeSource_To_core_VolumeSource(in *v1.VolumeSource, out *core.VolumeSource, s conversion.Scope) error {
- out.HostPath = (*core.HostPathVolumeSource)(unsafe.Pointer(in.HostPath))
- out.EmptyDir = (*core.EmptyDirVolumeSource)(unsafe.Pointer(in.EmptyDir))
- out.GCEPersistentDisk = (*core.GCEPersistentDiskVolumeSource)(unsafe.Pointer(in.GCEPersistentDisk))
- out.AWSElasticBlockStore = (*core.AWSElasticBlockStoreVolumeSource)(unsafe.Pointer(in.AWSElasticBlockStore))
- out.GitRepo = (*core.GitRepoVolumeSource)(unsafe.Pointer(in.GitRepo))
- out.Secret = (*core.SecretVolumeSource)(unsafe.Pointer(in.Secret))
- out.NFS = (*core.NFSVolumeSource)(unsafe.Pointer(in.NFS))
- out.ISCSI = (*core.ISCSIVolumeSource)(unsafe.Pointer(in.ISCSI))
- out.Glusterfs = (*core.GlusterfsVolumeSource)(unsafe.Pointer(in.Glusterfs))
- out.PersistentVolumeClaim = (*core.PersistentVolumeClaimVolumeSource)(unsafe.Pointer(in.PersistentVolumeClaim))
- out.RBD = (*core.RBDVolumeSource)(unsafe.Pointer(in.RBD))
- out.FlexVolume = (*core.FlexVolumeSource)(unsafe.Pointer(in.FlexVolume))
- out.Cinder = (*core.CinderVolumeSource)(unsafe.Pointer(in.Cinder))
- out.CephFS = (*core.CephFSVolumeSource)(unsafe.Pointer(in.CephFS))
- out.Flocker = (*core.FlockerVolumeSource)(unsafe.Pointer(in.Flocker))
- out.DownwardAPI = (*core.DownwardAPIVolumeSource)(unsafe.Pointer(in.DownwardAPI))
- out.FC = (*core.FCVolumeSource)(unsafe.Pointer(in.FC))
- out.AzureFile = (*core.AzureFileVolumeSource)(unsafe.Pointer(in.AzureFile))
- out.ConfigMap = (*core.ConfigMapVolumeSource)(unsafe.Pointer(in.ConfigMap))
- out.VsphereVolume = (*core.VsphereVirtualDiskVolumeSource)(unsafe.Pointer(in.VsphereVolume))
- out.Quobyte = (*core.QuobyteVolumeSource)(unsafe.Pointer(in.Quobyte))
- out.AzureDisk = (*core.AzureDiskVolumeSource)(unsafe.Pointer(in.AzureDisk))
- out.PhotonPersistentDisk = (*core.PhotonPersistentDiskVolumeSource)(unsafe.Pointer(in.PhotonPersistentDisk))
- out.Projected = (*core.ProjectedVolumeSource)(unsafe.Pointer(in.Projected))
- out.PortworxVolume = (*core.PortworxVolumeSource)(unsafe.Pointer(in.PortworxVolume))
- out.ScaleIO = (*core.ScaleIOVolumeSource)(unsafe.Pointer(in.ScaleIO))
- out.StorageOS = (*core.StorageOSVolumeSource)(unsafe.Pointer(in.StorageOS))
- return nil
-}
-
-// Convert_v1_VolumeSource_To_core_VolumeSource is an autogenerated conversion function.
-func Convert_v1_VolumeSource_To_core_VolumeSource(in *v1.VolumeSource, out *core.VolumeSource, s conversion.Scope) error {
- return autoConvert_v1_VolumeSource_To_core_VolumeSource(in, out, s)
-}
-
-func autoConvert_core_VolumeSource_To_v1_VolumeSource(in *core.VolumeSource, out *v1.VolumeSource, s conversion.Scope) error {
- out.HostPath = (*v1.HostPathVolumeSource)(unsafe.Pointer(in.HostPath))
- out.EmptyDir = (*v1.EmptyDirVolumeSource)(unsafe.Pointer(in.EmptyDir))
- out.GCEPersistentDisk = (*v1.GCEPersistentDiskVolumeSource)(unsafe.Pointer(in.GCEPersistentDisk))
- out.AWSElasticBlockStore = (*v1.AWSElasticBlockStoreVolumeSource)(unsafe.Pointer(in.AWSElasticBlockStore))
- out.GitRepo = (*v1.GitRepoVolumeSource)(unsafe.Pointer(in.GitRepo))
- out.Secret = (*v1.SecretVolumeSource)(unsafe.Pointer(in.Secret))
- out.NFS = (*v1.NFSVolumeSource)(unsafe.Pointer(in.NFS))
- out.ISCSI = (*v1.ISCSIVolumeSource)(unsafe.Pointer(in.ISCSI))
- out.Glusterfs = (*v1.GlusterfsVolumeSource)(unsafe.Pointer(in.Glusterfs))
- out.PersistentVolumeClaim = (*v1.PersistentVolumeClaimVolumeSource)(unsafe.Pointer(in.PersistentVolumeClaim))
- out.RBD = (*v1.RBDVolumeSource)(unsafe.Pointer(in.RBD))
- out.Quobyte = (*v1.QuobyteVolumeSource)(unsafe.Pointer(in.Quobyte))
- out.FlexVolume = (*v1.FlexVolumeSource)(unsafe.Pointer(in.FlexVolume))
- out.Cinder = (*v1.CinderVolumeSource)(unsafe.Pointer(in.Cinder))
- out.CephFS = (*v1.CephFSVolumeSource)(unsafe.Pointer(in.CephFS))
- out.Flocker = (*v1.FlockerVolumeSource)(unsafe.Pointer(in.Flocker))
- out.DownwardAPI = (*v1.DownwardAPIVolumeSource)(unsafe.Pointer(in.DownwardAPI))
- out.FC = (*v1.FCVolumeSource)(unsafe.Pointer(in.FC))
- out.AzureFile = (*v1.AzureFileVolumeSource)(unsafe.Pointer(in.AzureFile))
- out.ConfigMap = (*v1.ConfigMapVolumeSource)(unsafe.Pointer(in.ConfigMap))
- out.VsphereVolume = (*v1.VsphereVirtualDiskVolumeSource)(unsafe.Pointer(in.VsphereVolume))
- out.AzureDisk = (*v1.AzureDiskVolumeSource)(unsafe.Pointer(in.AzureDisk))
- out.PhotonPersistentDisk = (*v1.PhotonPersistentDiskVolumeSource)(unsafe.Pointer(in.PhotonPersistentDisk))
- out.Projected = (*v1.ProjectedVolumeSource)(unsafe.Pointer(in.Projected))
- out.PortworxVolume = (*v1.PortworxVolumeSource)(unsafe.Pointer(in.PortworxVolume))
- out.ScaleIO = (*v1.ScaleIOVolumeSource)(unsafe.Pointer(in.ScaleIO))
- out.StorageOS = (*v1.StorageOSVolumeSource)(unsafe.Pointer(in.StorageOS))
- return nil
-}
-
-// Convert_core_VolumeSource_To_v1_VolumeSource is an autogenerated conversion function.
-func Convert_core_VolumeSource_To_v1_VolumeSource(in *core.VolumeSource, out *v1.VolumeSource, s conversion.Scope) error {
- return autoConvert_core_VolumeSource_To_v1_VolumeSource(in, out, s)
-}
-
-func autoConvert_v1_VsphereVirtualDiskVolumeSource_To_core_VsphereVirtualDiskVolumeSource(in *v1.VsphereVirtualDiskVolumeSource, out *core.VsphereVirtualDiskVolumeSource, s conversion.Scope) error {
- out.VolumePath = in.VolumePath
- out.FSType = in.FSType
- out.StoragePolicyName = in.StoragePolicyName
- out.StoragePolicyID = in.StoragePolicyID
- return nil
-}
-
-// Convert_v1_VsphereVirtualDiskVolumeSource_To_core_VsphereVirtualDiskVolumeSource is an autogenerated conversion function.
-func Convert_v1_VsphereVirtualDiskVolumeSource_To_core_VsphereVirtualDiskVolumeSource(in *v1.VsphereVirtualDiskVolumeSource, out *core.VsphereVirtualDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_v1_VsphereVirtualDiskVolumeSource_To_core_VsphereVirtualDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_core_VsphereVirtualDiskVolumeSource_To_v1_VsphereVirtualDiskVolumeSource(in *core.VsphereVirtualDiskVolumeSource, out *v1.VsphereVirtualDiskVolumeSource, s conversion.Scope) error {
- out.VolumePath = in.VolumePath
- out.FSType = in.FSType
- out.StoragePolicyName = in.StoragePolicyName
- out.StoragePolicyID = in.StoragePolicyID
- return nil
-}
-
-// Convert_core_VsphereVirtualDiskVolumeSource_To_v1_VsphereVirtualDiskVolumeSource is an autogenerated conversion function.
-func Convert_core_VsphereVirtualDiskVolumeSource_To_v1_VsphereVirtualDiskVolumeSource(in *core.VsphereVirtualDiskVolumeSource, out *v1.VsphereVirtualDiskVolumeSource, s conversion.Scope) error {
- return autoConvert_core_VsphereVirtualDiskVolumeSource_To_v1_VsphereVirtualDiskVolumeSource(in, out, s)
-}
-
-func autoConvert_v1_WeightedPodAffinityTerm_To_core_WeightedPodAffinityTerm(in *v1.WeightedPodAffinityTerm, out *core.WeightedPodAffinityTerm, s conversion.Scope) error {
- out.Weight = in.Weight
- if err := Convert_v1_PodAffinityTerm_To_core_PodAffinityTerm(&in.PodAffinityTerm, &out.PodAffinityTerm, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_v1_WeightedPodAffinityTerm_To_core_WeightedPodAffinityTerm is an autogenerated conversion function.
-func Convert_v1_WeightedPodAffinityTerm_To_core_WeightedPodAffinityTerm(in *v1.WeightedPodAffinityTerm, out *core.WeightedPodAffinityTerm, s conversion.Scope) error {
- return autoConvert_v1_WeightedPodAffinityTerm_To_core_WeightedPodAffinityTerm(in, out, s)
-}
-
-func autoConvert_core_WeightedPodAffinityTerm_To_v1_WeightedPodAffinityTerm(in *core.WeightedPodAffinityTerm, out *v1.WeightedPodAffinityTerm, s conversion.Scope) error {
- out.Weight = in.Weight
- if err := Convert_core_PodAffinityTerm_To_v1_PodAffinityTerm(&in.PodAffinityTerm, &out.PodAffinityTerm, s); err != nil {
- return err
- }
- return nil
-}
-
-// Convert_core_WeightedPodAffinityTerm_To_v1_WeightedPodAffinityTerm is an autogenerated conversion function.
-func Convert_core_WeightedPodAffinityTerm_To_v1_WeightedPodAffinityTerm(in *core.WeightedPodAffinityTerm, out *v1.WeightedPodAffinityTerm, s conversion.Scope) error {
- return autoConvert_core_WeightedPodAffinityTerm_To_v1_WeightedPodAffinityTerm(in, out, s)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.defaults.go b/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.defaults.go
deleted file mode 100644
index 084c9516b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/v1/zz_generated.defaults.go
+++ /dev/null
@@ -1,638 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by defaulter-gen. DO NOT EDIT.
-
-package v1
-
-import (
- v1 "k8s.io/api/core/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
-)
-
-// RegisterDefaults adds defaulters functions to the given scheme.
-// Public to allow building arbitrary schemes.
-// All generated defaulters are covering - they call all nested defaulters.
-func RegisterDefaults(scheme *runtime.Scheme) error {
- scheme.AddTypeDefaultingFunc(&v1.ConfigMap{}, func(obj interface{}) { SetObjectDefaults_ConfigMap(obj.(*v1.ConfigMap)) })
- scheme.AddTypeDefaultingFunc(&v1.ConfigMapList{}, func(obj interface{}) { SetObjectDefaults_ConfigMapList(obj.(*v1.ConfigMapList)) })
- scheme.AddTypeDefaultingFunc(&v1.Endpoints{}, func(obj interface{}) { SetObjectDefaults_Endpoints(obj.(*v1.Endpoints)) })
- scheme.AddTypeDefaultingFunc(&v1.EndpointsList{}, func(obj interface{}) { SetObjectDefaults_EndpointsList(obj.(*v1.EndpointsList)) })
- scheme.AddTypeDefaultingFunc(&v1.LimitRange{}, func(obj interface{}) { SetObjectDefaults_LimitRange(obj.(*v1.LimitRange)) })
- scheme.AddTypeDefaultingFunc(&v1.LimitRangeList{}, func(obj interface{}) { SetObjectDefaults_LimitRangeList(obj.(*v1.LimitRangeList)) })
- scheme.AddTypeDefaultingFunc(&v1.Namespace{}, func(obj interface{}) { SetObjectDefaults_Namespace(obj.(*v1.Namespace)) })
- scheme.AddTypeDefaultingFunc(&v1.NamespaceList{}, func(obj interface{}) { SetObjectDefaults_NamespaceList(obj.(*v1.NamespaceList)) })
- scheme.AddTypeDefaultingFunc(&v1.Node{}, func(obj interface{}) { SetObjectDefaults_Node(obj.(*v1.Node)) })
- scheme.AddTypeDefaultingFunc(&v1.NodeList{}, func(obj interface{}) { SetObjectDefaults_NodeList(obj.(*v1.NodeList)) })
- scheme.AddTypeDefaultingFunc(&v1.PersistentVolume{}, func(obj interface{}) { SetObjectDefaults_PersistentVolume(obj.(*v1.PersistentVolume)) })
- scheme.AddTypeDefaultingFunc(&v1.PersistentVolumeClaim{}, func(obj interface{}) { SetObjectDefaults_PersistentVolumeClaim(obj.(*v1.PersistentVolumeClaim)) })
- scheme.AddTypeDefaultingFunc(&v1.PersistentVolumeClaimList{}, func(obj interface{}) {
- SetObjectDefaults_PersistentVolumeClaimList(obj.(*v1.PersistentVolumeClaimList))
- })
- scheme.AddTypeDefaultingFunc(&v1.PersistentVolumeList{}, func(obj interface{}) { SetObjectDefaults_PersistentVolumeList(obj.(*v1.PersistentVolumeList)) })
- scheme.AddTypeDefaultingFunc(&v1.Pod{}, func(obj interface{}) { SetObjectDefaults_Pod(obj.(*v1.Pod)) })
- scheme.AddTypeDefaultingFunc(&v1.PodList{}, func(obj interface{}) { SetObjectDefaults_PodList(obj.(*v1.PodList)) })
- scheme.AddTypeDefaultingFunc(&v1.PodTemplate{}, func(obj interface{}) { SetObjectDefaults_PodTemplate(obj.(*v1.PodTemplate)) })
- scheme.AddTypeDefaultingFunc(&v1.PodTemplateList{}, func(obj interface{}) { SetObjectDefaults_PodTemplateList(obj.(*v1.PodTemplateList)) })
- scheme.AddTypeDefaultingFunc(&v1.ReplicationController{}, func(obj interface{}) { SetObjectDefaults_ReplicationController(obj.(*v1.ReplicationController)) })
- scheme.AddTypeDefaultingFunc(&v1.ReplicationControllerList{}, func(obj interface{}) {
- SetObjectDefaults_ReplicationControllerList(obj.(*v1.ReplicationControllerList))
- })
- scheme.AddTypeDefaultingFunc(&v1.ResourceQuota{}, func(obj interface{}) { SetObjectDefaults_ResourceQuota(obj.(*v1.ResourceQuota)) })
- scheme.AddTypeDefaultingFunc(&v1.ResourceQuotaList{}, func(obj interface{}) { SetObjectDefaults_ResourceQuotaList(obj.(*v1.ResourceQuotaList)) })
- scheme.AddTypeDefaultingFunc(&v1.Secret{}, func(obj interface{}) { SetObjectDefaults_Secret(obj.(*v1.Secret)) })
- scheme.AddTypeDefaultingFunc(&v1.SecretList{}, func(obj interface{}) { SetObjectDefaults_SecretList(obj.(*v1.SecretList)) })
- scheme.AddTypeDefaultingFunc(&v1.Service{}, func(obj interface{}) { SetObjectDefaults_Service(obj.(*v1.Service)) })
- scheme.AddTypeDefaultingFunc(&v1.ServiceList{}, func(obj interface{}) { SetObjectDefaults_ServiceList(obj.(*v1.ServiceList)) })
- return nil
-}
-
-func SetObjectDefaults_ConfigMap(in *v1.ConfigMap) {
- SetDefaults_ConfigMap(in)
-}
-
-func SetObjectDefaults_ConfigMapList(in *v1.ConfigMapList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_ConfigMap(a)
- }
-}
-
-func SetObjectDefaults_Endpoints(in *v1.Endpoints) {
- SetDefaults_Endpoints(in)
-}
-
-func SetObjectDefaults_EndpointsList(in *v1.EndpointsList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Endpoints(a)
- }
-}
-
-func SetObjectDefaults_LimitRange(in *v1.LimitRange) {
- for i := range in.Spec.Limits {
- a := &in.Spec.Limits[i]
- SetDefaults_LimitRangeItem(a)
- SetDefaults_ResourceList(&a.Max)
- SetDefaults_ResourceList(&a.Min)
- SetDefaults_ResourceList(&a.Default)
- SetDefaults_ResourceList(&a.DefaultRequest)
- SetDefaults_ResourceList(&a.MaxLimitRequestRatio)
- }
-}
-
-func SetObjectDefaults_LimitRangeList(in *v1.LimitRangeList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_LimitRange(a)
- }
-}
-
-func SetObjectDefaults_Namespace(in *v1.Namespace) {
- SetDefaults_NamespaceStatus(&in.Status)
-}
-
-func SetObjectDefaults_NamespaceList(in *v1.NamespaceList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Namespace(a)
- }
-}
-
-func SetObjectDefaults_Node(in *v1.Node) {
- SetDefaults_Node(in)
- SetDefaults_NodeStatus(&in.Status)
- SetDefaults_ResourceList(&in.Status.Capacity)
- SetDefaults_ResourceList(&in.Status.Allocatable)
-}
-
-func SetObjectDefaults_NodeList(in *v1.NodeList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Node(a)
- }
-}
-
-func SetObjectDefaults_PersistentVolume(in *v1.PersistentVolume) {
- SetDefaults_PersistentVolume(in)
- SetDefaults_ResourceList(&in.Spec.Capacity)
- if in.Spec.PersistentVolumeSource.HostPath != nil {
- SetDefaults_HostPathVolumeSource(in.Spec.PersistentVolumeSource.HostPath)
- }
- if in.Spec.PersistentVolumeSource.RBD != nil {
- SetDefaults_RBDPersistentVolumeSource(in.Spec.PersistentVolumeSource.RBD)
- }
- if in.Spec.PersistentVolumeSource.ISCSI != nil {
- SetDefaults_ISCSIPersistentVolumeSource(in.Spec.PersistentVolumeSource.ISCSI)
- }
- if in.Spec.PersistentVolumeSource.AzureDisk != nil {
- SetDefaults_AzureDiskVolumeSource(in.Spec.PersistentVolumeSource.AzureDisk)
- }
- if in.Spec.PersistentVolumeSource.ScaleIO != nil {
- SetDefaults_ScaleIOPersistentVolumeSource(in.Spec.PersistentVolumeSource.ScaleIO)
- }
-}
-
-func SetObjectDefaults_PersistentVolumeClaim(in *v1.PersistentVolumeClaim) {
- SetDefaults_PersistentVolumeClaim(in)
- SetDefaults_ResourceList(&in.Spec.Resources.Limits)
- SetDefaults_ResourceList(&in.Spec.Resources.Requests)
- SetDefaults_ResourceList(&in.Status.Capacity)
-}
-
-func SetObjectDefaults_PersistentVolumeClaimList(in *v1.PersistentVolumeClaimList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_PersistentVolumeClaim(a)
- }
-}
-
-func SetObjectDefaults_PersistentVolumeList(in *v1.PersistentVolumeList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_PersistentVolume(a)
- }
-}
-
-func SetObjectDefaults_Pod(in *v1.Pod) {
- SetDefaults_Pod(in)
- SetDefaults_PodSpec(&in.Spec)
- for i := range in.Spec.Volumes {
- a := &in.Spec.Volumes[i]
- SetDefaults_Volume(a)
- if a.VolumeSource.HostPath != nil {
- SetDefaults_HostPathVolumeSource(a.VolumeSource.HostPath)
- }
- if a.VolumeSource.Secret != nil {
- SetDefaults_SecretVolumeSource(a.VolumeSource.Secret)
- }
- if a.VolumeSource.ISCSI != nil {
- SetDefaults_ISCSIVolumeSource(a.VolumeSource.ISCSI)
- }
- if a.VolumeSource.RBD != nil {
- SetDefaults_RBDVolumeSource(a.VolumeSource.RBD)
- }
- if a.VolumeSource.DownwardAPI != nil {
- SetDefaults_DownwardAPIVolumeSource(a.VolumeSource.DownwardAPI)
- for j := range a.VolumeSource.DownwardAPI.Items {
- b := &a.VolumeSource.DownwardAPI.Items[j]
- if b.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.FieldRef)
- }
- }
- }
- if a.VolumeSource.ConfigMap != nil {
- SetDefaults_ConfigMapVolumeSource(a.VolumeSource.ConfigMap)
- }
- if a.VolumeSource.AzureDisk != nil {
- SetDefaults_AzureDiskVolumeSource(a.VolumeSource.AzureDisk)
- }
- if a.VolumeSource.Projected != nil {
- SetDefaults_ProjectedVolumeSource(a.VolumeSource.Projected)
- for j := range a.VolumeSource.Projected.Sources {
- b := &a.VolumeSource.Projected.Sources[j]
- if b.DownwardAPI != nil {
- for k := range b.DownwardAPI.Items {
- c := &b.DownwardAPI.Items[k]
- if c.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(c.FieldRef)
- }
- }
- }
- }
- }
- if a.VolumeSource.ScaleIO != nil {
- SetDefaults_ScaleIOVolumeSource(a.VolumeSource.ScaleIO)
- }
- }
- for i := range in.Spec.InitContainers {
- a := &in.Spec.InitContainers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
- for i := range in.Spec.Containers {
- a := &in.Spec.Containers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
-}
-
-func SetObjectDefaults_PodList(in *v1.PodList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Pod(a)
- }
-}
-
-func SetObjectDefaults_PodTemplate(in *v1.PodTemplate) {
- SetDefaults_PodSpec(&in.Template.Spec)
- for i := range in.Template.Spec.Volumes {
- a := &in.Template.Spec.Volumes[i]
- SetDefaults_Volume(a)
- if a.VolumeSource.HostPath != nil {
- SetDefaults_HostPathVolumeSource(a.VolumeSource.HostPath)
- }
- if a.VolumeSource.Secret != nil {
- SetDefaults_SecretVolumeSource(a.VolumeSource.Secret)
- }
- if a.VolumeSource.ISCSI != nil {
- SetDefaults_ISCSIVolumeSource(a.VolumeSource.ISCSI)
- }
- if a.VolumeSource.RBD != nil {
- SetDefaults_RBDVolumeSource(a.VolumeSource.RBD)
- }
- if a.VolumeSource.DownwardAPI != nil {
- SetDefaults_DownwardAPIVolumeSource(a.VolumeSource.DownwardAPI)
- for j := range a.VolumeSource.DownwardAPI.Items {
- b := &a.VolumeSource.DownwardAPI.Items[j]
- if b.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.FieldRef)
- }
- }
- }
- if a.VolumeSource.ConfigMap != nil {
- SetDefaults_ConfigMapVolumeSource(a.VolumeSource.ConfigMap)
- }
- if a.VolumeSource.AzureDisk != nil {
- SetDefaults_AzureDiskVolumeSource(a.VolumeSource.AzureDisk)
- }
- if a.VolumeSource.Projected != nil {
- SetDefaults_ProjectedVolumeSource(a.VolumeSource.Projected)
- for j := range a.VolumeSource.Projected.Sources {
- b := &a.VolumeSource.Projected.Sources[j]
- if b.DownwardAPI != nil {
- for k := range b.DownwardAPI.Items {
- c := &b.DownwardAPI.Items[k]
- if c.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(c.FieldRef)
- }
- }
- }
- }
- }
- if a.VolumeSource.ScaleIO != nil {
- SetDefaults_ScaleIOVolumeSource(a.VolumeSource.ScaleIO)
- }
- }
- for i := range in.Template.Spec.InitContainers {
- a := &in.Template.Spec.InitContainers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
- for i := range in.Template.Spec.Containers {
- a := &in.Template.Spec.Containers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
-}
-
-func SetObjectDefaults_PodTemplateList(in *v1.PodTemplateList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_PodTemplate(a)
- }
-}
-
-func SetObjectDefaults_ReplicationController(in *v1.ReplicationController) {
- SetDefaults_ReplicationController(in)
- if in.Spec.Template != nil {
- SetDefaults_PodSpec(&in.Spec.Template.Spec)
- for i := range in.Spec.Template.Spec.Volumes {
- a := &in.Spec.Template.Spec.Volumes[i]
- SetDefaults_Volume(a)
- if a.VolumeSource.HostPath != nil {
- SetDefaults_HostPathVolumeSource(a.VolumeSource.HostPath)
- }
- if a.VolumeSource.Secret != nil {
- SetDefaults_SecretVolumeSource(a.VolumeSource.Secret)
- }
- if a.VolumeSource.ISCSI != nil {
- SetDefaults_ISCSIVolumeSource(a.VolumeSource.ISCSI)
- }
- if a.VolumeSource.RBD != nil {
- SetDefaults_RBDVolumeSource(a.VolumeSource.RBD)
- }
- if a.VolumeSource.DownwardAPI != nil {
- SetDefaults_DownwardAPIVolumeSource(a.VolumeSource.DownwardAPI)
- for j := range a.VolumeSource.DownwardAPI.Items {
- b := &a.VolumeSource.DownwardAPI.Items[j]
- if b.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.FieldRef)
- }
- }
- }
- if a.VolumeSource.ConfigMap != nil {
- SetDefaults_ConfigMapVolumeSource(a.VolumeSource.ConfigMap)
- }
- if a.VolumeSource.AzureDisk != nil {
- SetDefaults_AzureDiskVolumeSource(a.VolumeSource.AzureDisk)
- }
- if a.VolumeSource.Projected != nil {
- SetDefaults_ProjectedVolumeSource(a.VolumeSource.Projected)
- for j := range a.VolumeSource.Projected.Sources {
- b := &a.VolumeSource.Projected.Sources[j]
- if b.DownwardAPI != nil {
- for k := range b.DownwardAPI.Items {
- c := &b.DownwardAPI.Items[k]
- if c.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(c.FieldRef)
- }
- }
- }
- }
- }
- if a.VolumeSource.ScaleIO != nil {
- SetDefaults_ScaleIOVolumeSource(a.VolumeSource.ScaleIO)
- }
- }
- for i := range in.Spec.Template.Spec.InitContainers {
- a := &in.Spec.Template.Spec.InitContainers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
- for i := range in.Spec.Template.Spec.Containers {
- a := &in.Spec.Template.Spec.Containers[i]
- SetDefaults_Container(a)
- for j := range a.Ports {
- b := &a.Ports[j]
- SetDefaults_ContainerPort(b)
- }
- for j := range a.Env {
- b := &a.Env[j]
- if b.ValueFrom != nil {
- if b.ValueFrom.FieldRef != nil {
- SetDefaults_ObjectFieldSelector(b.ValueFrom.FieldRef)
- }
- }
- }
- SetDefaults_ResourceList(&a.Resources.Limits)
- SetDefaults_ResourceList(&a.Resources.Requests)
- if a.LivenessProbe != nil {
- SetDefaults_Probe(a.LivenessProbe)
- if a.LivenessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.LivenessProbe.Handler.HTTPGet)
- }
- }
- if a.ReadinessProbe != nil {
- SetDefaults_Probe(a.ReadinessProbe)
- if a.ReadinessProbe.Handler.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.ReadinessProbe.Handler.HTTPGet)
- }
- }
- if a.Lifecycle != nil {
- if a.Lifecycle.PostStart != nil {
- if a.Lifecycle.PostStart.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PostStart.HTTPGet)
- }
- }
- if a.Lifecycle.PreStop != nil {
- if a.Lifecycle.PreStop.HTTPGet != nil {
- SetDefaults_HTTPGetAction(a.Lifecycle.PreStop.HTTPGet)
- }
- }
- }
- }
- }
-}
-
-func SetObjectDefaults_ReplicationControllerList(in *v1.ReplicationControllerList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_ReplicationController(a)
- }
-}
-
-func SetObjectDefaults_ResourceQuota(in *v1.ResourceQuota) {
- SetDefaults_ResourceList(&in.Spec.Hard)
- SetDefaults_ResourceList(&in.Status.Hard)
- SetDefaults_ResourceList(&in.Status.Used)
-}
-
-func SetObjectDefaults_ResourceQuotaList(in *v1.ResourceQuotaList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_ResourceQuota(a)
- }
-}
-
-func SetObjectDefaults_Secret(in *v1.Secret) {
- SetDefaults_Secret(in)
-}
-
-func SetObjectDefaults_SecretList(in *v1.SecretList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Secret(a)
- }
-}
-
-func SetObjectDefaults_Service(in *v1.Service) {
- SetDefaults_Service(in)
-}
-
-func SetObjectDefaults_ServiceList(in *v1.ServiceList) {
- for i := range in.Items {
- a := &in.Items[i]
- SetObjectDefaults_Service(a)
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/core/validation/doc.go
deleted file mode 100644
index 0c1cfaab5..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package validation has functions for validating the correctness of api
-// objects and explaining what is wrong with them when they aren't valid.
-package validation // import "k8s.io/kubernetes/pkg/apis/core/validation"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/events.go b/vendor/k8s.io/kubernetes/pkg/apis/core/validation/events.go
deleted file mode 100644
index ab265bd76..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/events.go
+++ /dev/null
@@ -1,129 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package validation
-
-import (
- "fmt"
- "time"
-
- "k8s.io/apimachinery/pkg/api/meta"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/runtime/schema"
- "k8s.io/apimachinery/pkg/util/validation"
- "k8s.io/apimachinery/pkg/util/validation/field"
- "k8s.io/kubernetes/pkg/api/legacyscheme"
- "k8s.io/kubernetes/pkg/apis/core"
-)
-
-const (
- ReportingInstanceLengthLimit = 128
- ActionLengthLimit = 128
- ReasonLengthLimit = 128
- NoteLengthLimit = 1024
-)
-
-// ValidateEvent makes sure that the event makes sense.
-func ValidateEvent(event *core.Event) field.ErrorList {
- allErrs := field.ErrorList{}
- // Because go
- zeroTime := time.Time{}
-
- // "New" Events need to have EventTime set, so it's validating old object.
- if event.EventTime.Time == zeroTime {
- // Make sure event.Namespace and the involvedInvolvedObject.Namespace agree
- if len(event.InvolvedObject.Namespace) == 0 {
- // event.Namespace must also be empty (or "default", for compatibility with old clients)
- if event.Namespace != metav1.NamespaceNone && event.Namespace != metav1.NamespaceDefault {
- allErrs = append(allErrs, field.Invalid(field.NewPath("involvedObject", "namespace"), event.InvolvedObject.Namespace, "does not match event.namespace"))
- }
- } else {
- // event namespace must match
- if event.Namespace != event.InvolvedObject.Namespace {
- allErrs = append(allErrs, field.Invalid(field.NewPath("involvedObject", "namespace"), event.InvolvedObject.Namespace, "does not match event.namespace"))
- }
- }
-
- } else {
- if len(event.InvolvedObject.Namespace) == 0 && event.Namespace != metav1.NamespaceSystem {
- allErrs = append(allErrs, field.Invalid(field.NewPath("involvedObject", "namespace"), event.InvolvedObject.Namespace, "does not match event.namespace"))
- }
- if len(event.ReportingController) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("reportingController"), ""))
- }
- for _, msg := range validation.IsQualifiedName(event.ReportingController) {
- allErrs = append(allErrs, field.Invalid(field.NewPath("reportingController"), event.ReportingController, msg))
- }
- if len(event.ReportingInstance) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("reportingInstance"), ""))
- }
- if len(event.ReportingInstance) > ReportingInstanceLengthLimit {
- allErrs = append(allErrs, field.Invalid(field.NewPath("repotingIntance"), "", fmt.Sprintf("can have at most %v characters", ReportingInstanceLengthLimit)))
- }
- if len(event.Action) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("action"), ""))
- }
- if len(event.Action) > ActionLengthLimit {
- allErrs = append(allErrs, field.Invalid(field.NewPath("action"), "", fmt.Sprintf("can have at most %v characters", ActionLengthLimit)))
- }
- if len(event.Reason) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("reason"), ""))
- }
- if len(event.Reason) > ReasonLengthLimit {
- allErrs = append(allErrs, field.Invalid(field.NewPath("reason"), "", fmt.Sprintf("can have at most %v characters", ReasonLengthLimit)))
- }
- if len(event.Message) > NoteLengthLimit {
- allErrs = append(allErrs, field.Invalid(field.NewPath("message"), "", fmt.Sprintf("can have at most %v characters", NoteLengthLimit)))
- }
- }
-
- // For kinds we recognize, make sure InvolvedObject.Namespace is set for namespaced kinds
- if namespaced, err := isNamespacedKind(event.InvolvedObject.Kind, event.InvolvedObject.APIVersion); err == nil {
- if namespaced && len(event.InvolvedObject.Namespace) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("involvedObject", "namespace"), fmt.Sprintf("required for kind %s", event.InvolvedObject.Kind)))
- }
- if !namespaced && len(event.InvolvedObject.Namespace) > 0 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("involvedObject", "namespace"), event.InvolvedObject.Namespace, fmt.Sprintf("not allowed for kind %s", event.InvolvedObject.Kind)))
- }
- }
-
- for _, msg := range validation.IsDNS1123Subdomain(event.Namespace) {
- allErrs = append(allErrs, field.Invalid(field.NewPath("namespace"), event.Namespace, msg))
- }
- return allErrs
-}
-
-// Check whether the kind in groupVersion is scoped at the root of the api hierarchy
-func isNamespacedKind(kind, groupVersion string) (bool, error) {
- gv, err := schema.ParseGroupVersion(groupVersion)
- if err != nil {
- return false, err
- }
- g, err := legacyscheme.Registry.Group(gv.Group)
- if err != nil {
- return false, err
- }
-
- restMapping, err := g.RESTMapper.RESTMapping(schema.GroupKind{Group: gv.Group, Kind: kind}, gv.Version)
- if err != nil {
- return false, err
- }
- scopeName := restMapping.Scope.Name()
- if scopeName == meta.RESTScopeNameNamespace {
- return true, nil
- }
- return false, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/validation.go b/vendor/k8s.io/kubernetes/pkg/apis/core/validation/validation.go
deleted file mode 100644
index ec93a1968..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/validation/validation.go
+++ /dev/null
@@ -1,5066 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package validation
-
-import (
- "encoding/json"
- "fmt"
- "math"
- "net"
- "path"
- "path/filepath"
- "reflect"
- "regexp"
- "strings"
-
- "github.com/golang/glog"
-
- "k8s.io/api/core/v1"
- apiequality "k8s.io/apimachinery/pkg/api/equality"
- "k8s.io/apimachinery/pkg/api/resource"
- apimachineryvalidation "k8s.io/apimachinery/pkg/api/validation"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- unversionedvalidation "k8s.io/apimachinery/pkg/apis/meta/v1/validation"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/util/diff"
- "k8s.io/apimachinery/pkg/util/intstr"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/validation"
- "k8s.io/apimachinery/pkg/util/validation/field"
- utilfeature "k8s.io/apiserver/pkg/util/feature"
- apiservice "k8s.io/kubernetes/pkg/api/service"
- "k8s.io/kubernetes/pkg/apis/core"
- "k8s.io/kubernetes/pkg/apis/core/helper"
- podshelper "k8s.io/kubernetes/pkg/apis/core/pods"
- corev1 "k8s.io/kubernetes/pkg/apis/core/v1"
- v1helper "k8s.io/kubernetes/pkg/apis/core/v1/helper"
- "k8s.io/kubernetes/pkg/capabilities"
- "k8s.io/kubernetes/pkg/features"
- "k8s.io/kubernetes/pkg/fieldpath"
- "k8s.io/kubernetes/pkg/master/ports"
- "k8s.io/kubernetes/pkg/security/apparmor"
-)
-
-// TODO: delete this global variable when we enable the validation of common
-// fields by default.
-var RepairMalformedUpdates bool = apimachineryvalidation.RepairMalformedUpdates
-
-const isNegativeErrorMsg string = apimachineryvalidation.IsNegativeErrorMsg
-const isInvalidQuotaResource string = `must be a standard resource for quota`
-const fieldImmutableErrorMsg string = apimachineryvalidation.FieldImmutableErrorMsg
-const isNotIntegerErrorMsg string = `must be an integer`
-const isNotPositiveErrorMsg string = `must be greater than zero`
-const csiDriverNameRexpErrMsg string = "must consist of alphanumeric characters, '-', '_' or '.', and must start and end with an alphanumeric character"
-const csiDriverNameRexpFmt string = `^[a-zA-Z0-9][-a-zA-Z0-9_.]{0,61}[a-zA-Z-0-9]$`
-
-var pdPartitionErrorMsg string = validation.InclusiveRangeError(1, 255)
-var fileModeErrorMsg string = "must be a number between 0 and 0777 (octal), both inclusive"
-
-// BannedOwners is a black list of object that are not allowed to be owners.
-var BannedOwners = apimachineryvalidation.BannedOwners
-
-var iscsiInitiatorIqnRegex = regexp.MustCompile(`iqn\.\d{4}-\d{2}\.([[:alnum:]-.]+)(:[^,;*&$|\s]+)$`)
-var iscsiInitiatorEuiRegex = regexp.MustCompile(`^eui.[[:alnum:]]{16}$`)
-var iscsiInitiatorNaaRegex = regexp.MustCompile(`^naa.[[:alnum:]]{32}$`)
-
-var csiDriverNameRexp = regexp.MustCompile(csiDriverNameRexpFmt)
-
-// ValidateHasLabel requires that metav1.ObjectMeta has a Label with key and expectedValue
-func ValidateHasLabel(meta metav1.ObjectMeta, fldPath *field.Path, key, expectedValue string) field.ErrorList {
- allErrs := field.ErrorList{}
- actualValue, found := meta.Labels[key]
- if !found {
- allErrs = append(allErrs, field.Required(fldPath.Child("labels").Key(key),
- fmt.Sprintf("must be '%s'", expectedValue)))
- return allErrs
- }
- if actualValue != expectedValue {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("labels").Key(key), meta.Labels,
- fmt.Sprintf("must be '%s'", expectedValue)))
- }
- return allErrs
-}
-
-// ValidateAnnotations validates that a set of annotations are correctly defined.
-func ValidateAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- return apimachineryvalidation.ValidateAnnotations(annotations, fldPath)
-}
-
-func ValidateDNS1123Label(value string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, msg := range validation.IsDNS1123Label(value) {
- allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
- }
- return allErrs
-}
-
-// ValidateDNS1123Subdomain validates that a name is a proper DNS subdomain.
-func ValidateDNS1123Subdomain(value string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, msg := range validation.IsDNS1123Subdomain(value) {
- allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
- }
- return allErrs
-}
-
-func ValidatePodSpecificAnnotations(annotations map[string]string, spec *core.PodSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if value, isMirror := annotations[core.MirrorPodAnnotationKey]; isMirror {
- if len(spec.NodeName) == 0 {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(core.MirrorPodAnnotationKey), value, "must set spec.nodeName if mirror pod annotation is set"))
- }
- }
-
- if annotations[core.TolerationsAnnotationKey] != "" {
- allErrs = append(allErrs, ValidateTolerationsInPodAnnotations(annotations, fldPath)...)
- }
-
- allErrs = append(allErrs, ValidateSeccompPodAnnotations(annotations, fldPath)...)
- allErrs = append(allErrs, ValidateAppArmorPodAnnotations(annotations, spec, fldPath)...)
-
- sysctls, err := helper.SysctlsFromPodAnnotation(annotations[core.SysctlsPodAnnotationKey])
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(core.SysctlsPodAnnotationKey), annotations[core.SysctlsPodAnnotationKey], err.Error()))
- } else {
- allErrs = append(allErrs, validateSysctls(sysctls, fldPath.Key(core.SysctlsPodAnnotationKey))...)
- }
- unsafeSysctls, err := helper.SysctlsFromPodAnnotation(annotations[core.UnsafeSysctlsPodAnnotationKey])
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(core.UnsafeSysctlsPodAnnotationKey), annotations[core.UnsafeSysctlsPodAnnotationKey], err.Error()))
- } else {
- allErrs = append(allErrs, validateSysctls(unsafeSysctls, fldPath.Key(core.UnsafeSysctlsPodAnnotationKey))...)
- }
- inBoth := sysctlIntersection(sysctls, unsafeSysctls)
- if len(inBoth) > 0 {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(core.UnsafeSysctlsPodAnnotationKey), strings.Join(inBoth, ", "), "can not be safe and unsafe"))
- }
-
- return allErrs
-}
-
-// ValidateTolerationsInPodAnnotations tests that the serialized tolerations in Pod.Annotations has valid data
-func ValidateTolerationsInPodAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- tolerations, err := helper.GetTolerationsFromPodAnnotations(annotations)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath, core.TolerationsAnnotationKey, err.Error()))
- return allErrs
- }
-
- if len(tolerations) > 0 {
- allErrs = append(allErrs, ValidateTolerations(tolerations, fldPath.Child(core.TolerationsAnnotationKey))...)
- }
-
- return allErrs
-}
-
-func ValidatePodSpecificAnnotationUpdates(newPod, oldPod *core.Pod, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- newAnnotations := newPod.Annotations
- oldAnnotations := oldPod.Annotations
- for k, oldVal := range oldAnnotations {
- if newVal, exists := newAnnotations[k]; exists && newVal == oldVal {
- continue // No change.
- }
- if strings.HasPrefix(k, apparmor.ContainerAnnotationKeyPrefix) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not remove or update AppArmor annotations"))
- }
- if k == core.MirrorPodAnnotationKey {
- allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not remove or update mirror pod annotation"))
- }
- }
- // Check for additions
- for k := range newAnnotations {
- if _, ok := oldAnnotations[k]; ok {
- continue // No change.
- }
- if strings.HasPrefix(k, apparmor.ContainerAnnotationKeyPrefix) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not add AppArmor annotations"))
- }
- if k == core.MirrorPodAnnotationKey {
- allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not add mirror pod annotation"))
- }
- }
- allErrs = append(allErrs, ValidatePodSpecificAnnotations(newAnnotations, &newPod.Spec, fldPath)...)
- return allErrs
-}
-
-func ValidateEndpointsSpecificAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- return allErrs
-}
-
-// ValidateNameFunc validates that the provided name is valid for a given resource type.
-// Not all resources have the same validation rules for names. Prefix is true
-// if the name will have a value appended to it. If the name is not valid,
-// this returns a list of descriptions of individual characteristics of the
-// value that were not valid. Otherwise this returns an empty list or nil.
-type ValidateNameFunc apimachineryvalidation.ValidateNameFunc
-
-// ValidatePodName can be used to check whether the given pod name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidatePodName = NameIsDNSSubdomain
-
-// ValidateReplicationControllerName can be used to check whether the given replication
-// controller name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateReplicationControllerName = NameIsDNSSubdomain
-
-// ValidateServiceName can be used to check whether the given service name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateServiceName = NameIsDNS1035Label
-
-// ValidateNodeName can be used to check whether the given node name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateNodeName = NameIsDNSSubdomain
-
-// ValidateNamespaceName can be used to check whether the given namespace name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateNamespaceName = apimachineryvalidation.ValidateNamespaceName
-
-// ValidateLimitRangeName can be used to check whether the given limit range name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateLimitRangeName = NameIsDNSSubdomain
-
-// ValidateResourceQuotaName can be used to check whether the given
-// resource quota name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateResourceQuotaName = NameIsDNSSubdomain
-
-// ValidateSecretName can be used to check whether the given secret name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateSecretName = NameIsDNSSubdomain
-
-// ValidateServiceAccountName can be used to check whether the given service account name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateServiceAccountName = apimachineryvalidation.ValidateServiceAccountName
-
-// ValidateEndpointsName can be used to check whether the given endpoints name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateEndpointsName = NameIsDNSSubdomain
-
-// ValidateClusterName can be used to check whether the given cluster name is valid.
-var ValidateClusterName = apimachineryvalidation.ValidateClusterName
-
-// ValidateClassName can be used to check whether the given class name is valid.
-// It is defined here to avoid import cycle between pkg/apis/storage/validation
-// (where it should be) and this file.
-var ValidateClassName = NameIsDNSSubdomain
-
-// ValidatePiorityClassName can be used to check whether the given priority
-// class name is valid.
-var ValidatePriorityClassName = NameIsDNSSubdomain
-
-// TODO update all references to these functions to point to the apimachineryvalidation ones
-// NameIsDNSSubdomain is a ValidateNameFunc for names that must be a DNS subdomain.
-func NameIsDNSSubdomain(name string, prefix bool) []string {
- return apimachineryvalidation.NameIsDNSSubdomain(name, prefix)
-}
-
-// NameIsDNS1035Label is a ValidateNameFunc for names that must be a DNS 952 label.
-func NameIsDNS1035Label(name string, prefix bool) []string {
- return apimachineryvalidation.NameIsDNS1035Label(name, prefix)
-}
-
-// Validates that given value is not negative.
-func ValidateNonnegativeField(value int64, fldPath *field.Path) field.ErrorList {
- return apimachineryvalidation.ValidateNonnegativeField(value, fldPath)
-}
-
-// Validates that a Quantity is not negative
-func ValidateNonnegativeQuantity(value resource.Quantity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if value.Cmp(resource.Quantity{}) < 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, value.String(), isNegativeErrorMsg))
- }
- return allErrs
-}
-
-// Validates that a Quantity is positive
-func ValidatePositiveQuantityValue(value resource.Quantity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if value.Cmp(resource.Quantity{}) <= 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, value.String(), isNotPositiveErrorMsg))
- }
- return allErrs
-}
-
-func ValidateImmutableField(newVal, oldVal interface{}, fldPath *field.Path) field.ErrorList {
- return apimachineryvalidation.ValidateImmutableField(newVal, oldVal, fldPath)
-}
-
-func ValidateImmutableAnnotation(newVal string, oldVal string, annotation string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if oldVal != newVal {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("annotations", annotation), newVal, fieldImmutableErrorMsg))
- }
- return allErrs
-}
-
-// ValidateObjectMeta validates an object's metadata on creation. It expects that name generation has already
-// been performed.
-// It doesn't return an error for rootscoped resources with namespace, because namespace should already be cleared before.
-// TODO: Remove calls to this method scattered in validations of specific resources, e.g., ValidatePodUpdate.
-func ValidateObjectMeta(meta *metav1.ObjectMeta, requiresNamespace bool, nameFn ValidateNameFunc, fldPath *field.Path) field.ErrorList {
- allErrs := apimachineryvalidation.ValidateObjectMeta(meta, requiresNamespace, apimachineryvalidation.ValidateNameFunc(nameFn), fldPath)
- // run additional checks for the finalizer name
- for i := range meta.Finalizers {
- allErrs = append(allErrs, validateKubeFinalizerName(string(meta.Finalizers[i]), fldPath.Child("finalizers").Index(i))...)
- }
- return allErrs
-}
-
-// ValidateObjectMetaUpdate validates an object's metadata when updated
-func ValidateObjectMetaUpdate(newMeta, oldMeta *metav1.ObjectMeta, fldPath *field.Path) field.ErrorList {
- allErrs := apimachineryvalidation.ValidateObjectMetaUpdate(newMeta, oldMeta, fldPath)
- // run additional checks for the finalizer name
- for i := range newMeta.Finalizers {
- allErrs = append(allErrs, validateKubeFinalizerName(string(newMeta.Finalizers[i]), fldPath.Child("finalizers").Index(i))...)
- }
-
- return allErrs
-}
-
-func ValidateVolumes(volumes []core.Volume, fldPath *field.Path) (map[string]core.VolumeSource, field.ErrorList) {
- allErrs := field.ErrorList{}
-
- allNames := sets.String{}
- vols := make(map[string]core.VolumeSource)
- for i, vol := range volumes {
- idxPath := fldPath.Index(i)
- namePath := idxPath.Child("name")
- el := validateVolumeSource(&vol.VolumeSource, idxPath, vol.Name)
- if len(vol.Name) == 0 {
- el = append(el, field.Required(namePath, ""))
- } else {
- el = append(el, ValidateDNS1123Label(vol.Name, namePath)...)
- }
- if allNames.Has(vol.Name) {
- el = append(el, field.Duplicate(namePath, vol.Name))
- }
- if len(el) == 0 {
- allNames.Insert(vol.Name)
- vols[vol.Name] = vol.VolumeSource
- } else {
- allErrs = append(allErrs, el...)
- }
-
- }
- return vols, allErrs
-}
-
-func IsMatchedVolume(name string, volumes map[string]core.VolumeSource) bool {
- if _, ok := volumes[name]; ok {
- return true
- } else {
- return false
- }
-}
-
-func isMatchedDevice(name string, volumes map[string]core.VolumeSource) (bool, bool) {
- if source, ok := volumes[name]; ok {
- if source.PersistentVolumeClaim != nil {
- return true, true
- } else {
- return true, false
- }
- } else {
- return false, false
- }
-}
-
-func mountNameAlreadyExists(name string, devices map[string]string) bool {
- if _, ok := devices[name]; ok {
- return true
- } else {
- return false
- }
-}
-
-func mountPathAlreadyExists(mountPath string, devices map[string]string) bool {
- for _, devPath := range devices {
- if mountPath == devPath {
- return true
- }
- }
-
- return false
-}
-
-func deviceNameAlreadyExists(name string, mounts map[string]string) bool {
- if _, ok := mounts[name]; ok {
- return true
- } else {
- return false
- }
-}
-
-func devicePathAlreadyExists(devicePath string, mounts map[string]string) bool {
- for _, mountPath := range mounts {
- if mountPath == devicePath {
- return true
- }
- }
-
- return false
-}
-
-func validateVolumeSource(source *core.VolumeSource, fldPath *field.Path, volName string) field.ErrorList {
- numVolumes := 0
- allErrs := field.ErrorList{}
- if source.EmptyDir != nil {
- numVolumes++
- if !utilfeature.DefaultFeatureGate.Enabled(features.LocalStorageCapacityIsolation) {
- if source.EmptyDir.SizeLimit != nil && source.EmptyDir.SizeLimit.Cmp(resource.Quantity{}) != 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("emptyDir").Child("sizeLimit"), "SizeLimit field disabled by feature-gate for EmptyDir volumes"))
- }
- } else {
- if source.EmptyDir.SizeLimit != nil && source.EmptyDir.SizeLimit.Cmp(resource.Quantity{}) < 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("emptyDir").Child("sizeLimit"), "SizeLimit field must be a valid resource quantity"))
- }
- }
- if !utilfeature.DefaultFeatureGate.Enabled(features.HugePages) && source.EmptyDir.Medium == core.StorageMediumHugePages {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("emptyDir").Child("medium"), "HugePages medium is disabled by feature-gate for EmptyDir volumes"))
- }
- }
- if source.HostPath != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("hostPath"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateHostPathVolumeSource(source.HostPath, fldPath.Child("hostPath"))...)
- }
- }
- if source.GitRepo != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("gitRepo"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateGitRepoVolumeSource(source.GitRepo, fldPath.Child("gitRepo"))...)
- }
- }
- if source.GCEPersistentDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("gcePersistentDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateGCEPersistentDiskVolumeSource(source.GCEPersistentDisk, fldPath.Child("persistentDisk"))...)
- }
- }
- if source.AWSElasticBlockStore != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("awsElasticBlockStore"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAWSElasticBlockStoreVolumeSource(source.AWSElasticBlockStore, fldPath.Child("awsElasticBlockStore"))...)
- }
- }
- if source.Secret != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("secret"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateSecretVolumeSource(source.Secret, fldPath.Child("secret"))...)
- }
- }
- if source.NFS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("nfs"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateNFSVolumeSource(source.NFS, fldPath.Child("nfs"))...)
- }
- }
- if source.ISCSI != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("iscsi"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateISCSIVolumeSource(source.ISCSI, fldPath.Child("iscsi"))...)
- }
- if source.ISCSI.InitiatorName != nil && len(volName+":"+source.ISCSI.TargetPortal) > 64 {
- tooLongErr := "Total length of <volume name>:<iscsi.targetPortal> must be under 64 characters if iscsi.initiatorName is specified."
- allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), volName, tooLongErr))
- }
- }
- if source.Glusterfs != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("glusterfs"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateGlusterfsVolumeSource(source.Glusterfs, fldPath.Child("glusterfs"))...)
- }
- }
- if source.Flocker != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("flocker"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateFlockerVolumeSource(source.Flocker, fldPath.Child("flocker"))...)
- }
- }
- if source.PersistentVolumeClaim != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("persistentVolumeClaim"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validatePersistentClaimVolumeSource(source.PersistentVolumeClaim, fldPath.Child("persistentVolumeClaim"))...)
- }
- }
- if source.RBD != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("rbd"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateRBDVolumeSource(source.RBD, fldPath.Child("rbd"))...)
- }
- }
- if source.Cinder != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("cinder"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateCinderVolumeSource(source.Cinder, fldPath.Child("cinder"))...)
- }
- }
- if source.CephFS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("cephFS"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateCephFSVolumeSource(source.CephFS, fldPath.Child("cephfs"))...)
- }
- }
- if source.Quobyte != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("quobyte"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateQuobyteVolumeSource(source.Quobyte, fldPath.Child("quobyte"))...)
- }
- }
- if source.DownwardAPI != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("downwarAPI"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateDownwardAPIVolumeSource(source.DownwardAPI, fldPath.Child("downwardAPI"))...)
- }
- }
- if source.FC != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("fc"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateFCVolumeSource(source.FC, fldPath.Child("fc"))...)
- }
- }
- if source.FlexVolume != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("flexVolume"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateFlexVolumeSource(source.FlexVolume, fldPath.Child("flexVolume"))...)
- }
- }
- if source.ConfigMap != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("configMap"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateConfigMapVolumeSource(source.ConfigMap, fldPath.Child("configMap"))...)
- }
- }
-
- if source.AzureFile != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureFile"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAzureFile(source.AzureFile, fldPath.Child("azureFile"))...)
- }
- }
-
- if source.VsphereVolume != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("vsphereVolume"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateVsphereVolumeSource(source.VsphereVolume, fldPath.Child("vsphereVolume"))...)
- }
- }
- if source.PhotonPersistentDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("photonPersistentDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validatePhotonPersistentDiskVolumeSource(source.PhotonPersistentDisk, fldPath.Child("photonPersistentDisk"))...)
- }
- }
- if source.PortworxVolume != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("portworxVolume"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validatePortworxVolumeSource(source.PortworxVolume, fldPath.Child("portworxVolume"))...)
- }
- }
- if source.AzureDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAzureDisk(source.AzureDisk, fldPath.Child("azureDisk"))...)
- }
- }
- if source.StorageOS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("storageos"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateStorageOSVolumeSource(source.StorageOS, fldPath.Child("storageos"))...)
- }
- }
- if source.Projected != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("projected"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateProjectedVolumeSource(source.Projected, fldPath.Child("projected"))...)
- }
- }
- if source.ScaleIO != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("scaleIO"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateScaleIOVolumeSource(source.ScaleIO, fldPath.Child("scaleIO"))...)
- }
- }
-
- if numVolumes == 0 {
- allErrs = append(allErrs, field.Required(fldPath, "must specify a volume type"))
- }
-
- return allErrs
-}
-
-func validateHostPathVolumeSource(hostPath *core.HostPathVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(hostPath.Path) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- return allErrs
- }
-
- allErrs = append(allErrs, validatePathNoBacksteps(hostPath.Path, fldPath.Child("path"))...)
- allErrs = append(allErrs, validateHostPathType(hostPath.Type, fldPath.Child("type"))...)
- return allErrs
-}
-
-func validateGitRepoVolumeSource(gitRepo *core.GitRepoVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(gitRepo.Repository) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("repository"), ""))
- }
-
- pathErrs := validateLocalDescendingPath(gitRepo.Directory, fldPath.Child("directory"))
- allErrs = append(allErrs, pathErrs...)
- return allErrs
-}
-
-func validateISCSIVolumeSource(iscsi *core.ISCSIVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(iscsi.TargetPortal) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("targetPortal"), ""))
- }
- if len(iscsi.IQN) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("iqn"), ""))
- } else {
- if !strings.HasPrefix(iscsi.IQN, "iqn") && !strings.HasPrefix(iscsi.IQN, "eui") && !strings.HasPrefix(iscsi.IQN, "naa") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format starting with iqn, eui, or naa"))
- } else if strings.HasPrefix(iscsi.IQN, "iqn") && !iscsiInitiatorIqnRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- } else if strings.HasPrefix(iscsi.IQN, "eui") && !iscsiInitiatorEuiRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- } else if strings.HasPrefix(iscsi.IQN, "naa") && !iscsiInitiatorNaaRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- }
- }
- if iscsi.Lun < 0 || iscsi.Lun > 255 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), iscsi.Lun, validation.InclusiveRangeError(0, 255)))
- }
- if (iscsi.DiscoveryCHAPAuth || iscsi.SessionCHAPAuth) && iscsi.SecretRef == nil {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef"), ""))
- }
- if iscsi.InitiatorName != nil {
- initiator := *iscsi.InitiatorName
- if !strings.HasPrefix(initiator, "iqn") && !strings.HasPrefix(initiator, "eui") && !strings.HasPrefix(initiator, "naa") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format starting with iqn, eui, or naa"))
- }
- if strings.HasPrefix(initiator, "iqn") && !iscsiInitiatorIqnRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- } else if strings.HasPrefix(initiator, "eui") && !iscsiInitiatorEuiRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- } else if strings.HasPrefix(initiator, "naa") && !iscsiInitiatorNaaRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- }
- }
- return allErrs
-}
-
-func validateISCSIPersistentVolumeSource(iscsi *core.ISCSIPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(iscsi.TargetPortal) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("targetPortal"), ""))
- }
- if len(iscsi.IQN) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("iqn"), ""))
- } else {
- if !strings.HasPrefix(iscsi.IQN, "iqn") && !strings.HasPrefix(iscsi.IQN, "eui") && !strings.HasPrefix(iscsi.IQN, "naa") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- } else if strings.HasPrefix(iscsi.IQN, "iqn") && !iscsiInitiatorIqnRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- } else if strings.HasPrefix(iscsi.IQN, "eui") && !iscsiInitiatorEuiRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- } else if strings.HasPrefix(iscsi.IQN, "naa") && !iscsiInitiatorNaaRegex.MatchString(iscsi.IQN) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
- }
- }
- if iscsi.Lun < 0 || iscsi.Lun > 255 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), iscsi.Lun, validation.InclusiveRangeError(0, 255)))
- }
- if (iscsi.DiscoveryCHAPAuth || iscsi.SessionCHAPAuth) && iscsi.SecretRef == nil {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef"), ""))
- }
- if iscsi.SecretRef != nil {
- if len(iscsi.SecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
- }
- }
- if iscsi.InitiatorName != nil {
- initiator := *iscsi.InitiatorName
- if !strings.HasPrefix(initiator, "iqn") && !strings.HasPrefix(initiator, "eui") && !strings.HasPrefix(initiator, "naa") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- }
- if strings.HasPrefix(initiator, "iqn") && !iscsiInitiatorIqnRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- } else if strings.HasPrefix(initiator, "eui") && !iscsiInitiatorEuiRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- } else if strings.HasPrefix(initiator, "naa") && !iscsiInitiatorNaaRegex.MatchString(initiator) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
- }
- }
- return allErrs
-}
-
-func validateFCVolumeSource(fc *core.FCVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(fc.TargetWWNs) < 1 && len(fc.WWIDs) < 1 {
- allErrs = append(allErrs, field.Required(fldPath.Child("targetWWNs"), "must specify either targetWWNs or wwids, but not both"))
- }
-
- if len(fc.TargetWWNs) != 0 && len(fc.WWIDs) != 0 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("targetWWNs"), fc.TargetWWNs, "targetWWNs and wwids can not be specified simultaneously"))
- }
-
- if len(fc.TargetWWNs) != 0 {
- if fc.Lun == nil {
- allErrs = append(allErrs, field.Required(fldPath.Child("lun"), "lun is required if targetWWNs is specified"))
- } else {
- if *fc.Lun < 0 || *fc.Lun > 255 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), fc.Lun, validation.InclusiveRangeError(0, 255)))
- }
- }
- }
- return allErrs
-}
-
-func validateGCEPersistentDiskVolumeSource(pd *core.GCEPersistentDiskVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(pd.PDName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("pdName"), ""))
- }
- if pd.Partition < 0 || pd.Partition > 255 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("partition"), pd.Partition, pdPartitionErrorMsg))
- }
- return allErrs
-}
-
-func validateAWSElasticBlockStoreVolumeSource(PD *core.AWSElasticBlockStoreVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(PD.VolumeID) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
- }
- if PD.Partition < 0 || PD.Partition > 255 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("partition"), PD.Partition, pdPartitionErrorMsg))
- }
- return allErrs
-}
-
-func validateSecretVolumeSource(secretSource *core.SecretVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(secretSource.SecretName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
- }
-
- secretMode := secretSource.DefaultMode
- if secretMode != nil && (*secretMode > 0777 || *secretMode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *secretMode, fileModeErrorMsg))
- }
-
- itemsPath := fldPath.Child("items")
- for i, kp := range secretSource.Items {
- itemPath := itemsPath.Index(i)
- allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
- }
- return allErrs
-}
-
-func validateConfigMapVolumeSource(configMapSource *core.ConfigMapVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(configMapSource.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- }
-
- configMapMode := configMapSource.DefaultMode
- if configMapMode != nil && (*configMapMode > 0777 || *configMapMode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *configMapMode, fileModeErrorMsg))
- }
-
- itemsPath := fldPath.Child("items")
- for i, kp := range configMapSource.Items {
- itemPath := itemsPath.Index(i)
- allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
- }
- return allErrs
-}
-
-func validateKeyToPath(kp *core.KeyToPath, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(kp.Key) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
- }
- if len(kp.Path) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- }
- allErrs = append(allErrs, validateLocalNonReservedPath(kp.Path, fldPath.Child("path"))...)
- if kp.Mode != nil && (*kp.Mode > 0777 || *kp.Mode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("mode"), *kp.Mode, fileModeErrorMsg))
- }
-
- return allErrs
-}
-
-func validatePersistentClaimVolumeSource(claim *core.PersistentVolumeClaimVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(claim.ClaimName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("claimName"), ""))
- }
- return allErrs
-}
-
-func validateNFSVolumeSource(nfs *core.NFSVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(nfs.Server) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("server"), ""))
- }
- if len(nfs.Path) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- }
- if !path.IsAbs(nfs.Path) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("path"), nfs.Path, "must be an absolute path"))
- }
- return allErrs
-}
-
-func validateQuobyteVolumeSource(quobyte *core.QuobyteVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(quobyte.Registry) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("registry"), "must be a host:port pair or multiple pairs separated by commas"))
- } else {
- for _, hostPortPair := range strings.Split(quobyte.Registry, ",") {
- if _, _, err := net.SplitHostPort(hostPortPair); err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("registry"), quobyte.Registry, "must be a host:port pair or multiple pairs separated by commas"))
- }
- }
- }
-
- if len(quobyte.Volume) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volume"), ""))
- }
- return allErrs
-}
-
-func validateGlusterfsVolumeSource(glusterfs *core.GlusterfsVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(glusterfs.EndpointsName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("endpoints"), ""))
- }
- if len(glusterfs.Path) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- }
- return allErrs
-}
-
-func validateFlockerVolumeSource(flocker *core.FlockerVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(flocker.DatasetName) == 0 && len(flocker.DatasetUUID) == 0 {
- //TODO: consider adding a RequiredOneOf() error for this and similar cases
- allErrs = append(allErrs, field.Required(fldPath, "one of datasetName and datasetUUID is required"))
- }
- if len(flocker.DatasetName) != 0 && len(flocker.DatasetUUID) != 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, "resource", "datasetName and datasetUUID can not be specified simultaneously"))
- }
- if strings.Contains(flocker.DatasetName, "/") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("datasetName"), flocker.DatasetName, "must not contain '/'"))
- }
- return allErrs
-}
-
-var validVolumeDownwardAPIFieldPathExpressions = sets.NewString(
- "metadata.name",
- "metadata.namespace",
- "metadata.labels",
- "metadata.annotations",
- "metadata.uid")
-
-func validateDownwardAPIVolumeFile(file *core.DownwardAPIVolumeFile, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if len(file.Path) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- }
- allErrs = append(allErrs, validateLocalNonReservedPath(file.Path, fldPath.Child("path"))...)
- if file.FieldRef != nil {
- allErrs = append(allErrs, validateObjectFieldSelector(file.FieldRef, &validVolumeDownwardAPIFieldPathExpressions, fldPath.Child("fieldRef"))...)
- if file.ResourceFieldRef != nil {
- allErrs = append(allErrs, field.Invalid(fldPath, "resource", "fieldRef and resourceFieldRef can not be specified simultaneously"))
- }
- } else if file.ResourceFieldRef != nil {
- allErrs = append(allErrs, validateContainerResourceFieldSelector(file.ResourceFieldRef, &validContainerResourceFieldPathExpressions, fldPath.Child("resourceFieldRef"), true)...)
- } else {
- allErrs = append(allErrs, field.Required(fldPath, "one of fieldRef and resourceFieldRef is required"))
- }
- if file.Mode != nil && (*file.Mode > 0777 || *file.Mode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("mode"), *file.Mode, fileModeErrorMsg))
- }
-
- return allErrs
-}
-
-func validateDownwardAPIVolumeSource(downwardAPIVolume *core.DownwardAPIVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- downwardAPIMode := downwardAPIVolume.DefaultMode
- if downwardAPIMode != nil && (*downwardAPIMode > 0777 || *downwardAPIMode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *downwardAPIMode, fileModeErrorMsg))
- }
-
- for _, file := range downwardAPIVolume.Items {
- allErrs = append(allErrs, validateDownwardAPIVolumeFile(&file, fldPath)...)
- }
- return allErrs
-}
-
-func validateProjectionSources(projection *core.ProjectedVolumeSource, projectionMode *int32, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allPaths := sets.String{}
-
- for _, source := range projection.Sources {
- numSources := 0
- if source.Secret != nil {
- if numSources > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("secret"), "may not specify more than 1 volume type"))
- } else {
- numSources++
- if len(source.Secret.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- }
- itemsPath := fldPath.Child("items")
- for i, kp := range source.Secret.Items {
- itemPath := itemsPath.Index(i)
- allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
- if len(kp.Path) > 0 {
- curPath := kp.Path
- if !allPaths.Has(curPath) {
- allPaths.Insert(curPath)
- } else {
- allErrs = append(allErrs, field.Invalid(fldPath, source.Secret.Name, "conflicting duplicate paths"))
- }
- }
- }
- }
- }
- if source.ConfigMap != nil {
- if numSources > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("configMap"), "may not specify more than 1 volume type"))
- } else {
- numSources++
- if len(source.ConfigMap.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- }
- itemsPath := fldPath.Child("items")
- for i, kp := range source.ConfigMap.Items {
- itemPath := itemsPath.Index(i)
- allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
- if len(kp.Path) > 0 {
- curPath := kp.Path
- if !allPaths.Has(curPath) {
- allPaths.Insert(curPath)
- } else {
- allErrs = append(allErrs, field.Invalid(fldPath, source.ConfigMap.Name, "conflicting duplicate paths"))
- }
-
- }
- }
- }
- }
- if source.DownwardAPI != nil {
- if numSources > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("downwardAPI"), "may not specify more than 1 volume type"))
- } else {
- numSources++
- for _, file := range source.DownwardAPI.Items {
- allErrs = append(allErrs, validateDownwardAPIVolumeFile(&file, fldPath.Child("downwardAPI"))...)
- if len(file.Path) > 0 {
- curPath := file.Path
- if !allPaths.Has(curPath) {
- allPaths.Insert(curPath)
- } else {
- allErrs = append(allErrs, field.Invalid(fldPath, curPath, "conflicting duplicate paths"))
- }
-
- }
- }
- }
- }
- }
- return allErrs
-}
-
-func validateProjectedVolumeSource(projection *core.ProjectedVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- projectionMode := projection.DefaultMode
- if projectionMode != nil && (*projectionMode > 0777 || *projectionMode < 0) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *projectionMode, fileModeErrorMsg))
- }
-
- allErrs = append(allErrs, validateProjectionSources(projection, projectionMode, fldPath)...)
- return allErrs
-}
-
-var supportedHostPathTypes = sets.NewString(
- string(core.HostPathUnset),
- string(core.HostPathDirectoryOrCreate),
- string(core.HostPathDirectory),
- string(core.HostPathFileOrCreate),
- string(core.HostPathFile),
- string(core.HostPathSocket),
- string(core.HostPathCharDev),
- string(core.HostPathBlockDev))
-
-func validateHostPathType(hostPathType *core.HostPathType, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if hostPathType != nil && !supportedHostPathTypes.Has(string(*hostPathType)) {
- allErrs = append(allErrs, field.NotSupported(fldPath, hostPathType, supportedHostPathTypes.List()))
- }
-
- return allErrs
-}
-
-// This validate will make sure targetPath:
-// 1. is not abs path
-// 2. does not have any element which is ".."
-func validateLocalDescendingPath(targetPath string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if path.IsAbs(targetPath) {
- allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must be a relative path"))
- }
-
- allErrs = append(allErrs, validatePathNoBacksteps(targetPath, fldPath)...)
-
- return allErrs
-}
-
-// validatePathNoBacksteps makes sure the targetPath does not have any `..` path elements when split
-//
-// This assumes the OS of the apiserver and the nodes are the same. The same check should be done
-// on the node to ensure there are no backsteps.
-func validatePathNoBacksteps(targetPath string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- parts := strings.Split(filepath.ToSlash(targetPath), "/")
- for _, item := range parts {
- if item == ".." {
- allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must not contain '..'"))
- break // even for `../../..`, one error is sufficient to make the point
- }
- }
- return allErrs
-}
-
-// validateMountPropagation verifies that MountPropagation field is valid and
-// allowed for given container.
-func validateMountPropagation(mountPropagation *core.MountPropagationMode, container *core.Container, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if mountPropagation == nil {
- return allErrs
- }
- if !utilfeature.DefaultFeatureGate.Enabled(features.MountPropagation) {
- allErrs = append(allErrs, field.Forbidden(fldPath, "mount propagation is disabled by feature-gate"))
- return allErrs
- }
-
- supportedMountPropagations := sets.NewString(string(core.MountPropagationBidirectional), string(core.MountPropagationHostToContainer))
- if !supportedMountPropagations.Has(string(*mountPropagation)) {
- allErrs = append(allErrs, field.NotSupported(fldPath, *mountPropagation, supportedMountPropagations.List()))
- }
-
- if container == nil {
- // The container is not available yet, e.g. during validation of
- // PodPreset. Stop validation now, Pod validation will refuse final
- // Pods with Bidirectional propagation in non-privileged containers.
- return allErrs
- }
-
- privileged := container.SecurityContext != nil && container.SecurityContext.Privileged != nil && *container.SecurityContext.Privileged
- if *mountPropagation == core.MountPropagationBidirectional && !privileged {
- allErrs = append(allErrs, field.Forbidden(fldPath, "Bidirectional mount propagation is available only to privileged containers"))
- }
- return allErrs
-}
-
-// This validate will make sure targetPath:
-// 1. is not abs path
-// 2. does not contain any '..' elements
-// 3. does not start with '..'
-func validateLocalNonReservedPath(targetPath string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, validateLocalDescendingPath(targetPath, fldPath)...)
- // Don't report this error if the check for .. elements already caught it.
- if strings.HasPrefix(targetPath, "..") && !strings.HasPrefix(targetPath, "../") {
- allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must not start with '..'"))
- }
- return allErrs
-}
-
-func validateRBDVolumeSource(rbd *core.RBDVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(rbd.CephMonitors) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
- }
- if len(rbd.RBDImage) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("image"), ""))
- }
- return allErrs
-}
-
-func validateRBDPersistentVolumeSource(rbd *core.RBDPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(rbd.CephMonitors) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
- }
- if len(rbd.RBDImage) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("image"), ""))
- }
- return allErrs
-}
-
-func validateCinderVolumeSource(cd *core.CinderVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(cd.VolumeID) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
- }
- return allErrs
-}
-
-func validateCephFSVolumeSource(cephfs *core.CephFSVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(cephfs.Monitors) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
- }
- return allErrs
-}
-
-func validateCephFSPersistentVolumeSource(cephfs *core.CephFSPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(cephfs.Monitors) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
- }
- return allErrs
-}
-
-func validateFlexVolumeSource(fv *core.FlexVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(fv.Driver) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("driver"), ""))
- }
-
- // Make sure user-specified options don't use kubernetes namespaces
- for k := range fv.Options {
- namespace := k
- if parts := strings.SplitN(k, "/", 2); len(parts) == 2 {
- namespace = parts[0]
- }
- normalized := "." + strings.ToLower(namespace)
- if strings.HasSuffix(normalized, ".kubernetes.io") || strings.HasSuffix(normalized, ".k8s.io") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("options").Key(k), k, "kubernetes.io and k8s.io namespaces are reserved"))
- }
- }
-
- return allErrs
-}
-
-func validateFlexPersistentVolumeSource(fv *core.FlexPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(fv.Driver) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("driver"), ""))
- }
-
- // Make sure user-specified options don't use kubernetes namespaces
- for k := range fv.Options {
- namespace := k
- if parts := strings.SplitN(k, "/", 2); len(parts) == 2 {
- namespace = parts[0]
- }
- normalized := "." + strings.ToLower(namespace)
- if strings.HasSuffix(normalized, ".kubernetes.io") || strings.HasSuffix(normalized, ".k8s.io") {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("options").Key(k), k, "kubernetes.io and k8s.io namespaces are reserved"))
- }
- }
-
- return allErrs
-}
-
-func validateAzureFile(azure *core.AzureFileVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if azure.SecretName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
- }
- if azure.ShareName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("shareName"), ""))
- }
- return allErrs
-}
-
-func validateAzureFilePV(azure *core.AzureFilePersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if azure.SecretName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
- }
- if azure.ShareName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("shareName"), ""))
- }
- if azure.SecretNamespace != nil {
- if len(*azure.SecretNamespace) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretNamespace"), ""))
- }
- }
- return allErrs
-}
-
-func validateAzureDisk(azure *core.AzureDiskVolumeSource, fldPath *field.Path) field.ErrorList {
- var supportedCachingModes = sets.NewString(string(core.AzureDataDiskCachingNone), string(core.AzureDataDiskCachingReadOnly), string(core.AzureDataDiskCachingReadWrite))
- var supportedDiskKinds = sets.NewString(string(core.AzureSharedBlobDisk), string(core.AzureDedicatedBlobDisk), string(core.AzureManagedDisk))
-
- diskUriSupportedManaged := []string{"/subscriptions/{sub-id}/resourcegroups/{group-name}/providers/microsoft.compute/disks/{disk-id}"}
- diskUriSupportedblob := []string{"https://{account-name}.blob.core.windows.net/{container-name}/{disk-name}.vhd"}
-
- allErrs := field.ErrorList{}
- if azure.DiskName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("diskName"), ""))
- }
-
- if azure.DataDiskURI == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("diskURI"), ""))
- }
-
- if azure.CachingMode != nil && !supportedCachingModes.Has(string(*azure.CachingMode)) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("cachingMode"), *azure.CachingMode, supportedCachingModes.List()))
- }
-
- if azure.Kind != nil && !supportedDiskKinds.Has(string(*azure.Kind)) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("kind"), *azure.Kind, supportedDiskKinds.List()))
- }
-
- // validate that DiskUri is the correct format
- if azure.Kind != nil && *azure.Kind == core.AzureManagedDisk && strings.Index(azure.DataDiskURI, "/subscriptions/") != 0 {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("diskURI"), azure.DataDiskURI, diskUriSupportedManaged))
- }
-
- if azure.Kind != nil && *azure.Kind != core.AzureManagedDisk && strings.Index(azure.DataDiskURI, "https://") != 0 {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("diskURI"), azure.DataDiskURI, diskUriSupportedblob))
- }
-
- return allErrs
-}
-
-func validateVsphereVolumeSource(cd *core.VsphereVirtualDiskVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(cd.VolumePath) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumePath"), ""))
- }
- return allErrs
-}
-
-func validatePhotonPersistentDiskVolumeSource(cd *core.PhotonPersistentDiskVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(cd.PdID) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("pdID"), ""))
- }
- return allErrs
-}
-
-func validatePortworxVolumeSource(pwx *core.PortworxVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(pwx.VolumeID) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
- }
- return allErrs
-}
-
-func validateScaleIOVolumeSource(sio *core.ScaleIOVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if sio.Gateway == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("gateway"), ""))
- }
- if sio.System == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("system"), ""))
- }
- if sio.VolumeName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
- }
- return allErrs
-}
-
-func validateScaleIOPersistentVolumeSource(sio *core.ScaleIOPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if sio.Gateway == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("gateway"), ""))
- }
- if sio.System == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("system"), ""))
- }
- if sio.VolumeName == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
- }
- return allErrs
-}
-
-func validateLocalVolumeSource(ls *core.LocalVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if ls.Path == "" {
- allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
- return allErrs
- }
-
- if !path.IsAbs(ls.Path) {
- allErrs = append(allErrs, field.Invalid(fldPath, ls.Path, "must be an absolute path"))
- }
- allErrs = append(allErrs, validatePathNoBacksteps(ls.Path, fldPath.Child("path"))...)
- return allErrs
-}
-
-func validateStorageOSVolumeSource(storageos *core.StorageOSVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(storageos.VolumeName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeName, fldPath.Child("volumeName"))...)
- }
- if len(storageos.VolumeNamespace) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeNamespace, fldPath.Child("volumeNamespace"))...)
- }
- if storageos.SecretRef != nil {
- if len(storageos.SecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
- }
- }
- return allErrs
-}
-
-func validateStorageOSPersistentVolumeSource(storageos *core.StorageOSPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(storageos.VolumeName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeName, fldPath.Child("volumeName"))...)
- }
- if len(storageos.VolumeNamespace) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeNamespace, fldPath.Child("volumeNamespace"))...)
- }
- if storageos.SecretRef != nil {
- if len(storageos.SecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
- }
- if len(storageos.SecretRef.Namespace) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "namespace"), ""))
- }
- }
- return allErrs
-}
-
-func validateCSIPersistentVolumeSource(csi *core.CSIPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if !utilfeature.DefaultFeatureGate.Enabled(features.CSIPersistentVolume) {
- allErrs = append(allErrs, field.Forbidden(fldPath, "CSIPersistentVolume disabled by feature-gate"))
- }
-
- if len(csi.Driver) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("driver"), ""))
- }
-
- if len(csi.Driver) > 63 {
- allErrs = append(allErrs, field.TooLong(fldPath.Child("driver"), csi.Driver, 63))
- }
-
- if !csiDriverNameRexp.MatchString(csi.Driver) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("driver"), csi.Driver, validation.RegexError(csiDriverNameRexpErrMsg, csiDriverNameRexpFmt, "csi-hostpath")))
- }
-
- if len(csi.VolumeHandle) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("volumeHandle"), ""))
- }
-
- if csi.ControllerPublishSecretRef != nil {
- if len(csi.ControllerPublishSecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("controllerPublishSecretRef", "name"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerPublishSecretRef.Name, fldPath.Child("name"))...)
- }
- if len(csi.ControllerPublishSecretRef.Namespace) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("controllerPublishSecretRef", "namespace"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerPublishSecretRef.Namespace, fldPath.Child("namespace"))...)
- }
- }
-
- if csi.NodePublishSecretRef != nil {
- if len(csi.NodePublishSecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("nodePublishSecretRef ", "name"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.NodePublishSecretRef.Name, fldPath.Child("name"))...)
- }
- if len(csi.NodePublishSecretRef.Namespace) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("nodePublishSecretRef ", "namespace"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.NodePublishSecretRef.Namespace, fldPath.Child("namespace"))...)
- }
- }
-
- if csi.NodeStageSecretRef != nil {
- if len(csi.NodeStageSecretRef.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("nodeStageSecretRef", "name"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.NodeStageSecretRef.Name, fldPath.Child("name"))...)
- }
- if len(csi.NodeStageSecretRef.Namespace) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("nodeStageSecretRef", "namespace"), ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(csi.NodeStageSecretRef.Namespace, fldPath.Child("namespace"))...)
- }
- }
-
- return allErrs
-}
-
-// ValidatePersistentVolumeName checks that a name is appropriate for a
-// PersistentVolumeName object.
-var ValidatePersistentVolumeName = NameIsDNSSubdomain
-
-var supportedAccessModes = sets.NewString(string(core.ReadWriteOnce), string(core.ReadOnlyMany), string(core.ReadWriteMany))
-
-var supportedReclaimPolicy = sets.NewString(string(core.PersistentVolumeReclaimDelete), string(core.PersistentVolumeReclaimRecycle), string(core.PersistentVolumeReclaimRetain))
-
-var supportedVolumeModes = sets.NewString(string(core.PersistentVolumeBlock), string(core.PersistentVolumeFilesystem))
-
-func ValidatePersistentVolume(pv *core.PersistentVolume) field.ErrorList {
- metaPath := field.NewPath("metadata")
- allErrs := ValidateObjectMeta(&pv.ObjectMeta, false, ValidatePersistentVolumeName, metaPath)
-
- specPath := field.NewPath("spec")
- if len(pv.Spec.AccessModes) == 0 {
- allErrs = append(allErrs, field.Required(specPath.Child("accessModes"), ""))
- }
- for _, mode := range pv.Spec.AccessModes {
- if !supportedAccessModes.Has(string(mode)) {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("accessModes"), mode, supportedAccessModes.List()))
- }
- }
-
- if len(pv.Spec.Capacity) == 0 {
- allErrs = append(allErrs, field.Required(specPath.Child("capacity"), ""))
- }
-
- if _, ok := pv.Spec.Capacity[core.ResourceStorage]; !ok || len(pv.Spec.Capacity) > 1 {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("capacity"), pv.Spec.Capacity, []string{string(core.ResourceStorage)}))
- }
- capPath := specPath.Child("capacity")
- for r, qty := range pv.Spec.Capacity {
- allErrs = append(allErrs, validateBasicResource(qty, capPath.Key(string(r)))...)
- allErrs = append(allErrs, ValidatePositiveQuantityValue(qty, capPath.Key(string(r)))...)
- }
- if len(string(pv.Spec.PersistentVolumeReclaimPolicy)) > 0 {
- if !supportedReclaimPolicy.Has(string(pv.Spec.PersistentVolumeReclaimPolicy)) {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("persistentVolumeReclaimPolicy"), pv.Spec.PersistentVolumeReclaimPolicy, supportedReclaimPolicy.List()))
- }
- }
-
- nodeAffinitySpecified, errs := validateStorageNodeAffinityAnnotation(pv.ObjectMeta.Annotations, metaPath.Child("annotations"))
- allErrs = append(allErrs, errs...)
-
- volumeNodeAffinitySpecified, errs := validateVolumeNodeAffinity(pv.Spec.NodeAffinity, specPath.Child("nodeAffinity"))
- allErrs = append(allErrs, errs...)
-
- if nodeAffinitySpecified && volumeNodeAffinitySpecified {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("nodeAffinity"), "may not specify both alpha nodeAffinity annotation and nodeAffinity field"))
- }
-
- nodeAffinitySpecified = nodeAffinitySpecified || volumeNodeAffinitySpecified
-
- numVolumes := 0
- if pv.Spec.HostPath != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("hostPath"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateHostPathVolumeSource(pv.Spec.HostPath, specPath.Child("hostPath"))...)
- }
- }
- if pv.Spec.GCEPersistentDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("gcePersistentDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateGCEPersistentDiskVolumeSource(pv.Spec.GCEPersistentDisk, specPath.Child("persistentDisk"))...)
- }
- }
- if pv.Spec.AWSElasticBlockStore != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("awsElasticBlockStore"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAWSElasticBlockStoreVolumeSource(pv.Spec.AWSElasticBlockStore, specPath.Child("awsElasticBlockStore"))...)
- }
- }
- if pv.Spec.Glusterfs != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("glusterfs"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateGlusterfsVolumeSource(pv.Spec.Glusterfs, specPath.Child("glusterfs"))...)
- }
- }
- if pv.Spec.Flocker != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("flocker"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateFlockerVolumeSource(pv.Spec.Flocker, specPath.Child("flocker"))...)
- }
- }
- if pv.Spec.NFS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("nfs"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateNFSVolumeSource(pv.Spec.NFS, specPath.Child("nfs"))...)
- }
- }
- if pv.Spec.RBD != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("rbd"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateRBDPersistentVolumeSource(pv.Spec.RBD, specPath.Child("rbd"))...)
- }
- }
- if pv.Spec.Quobyte != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("quobyte"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateQuobyteVolumeSource(pv.Spec.Quobyte, specPath.Child("quobyte"))...)
- }
- }
- if pv.Spec.CephFS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("cephFS"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateCephFSPersistentVolumeSource(pv.Spec.CephFS, specPath.Child("cephfs"))...)
- }
- }
- if pv.Spec.ISCSI != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("iscsi"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateISCSIPersistentVolumeSource(pv.Spec.ISCSI, specPath.Child("iscsi"))...)
- }
- if pv.Spec.ISCSI.InitiatorName != nil && len(pv.ObjectMeta.Name+":"+pv.Spec.ISCSI.TargetPortal) > 64 {
- tooLongErr := "Total length of <volume name>:<iscsi.targetPortal> must be under 64 characters if iscsi.initiatorName is specified."
- allErrs = append(allErrs, field.Invalid(metaPath.Child("name"), pv.ObjectMeta.Name, tooLongErr))
- }
- }
- if pv.Spec.Cinder != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("cinder"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateCinderVolumeSource(pv.Spec.Cinder, specPath.Child("cinder"))...)
- }
- }
- if pv.Spec.FC != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("fc"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateFCVolumeSource(pv.Spec.FC, specPath.Child("fc"))...)
- }
- }
- if pv.Spec.FlexVolume != nil {
- numVolumes++
- allErrs = append(allErrs, validateFlexPersistentVolumeSource(pv.Spec.FlexVolume, specPath.Child("flexVolume"))...)
- }
- if pv.Spec.AzureFile != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("azureFile"), "may not specify more than 1 volume type"))
-
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAzureFilePV(pv.Spec.AzureFile, specPath.Child("azureFile"))...)
- }
- }
-
- if pv.Spec.VsphereVolume != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("vsphereVolume"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateVsphereVolumeSource(pv.Spec.VsphereVolume, specPath.Child("vsphereVolume"))...)
- }
- }
- if pv.Spec.PhotonPersistentDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("photonPersistentDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validatePhotonPersistentDiskVolumeSource(pv.Spec.PhotonPersistentDisk, specPath.Child("photonPersistentDisk"))...)
- }
- }
- if pv.Spec.PortworxVolume != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("portworxVolume"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validatePortworxVolumeSource(pv.Spec.PortworxVolume, specPath.Child("portworxVolume"))...)
- }
- }
- if pv.Spec.AzureDisk != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("azureDisk"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateAzureDisk(pv.Spec.AzureDisk, specPath.Child("azureDisk"))...)
- }
- }
- if pv.Spec.ScaleIO != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("scaleIO"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateScaleIOPersistentVolumeSource(pv.Spec.ScaleIO, specPath.Child("scaleIO"))...)
- }
- }
- if pv.Spec.Local != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("local"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- if !utilfeature.DefaultFeatureGate.Enabled(features.PersistentLocalVolumes) {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("local"), "Local volumes are disabled by feature-gate"))
- }
- allErrs = append(allErrs, validateLocalVolumeSource(pv.Spec.Local, specPath.Child("local"))...)
-
- // NodeAffinity is required
- if !nodeAffinitySpecified {
- allErrs = append(allErrs, field.Required(metaPath.Child("annotations"), "Local volume requires node affinity"))
- }
- }
- }
- if pv.Spec.StorageOS != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("storageos"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateStorageOSPersistentVolumeSource(pv.Spec.StorageOS, specPath.Child("storageos"))...)
- }
- }
-
- if pv.Spec.CSI != nil {
- if numVolumes > 0 {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("csi"), "may not specify more than 1 volume type"))
- } else {
- numVolumes++
- allErrs = append(allErrs, validateCSIPersistentVolumeSource(pv.Spec.CSI, specPath.Child("csi"))...)
- }
- }
-
- if numVolumes == 0 {
- allErrs = append(allErrs, field.Required(specPath, "must specify a volume type"))
- }
-
- // do not allow hostPath mounts of '/' to have a 'recycle' reclaim policy
- if pv.Spec.HostPath != nil && path.Clean(pv.Spec.HostPath.Path) == "/" && pv.Spec.PersistentVolumeReclaimPolicy == core.PersistentVolumeReclaimRecycle {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("persistentVolumeReclaimPolicy"), "may not be 'recycle' for a hostPath mount of '/'"))
- }
-
- if len(pv.Spec.StorageClassName) > 0 {
- for _, msg := range ValidateClassName(pv.Spec.StorageClassName, false) {
- allErrs = append(allErrs, field.Invalid(specPath.Child("storageClassName"), pv.Spec.StorageClassName, msg))
- }
- }
- if pv.Spec.VolumeMode != nil && !utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("volumeMode"), "PersistentVolume volumeMode is disabled by feature-gate"))
- } else if pv.Spec.VolumeMode != nil && !supportedVolumeModes.Has(string(*pv.Spec.VolumeMode)) {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("volumeMode"), *pv.Spec.VolumeMode, supportedVolumeModes.List()))
- }
- return allErrs
-}
-
-// ValidatePersistentVolumeUpdate tests to see if the update is legal for an end user to make.
-// newPv is updated with fields that cannot be changed.
-func ValidatePersistentVolumeUpdate(newPv, oldPv *core.PersistentVolume) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = ValidatePersistentVolume(newPv)
-
- // PersistentVolumeSource should be immutable after creation.
- if !apiequality.Semantic.DeepEqual(newPv.Spec.PersistentVolumeSource, oldPv.Spec.PersistentVolumeSource) {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "persistentvolumesource"), "is immutable after creation"))
- }
-
- newPv.Status = oldPv.Status
-
- if utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- allErrs = append(allErrs, ValidateImmutableField(newPv.Spec.VolumeMode, oldPv.Spec.VolumeMode, field.NewPath("volumeMode"))...)
- }
-
- if utilfeature.DefaultFeatureGate.Enabled(features.VolumeScheduling) {
- // Allow setting NodeAffinity if oldPv NodeAffinity was not set
- if oldPv.Spec.NodeAffinity != nil {
- allErrs = append(allErrs, ValidateImmutableField(newPv.Spec.NodeAffinity, oldPv.Spec.NodeAffinity, field.NewPath("nodeAffinity"))...)
- }
- }
-
- return allErrs
-}
-
-// ValidatePersistentVolumeStatusUpdate tests to see if the status update is legal for an end user to make.
-// newPv is updated with fields that cannot be changed.
-func ValidatePersistentVolumeStatusUpdate(newPv, oldPv *core.PersistentVolume) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newPv.ObjectMeta, &oldPv.ObjectMeta, field.NewPath("metadata"))
- if len(newPv.ResourceVersion) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
- }
- newPv.Spec = oldPv.Spec
- return allErrs
-}
-
-// ValidatePersistentVolumeClaim validates a PersistentVolumeClaim
-func ValidatePersistentVolumeClaim(pvc *core.PersistentVolumeClaim) field.ErrorList {
- allErrs := ValidateObjectMeta(&pvc.ObjectMeta, true, ValidatePersistentVolumeName, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidatePersistentVolumeClaimSpec(&pvc.Spec, field.NewPath("spec"))...)
- return allErrs
-}
-
-// ValidatePersistentVolumeClaimSpec validates a PersistentVolumeClaimSpec
-func ValidatePersistentVolumeClaimSpec(spec *core.PersistentVolumeClaimSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(spec.AccessModes) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("accessModes"), "at least 1 access mode is required"))
- }
- if spec.Selector != nil {
- allErrs = append(allErrs, unversionedvalidation.ValidateLabelSelector(spec.Selector, fldPath.Child("selector"))...)
- }
- for _, mode := range spec.AccessModes {
- if mode != core.ReadWriteOnce && mode != core.ReadOnlyMany && mode != core.ReadWriteMany {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("accessModes"), mode, supportedAccessModes.List()))
- }
- }
- storageValue, ok := spec.Resources.Requests[core.ResourceStorage]
- if !ok {
- allErrs = append(allErrs, field.Required(fldPath.Child("resources").Key(string(core.ResourceStorage)), ""))
- } else {
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(core.ResourceStorage), storageValue, fldPath.Child("resources").Key(string(core.ResourceStorage)))...)
- allErrs = append(allErrs, ValidatePositiveQuantityValue(storageValue, fldPath.Child("resources").Key(string(core.ResourceStorage)))...)
- }
-
- if spec.StorageClassName != nil && len(*spec.StorageClassName) > 0 {
- for _, msg := range ValidateClassName(*spec.StorageClassName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("storageClassName"), *spec.StorageClassName, msg))
- }
- }
- if spec.VolumeMode != nil && !utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("volumeMode"), "PersistentVolumeClaim volumeMode is disabled by feature-gate"))
- } else if spec.VolumeMode != nil && !supportedVolumeModes.Has(string(*spec.VolumeMode)) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("volumeMode"), *spec.VolumeMode, supportedVolumeModes.List()))
- }
- return allErrs
-}
-
-// ValidatePersistentVolumeClaimUpdate validates an update to a PersistentVolumeClaim
-func ValidatePersistentVolumeClaimUpdate(newPvc, oldPvc *core.PersistentVolumeClaim) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newPvc.ObjectMeta, &oldPvc.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidatePersistentVolumeClaim(newPvc)...)
- newPvcClone := newPvc.DeepCopy()
- oldPvcClone := oldPvc.DeepCopy()
-
- // PVController needs to update PVC.Spec w/ VolumeName.
- // Claims are immutable in order to enforce quota, range limits, etc. without gaming the system.
- if len(oldPvc.Spec.VolumeName) == 0 {
- // volumeName changes are allowed once.
- oldPvcClone.Spec.VolumeName = newPvcClone.Spec.VolumeName
- }
-
- if validateStorageClassUpgrade(oldPvcClone.Annotations, newPvcClone.Annotations,
- oldPvcClone.Spec.StorageClassName, newPvcClone.Spec.StorageClassName) {
- newPvcClone.Spec.StorageClassName = nil
- metav1.SetMetaDataAnnotation(&newPvcClone.ObjectMeta, core.BetaStorageClassAnnotation, oldPvcClone.Annotations[core.BetaStorageClassAnnotation])
- } else {
- // storageclass annotation should be immutable after creation
- // TODO: remove Beta when no longer needed
- allErrs = append(allErrs, ValidateImmutableAnnotation(newPvc.ObjectMeta.Annotations[v1.BetaStorageClassAnnotation], oldPvc.ObjectMeta.Annotations[v1.BetaStorageClassAnnotation], v1.BetaStorageClassAnnotation, field.NewPath("metadata"))...)
- }
-
- if utilfeature.DefaultFeatureGate.Enabled(features.ExpandPersistentVolumes) {
- // lets make sure storage values are same.
- if newPvc.Status.Phase == core.ClaimBound && newPvcClone.Spec.Resources.Requests != nil {
- newPvcClone.Spec.Resources.Requests["storage"] = oldPvc.Spec.Resources.Requests["storage"]
- }
-
- oldSize := oldPvc.Spec.Resources.Requests["storage"]
- newSize := newPvc.Spec.Resources.Requests["storage"]
-
- if !apiequality.Semantic.DeepEqual(newPvcClone.Spec, oldPvcClone.Spec) {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec"), "is immutable after creation except resources.requests for bound claims"))
- }
- if newSize.Cmp(oldSize) < 0 {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "resources", "requests", "storage"), "field can not be less than previous value"))
- }
-
- } else {
- // changes to Spec are not allowed, but updates to label/and some annotations are OK.
- // no-op updates pass validation.
- if !apiequality.Semantic.DeepEqual(newPvcClone.Spec, oldPvcClone.Spec) {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec"), "field is immutable after creation"))
- }
- }
-
- if utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- allErrs = append(allErrs, ValidateImmutableField(newPvc.Spec.VolumeMode, oldPvc.Spec.VolumeMode, field.NewPath("volumeMode"))...)
- }
- return allErrs
-}
-
-// Provide an upgrade path from PVC with storage class specified in beta
-// annotation to storage class specified in attribute. We allow update of
-// StorageClassName only if following four conditions are met at the same time:
-// 1. The old pvc's StorageClassAnnotation is set
-// 2. The old pvc's StorageClassName is not set
-// 3. The new pvc's StorageClassName is set and equal to the old value in annotation
-// 4. If the new pvc's StorageClassAnnotation is set,it must be equal to the old pv/pvc's StorageClassAnnotation
-func validateStorageClassUpgrade(oldAnnotations, newAnnotations map[string]string, oldScName, newScName *string) bool {
- oldSc, oldAnnotationExist := oldAnnotations[core.BetaStorageClassAnnotation]
- newScInAnnotation, newAnnotationExist := newAnnotations[core.BetaStorageClassAnnotation]
- return oldAnnotationExist /* condition 1 */ &&
- oldScName == nil /* condition 2*/ &&
- (newScName != nil && *newScName == oldSc) /* condition 3 */ &&
- (!newAnnotationExist || newScInAnnotation == oldSc) /* condition 4 */
-}
-
-// ValidatePersistentVolumeClaimStatusUpdate validates an update to status of a PersistentVolumeClaim
-func ValidatePersistentVolumeClaimStatusUpdate(newPvc, oldPvc *core.PersistentVolumeClaim) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newPvc.ObjectMeta, &oldPvc.ObjectMeta, field.NewPath("metadata"))
- if len(newPvc.ResourceVersion) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
- }
- if len(newPvc.Spec.AccessModes) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("Spec", "accessModes"), ""))
- }
- if !utilfeature.DefaultFeatureGate.Enabled(features.ExpandPersistentVolumes) && len(newPvc.Status.Conditions) > 0 {
- conditionPath := field.NewPath("status", "conditions")
- allErrs = append(allErrs, field.Forbidden(conditionPath, "invalid field"))
- }
- capPath := field.NewPath("status", "capacity")
- for r, qty := range newPvc.Status.Capacity {
- allErrs = append(allErrs, validateBasicResource(qty, capPath.Key(string(r)))...)
- }
- newPvc.Spec = oldPvc.Spec
- return allErrs
-}
-
-var supportedPortProtocols = sets.NewString(string(core.ProtocolTCP), string(core.ProtocolUDP))
-
-func validateContainerPorts(ports []core.ContainerPort, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- allNames := sets.String{}
- for i, port := range ports {
- idxPath := fldPath.Index(i)
- if len(port.Name) > 0 {
- if msgs := validation.IsValidPortName(port.Name); len(msgs) != 0 {
- for i = range msgs {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), port.Name, msgs[i]))
- }
- } else if allNames.Has(port.Name) {
- allErrs = append(allErrs, field.Duplicate(idxPath.Child("name"), port.Name))
- } else {
- allNames.Insert(port.Name)
- }
- }
- if port.ContainerPort == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("containerPort"), ""))
- } else {
- for _, msg := range validation.IsValidPortNum(int(port.ContainerPort)) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("containerPort"), port.ContainerPort, msg))
- }
- }
- if port.HostPort != 0 {
- for _, msg := range validation.IsValidPortNum(int(port.HostPort)) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("hostPort"), port.HostPort, msg))
- }
- }
- if len(port.Protocol) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("protocol"), ""))
- } else if !supportedPortProtocols.Has(string(port.Protocol)) {
- allErrs = append(allErrs, field.NotSupported(idxPath.Child("protocol"), port.Protocol, supportedPortProtocols.List()))
- }
- }
- return allErrs
-}
-
-// ValidateEnv validates env vars
-func ValidateEnv(vars []core.EnvVar, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- for i, ev := range vars {
- idxPath := fldPath.Index(i)
- if len(ev.Name) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
- } else {
- for _, msg := range validation.IsEnvVarName(ev.Name) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), ev.Name, msg))
- }
- }
- allErrs = append(allErrs, validateEnvVarValueFrom(ev, idxPath.Child("valueFrom"))...)
- }
- return allErrs
-}
-
-var validEnvDownwardAPIFieldPathExpressions = sets.NewString(
- "metadata.name",
- "metadata.namespace",
- "metadata.uid",
- "spec.nodeName",
- "spec.serviceAccountName",
- "status.hostIP",
- "status.podIP")
-var validContainerResourceFieldPathExpressions = sets.NewString("limits.cpu", "limits.memory", "limits.ephemeral-storage", "requests.cpu", "requests.memory", "requests.ephemeral-storage")
-
-func validateEnvVarValueFrom(ev core.EnvVar, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if ev.ValueFrom == nil {
- return allErrs
- }
-
- numSources := 0
-
- if ev.ValueFrom.FieldRef != nil {
- numSources++
- allErrs = append(allErrs, validateObjectFieldSelector(ev.ValueFrom.FieldRef, &validEnvDownwardAPIFieldPathExpressions, fldPath.Child("fieldRef"))...)
- }
- if ev.ValueFrom.ResourceFieldRef != nil {
- numSources++
- allErrs = append(allErrs, validateContainerResourceFieldSelector(ev.ValueFrom.ResourceFieldRef, &validContainerResourceFieldPathExpressions, fldPath.Child("resourceFieldRef"), false)...)
- }
- if ev.ValueFrom.ConfigMapKeyRef != nil {
- numSources++
- allErrs = append(allErrs, validateConfigMapKeySelector(ev.ValueFrom.ConfigMapKeyRef, fldPath.Child("configMapKeyRef"))...)
- }
- if ev.ValueFrom.SecretKeyRef != nil {
- numSources++
- allErrs = append(allErrs, validateSecretKeySelector(ev.ValueFrom.SecretKeyRef, fldPath.Child("secretKeyRef"))...)
- }
-
- if numSources == 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, "", "must specify one of: `fieldRef`, `resourceFieldRef`, `configMapKeyRef` or `secretKeyRef`"))
- } else if len(ev.Value) != 0 {
- if numSources != 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, "", "may not be specified when `value` is not empty"))
- }
- } else if numSources > 1 {
- allErrs = append(allErrs, field.Invalid(fldPath, "", "may not have more than one field specified at a time"))
- }
-
- return allErrs
-}
-
-func validateObjectFieldSelector(fs *core.ObjectFieldSelector, expressions *sets.String, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if len(fs.APIVersion) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("apiVersion"), ""))
- return allErrs
- }
- if len(fs.FieldPath) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("fieldPath"), ""))
- return allErrs
- }
-
- internalFieldPath, _, err := podshelper.ConvertDownwardAPIFieldLabel(fs.APIVersion, fs.FieldPath, "")
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("fieldPath"), fs.FieldPath, fmt.Sprintf("error converting fieldPath: %v", err)))
- return allErrs
- }
-
- if path, subscript, ok := fieldpath.SplitMaybeSubscriptedPath(internalFieldPath); ok {
- switch path {
- case "metadata.annotations":
- for _, msg := range validation.IsQualifiedName(strings.ToLower(subscript)) {
- allErrs = append(allErrs, field.Invalid(fldPath, subscript, msg))
- }
- case "metadata.labels":
- for _, msg := range validation.IsQualifiedName(subscript) {
- allErrs = append(allErrs, field.Invalid(fldPath, subscript, msg))
- }
- default:
- allErrs = append(allErrs, field.Invalid(fldPath, path, "does not support subscript"))
- }
- } else if !expressions.Has(path) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("fieldPath"), path, expressions.List()))
- return allErrs
- }
-
- return allErrs
-}
-
-func fsResourceIsEphemeralStorage(resource string) bool {
- if resource == "limits.ephemeral-storage" || resource == "requests.ephemeral-storage" {
- return true
- }
- return false
-}
-
-func validateContainerResourceFieldSelector(fs *core.ResourceFieldSelector, expressions *sets.String, fldPath *field.Path, volume bool) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if volume && len(fs.ContainerName) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("containerName"), ""))
- } else if len(fs.Resource) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("resource"), ""))
- } else if !expressions.Has(fs.Resource) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("resource"), fs.Resource, expressions.List()))
- } else if fsResourceIsEphemeralStorage(fs.Resource) && !utilfeature.DefaultFeatureGate.Enabled(features.LocalStorageCapacityIsolation) {
- allErrs = append(allErrs, field.Forbidden(fldPath, "Containers' ephemeral storage requests/limits disabled by feature-gate for Downward API"))
- }
- allErrs = append(allErrs, validateContainerResourceDivisor(fs.Resource, fs.Divisor, fldPath)...)
- return allErrs
-}
-
-func ValidateEnvFrom(vars []core.EnvFromSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i, ev := range vars {
- idxPath := fldPath.Index(i)
- if len(ev.Prefix) > 0 {
- for _, msg := range validation.IsEnvVarName(ev.Prefix) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("prefix"), ev.Prefix, msg))
- }
- }
-
- numSources := 0
- if ev.ConfigMapRef != nil {
- numSources++
- allErrs = append(allErrs, validateConfigMapEnvSource(ev.ConfigMapRef, idxPath.Child("configMapRef"))...)
- }
- if ev.SecretRef != nil {
- numSources++
- allErrs = append(allErrs, validateSecretEnvSource(ev.SecretRef, idxPath.Child("secretRef"))...)
- }
-
- if numSources == 0 {
- allErrs = append(allErrs, field.Invalid(fldPath, "", "must specify one of: `configMapRef` or `secretRef`"))
- } else if numSources > 1 {
- allErrs = append(allErrs, field.Invalid(fldPath, "", "may not have more than one field specified at a time"))
- }
- }
- return allErrs
-}
-
-func validateConfigMapEnvSource(configMapSource *core.ConfigMapEnvSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(configMapSource.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- } else {
- for _, msg := range ValidateConfigMapName(configMapSource.Name, true) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), configMapSource.Name, msg))
- }
- }
- return allErrs
-}
-
-func validateSecretEnvSource(secretSource *core.SecretEnvSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(secretSource.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- } else {
- for _, msg := range ValidateSecretName(secretSource.Name, true) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), secretSource.Name, msg))
- }
- }
- return allErrs
-}
-
-var validContainerResourceDivisorForCPU = sets.NewString("1m", "1")
-var validContainerResourceDivisorForMemory = sets.NewString("1", "1k", "1M", "1G", "1T", "1P", "1E", "1Ki", "1Mi", "1Gi", "1Ti", "1Pi", "1Ei")
-var validContainerResourceDivisorForEphemeralStorage = sets.NewString("1", "1k", "1M", "1G", "1T", "1P", "1E", "1Ki", "1Mi", "1Gi", "1Ti", "1Pi", "1Ei")
-
-func validateContainerResourceDivisor(rName string, divisor resource.Quantity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- unsetDivisor := resource.Quantity{}
- if unsetDivisor.Cmp(divisor) == 0 {
- return allErrs
- }
- switch rName {
- case "limits.cpu", "requests.cpu":
- if !validContainerResourceDivisorForCPU.Has(divisor.String()) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1m and 1 are supported with the cpu resource"))
- }
- case "limits.memory", "requests.memory":
- if !validContainerResourceDivisorForMemory.Has(divisor.String()) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1, 1k, 1M, 1G, 1T, 1P, 1E, 1Ki, 1Mi, 1Gi, 1Ti, 1Pi, 1Ei are supported with the memory resource"))
- }
- case "limits.ephemeral-storage", "requests.ephemeral-storage":
- if !validContainerResourceDivisorForEphemeralStorage.Has(divisor.String()) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1, 1k, 1M, 1G, 1T, 1P, 1E, 1Ki, 1Mi, 1Gi, 1Ti, 1Pi, 1Ei are supported with the local ephemeral storage resource"))
- }
- }
- return allErrs
-}
-
-func validateConfigMapKeySelector(s *core.ConfigMapKeySelector, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- nameFn := ValidateNameFunc(ValidateSecretName)
- for _, msg := range nameFn(s.Name, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), s.Name, msg))
- }
- if len(s.Key) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
- } else {
- for _, msg := range validation.IsConfigMapKey(s.Key) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("key"), s.Key, msg))
- }
- }
-
- return allErrs
-}
-
-func validateSecretKeySelector(s *core.SecretKeySelector, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- nameFn := ValidateNameFunc(ValidateSecretName)
- for _, msg := range nameFn(s.Name, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), s.Name, msg))
- }
- if len(s.Key) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
- } else {
- for _, msg := range validation.IsConfigMapKey(s.Key) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("key"), s.Key, msg))
- }
- }
-
- return allErrs
-}
-
-func GetVolumeMountMap(mounts []core.VolumeMount) map[string]string {
- volmounts := make(map[string]string)
-
- for _, mnt := range mounts {
- volmounts[mnt.Name] = mnt.MountPath
- }
-
- return volmounts
-}
-
-func GetVolumeDeviceMap(devices []core.VolumeDevice) map[string]string {
- voldevices := make(map[string]string)
-
- for _, dev := range devices {
- voldevices[dev.Name] = dev.DevicePath
- }
-
- return voldevices
-}
-
-func ValidateVolumeMounts(mounts []core.VolumeMount, voldevices map[string]string, volumes map[string]core.VolumeSource, container *core.Container, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- mountpoints := sets.NewString()
-
- for i, mnt := range mounts {
- idxPath := fldPath.Index(i)
- if len(mnt.Name) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
- }
- if !IsMatchedVolume(mnt.Name, volumes) {
- allErrs = append(allErrs, field.NotFound(idxPath.Child("name"), mnt.Name))
- }
- if len(mnt.MountPath) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("mountPath"), ""))
- }
- if mountpoints.Has(mnt.MountPath) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("mountPath"), mnt.MountPath, "must be unique"))
- }
- mountpoints.Insert(mnt.MountPath)
-
- // check for overlap with VolumeDevice
- if mountNameAlreadyExists(mnt.Name, voldevices) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), mnt.Name, "must not already exist in volumeDevices"))
- }
- if mountPathAlreadyExists(mnt.MountPath, voldevices) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("mountPath"), mnt.MountPath, "must not already exist as a path in volumeDevices"))
- }
-
- if len(mnt.SubPath) > 0 {
- allErrs = append(allErrs, validateLocalDescendingPath(mnt.SubPath, fldPath.Child("subPath"))...)
- }
-
- if mnt.MountPropagation != nil {
- allErrs = append(allErrs, validateMountPropagation(mnt.MountPropagation, container, fldPath.Child("mountPropagation"))...)
- }
- }
- return allErrs
-}
-
-func ValidateVolumeDevices(devices []core.VolumeDevice, volmounts map[string]string, volumes map[string]core.VolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- devicepath := sets.NewString()
- devicename := sets.NewString()
-
- if devices != nil && !utilfeature.DefaultFeatureGate.Enabled(features.BlockVolume) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("volumeDevices"), "Container volumeDevices is disabled by feature-gate"))
- return allErrs
- }
- if devices != nil {
- for i, dev := range devices {
- idxPath := fldPath.Index(i)
- devName := dev.Name
- devPath := dev.DevicePath
- didMatch, isPVC := isMatchedDevice(devName, volumes)
- if len(devName) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
- }
- if devicename.Has(devName) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "must be unique"))
- }
- // Must be PersistentVolumeClaim volume source
- if didMatch && !isPVC {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "can only use volume source type of PersistentVolumeClaim for block mode"))
- }
- if !didMatch {
- allErrs = append(allErrs, field.NotFound(idxPath.Child("name"), devName))
- }
- if len(devPath) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("devicePath"), ""))
- }
- if devicepath.Has(devPath) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "must be unique"))
- }
- if len(devPath) > 0 && len(validatePathNoBacksteps(devPath, fldPath.Child("devicePath"))) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "can not contain backsteps ('..')"))
- } else {
- devicepath.Insert(devPath)
- }
- // check for overlap with VolumeMount
- if deviceNameAlreadyExists(devName, volmounts) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "must not already exist in volumeMounts"))
- }
- if devicePathAlreadyExists(devPath, volmounts) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "must not already exist as a path in volumeMounts"))
- }
- if len(devName) > 0 {
- devicename.Insert(devName)
- }
- }
- }
- return allErrs
-}
-
-func validateProbe(probe *core.Probe, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if probe == nil {
- return allErrs
- }
- allErrs = append(allErrs, validateHandler(&probe.Handler, fldPath)...)
-
- allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.InitialDelaySeconds), fldPath.Child("initialDelaySeconds"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.TimeoutSeconds), fldPath.Child("timeoutSeconds"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.PeriodSeconds), fldPath.Child("periodSeconds"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.SuccessThreshold), fldPath.Child("successThreshold"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.FailureThreshold), fldPath.Child("failureThreshold"))...)
- return allErrs
-}
-
-func validateClientIPAffinityConfig(config *core.SessionAffinityConfig, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if config == nil {
- allErrs = append(allErrs, field.Required(fldPath, fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
- return allErrs
- }
- if config.ClientIP == nil {
- allErrs = append(allErrs, field.Required(fldPath.Child("clientIP"), fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
- return allErrs
- }
- if config.ClientIP.TimeoutSeconds == nil {
- allErrs = append(allErrs, field.Required(fldPath.Child("clientIP").Child("timeoutSeconds"), fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
- return allErrs
- }
- allErrs = append(allErrs, validateAffinityTimeout(config.ClientIP.TimeoutSeconds, fldPath.Child("clientIP").Child("timeoutSeconds"))...)
-
- return allErrs
-}
-
-func validateAffinityTimeout(timeout *int32, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if *timeout <= 0 || *timeout > core.MaxClientIPServiceAffinitySeconds {
- allErrs = append(allErrs, field.Invalid(fldPath, timeout, fmt.Sprintf("must be greater than 0 and less than %d", core.MaxClientIPServiceAffinitySeconds)))
- }
- return allErrs
-}
-
-// AccumulateUniqueHostPorts extracts each HostPort of each Container,
-// accumulating the results and returning an error if any ports conflict.
-func AccumulateUniqueHostPorts(containers []core.Container, accumulator *sets.String, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- for ci, ctr := range containers {
- idxPath := fldPath.Index(ci)
- portsPath := idxPath.Child("ports")
- for pi := range ctr.Ports {
- idxPath := portsPath.Index(pi)
- port := ctr.Ports[pi].HostPort
- if port == 0 {
- continue
- }
- str := fmt.Sprintf("%s/%s/%d", ctr.Ports[pi].Protocol, ctr.Ports[pi].HostIP, port)
- if accumulator.Has(str) {
- allErrs = append(allErrs, field.Duplicate(idxPath.Child("hostPort"), str))
- } else {
- accumulator.Insert(str)
- }
- }
- }
- return allErrs
-}
-
-// checkHostPortConflicts checks for colliding Port.HostPort values across
-// a slice of containers.
-func checkHostPortConflicts(containers []core.Container, fldPath *field.Path) field.ErrorList {
- allPorts := sets.String{}
- return AccumulateUniqueHostPorts(containers, &allPorts, fldPath)
-}
-
-func validateExecAction(exec *core.ExecAction, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- if len(exec.Command) == 0 {
- allErrors = append(allErrors, field.Required(fldPath.Child("command"), ""))
- }
- return allErrors
-}
-
-var supportedHTTPSchemes = sets.NewString(string(core.URISchemeHTTP), string(core.URISchemeHTTPS))
-
-func validateHTTPGetAction(http *core.HTTPGetAction, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- if len(http.Path) == 0 {
- allErrors = append(allErrors, field.Required(fldPath.Child("path"), ""))
- }
- allErrors = append(allErrors, ValidatePortNumOrName(http.Port, fldPath.Child("port"))...)
- if !supportedHTTPSchemes.Has(string(http.Scheme)) {
- allErrors = append(allErrors, field.NotSupported(fldPath.Child("scheme"), http.Scheme, supportedHTTPSchemes.List()))
- }
- for _, header := range http.HTTPHeaders {
- for _, msg := range validation.IsHTTPHeaderName(header.Name) {
- allErrors = append(allErrors, field.Invalid(fldPath.Child("httpHeaders"), header.Name, msg))
- }
- }
- return allErrors
-}
-
-func ValidatePortNumOrName(port intstr.IntOrString, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if port.Type == intstr.Int {
- for _, msg := range validation.IsValidPortNum(port.IntValue()) {
- allErrs = append(allErrs, field.Invalid(fldPath, port.IntValue(), msg))
- }
- } else if port.Type == intstr.String {
- for _, msg := range validation.IsValidPortName(port.StrVal) {
- allErrs = append(allErrs, field.Invalid(fldPath, port.StrVal, msg))
- }
- } else {
- allErrs = append(allErrs, field.InternalError(fldPath, fmt.Errorf("unknown type: %v", port.Type)))
- }
- return allErrs
-}
-
-func validateTCPSocketAction(tcp *core.TCPSocketAction, fldPath *field.Path) field.ErrorList {
- return ValidatePortNumOrName(tcp.Port, fldPath.Child("port"))
-}
-
-func validateHandler(handler *core.Handler, fldPath *field.Path) field.ErrorList {
- numHandlers := 0
- allErrors := field.ErrorList{}
- if handler.Exec != nil {
- if numHandlers > 0 {
- allErrors = append(allErrors, field.Forbidden(fldPath.Child("exec"), "may not specify more than 1 handler type"))
- } else {
- numHandlers++
- allErrors = append(allErrors, validateExecAction(handler.Exec, fldPath.Child("exec"))...)
- }
- }
- if handler.HTTPGet != nil {
- if numHandlers > 0 {
- allErrors = append(allErrors, field.Forbidden(fldPath.Child("httpGet"), "may not specify more than 1 handler type"))
- } else {
- numHandlers++
- allErrors = append(allErrors, validateHTTPGetAction(handler.HTTPGet, fldPath.Child("httpGet"))...)
- }
- }
- if handler.TCPSocket != nil {
- if numHandlers > 0 {
- allErrors = append(allErrors, field.Forbidden(fldPath.Child("tcpSocket"), "may not specify more than 1 handler type"))
- } else {
- numHandlers++
- allErrors = append(allErrors, validateTCPSocketAction(handler.TCPSocket, fldPath.Child("tcpSocket"))...)
- }
- }
- if numHandlers == 0 {
- allErrors = append(allErrors, field.Required(fldPath, "must specify a handler type"))
- }
- return allErrors
-}
-
-func validateLifecycle(lifecycle *core.Lifecycle, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if lifecycle.PostStart != nil {
- allErrs = append(allErrs, validateHandler(lifecycle.PostStart, fldPath.Child("postStart"))...)
- }
- if lifecycle.PreStop != nil {
- allErrs = append(allErrs, validateHandler(lifecycle.PreStop, fldPath.Child("preStop"))...)
- }
- return allErrs
-}
-
-var supportedPullPolicies = sets.NewString(string(core.PullAlways), string(core.PullIfNotPresent), string(core.PullNever))
-
-func validatePullPolicy(policy core.PullPolicy, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
-
- switch policy {
- case core.PullAlways, core.PullIfNotPresent, core.PullNever:
- break
- case "":
- allErrors = append(allErrors, field.Required(fldPath, ""))
- default:
- allErrors = append(allErrors, field.NotSupported(fldPath, policy, supportedPullPolicies.List()))
- }
-
- return allErrors
-}
-
-func validateInitContainers(containers, otherContainers []core.Container, deviceVolumes map[string]core.VolumeSource, fldPath *field.Path) field.ErrorList {
- var allErrs field.ErrorList
- if len(containers) > 0 {
- allErrs = append(allErrs, validateContainers(containers, deviceVolumes, fldPath)...)
- }
-
- allNames := sets.String{}
- for _, ctr := range otherContainers {
- allNames.Insert(ctr.Name)
- }
- for i, ctr := range containers {
- idxPath := fldPath.Index(i)
- if allNames.Has(ctr.Name) {
- allErrs = append(allErrs, field.Duplicate(idxPath.Child("name"), ctr.Name))
- }
- if len(ctr.Name) > 0 {
- allNames.Insert(ctr.Name)
- }
- if ctr.Lifecycle != nil {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("lifecycle"), ctr.Lifecycle, "must not be set for init containers"))
- }
- if ctr.LivenessProbe != nil {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("livenessProbe"), ctr.LivenessProbe, "must not be set for init containers"))
- }
- if ctr.ReadinessProbe != nil {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("readinessProbe"), ctr.ReadinessProbe, "must not be set for init containers"))
- }
- }
- return allErrs
-}
-
-func validateContainers(containers []core.Container, volumes map[string]core.VolumeSource, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if len(containers) == 0 {
- return append(allErrs, field.Required(fldPath, ""))
- }
-
- allNames := sets.String{}
- for i, ctr := range containers {
- idxPath := fldPath.Index(i)
- namePath := idxPath.Child("name")
- volMounts := GetVolumeMountMap(ctr.VolumeMounts)
- volDevices := GetVolumeDeviceMap(ctr.VolumeDevices)
-
- if len(ctr.Name) == 0 {
- allErrs = append(allErrs, field.Required(namePath, ""))
- } else {
- allErrs = append(allErrs, ValidateDNS1123Label(ctr.Name, namePath)...)
- }
- if allNames.Has(ctr.Name) {
- allErrs = append(allErrs, field.Duplicate(namePath, ctr.Name))
- } else {
- allNames.Insert(ctr.Name)
- }
- // TODO: do not validate leading and trailing whitespace to preserve backward compatibility.
- // for example: https://github.com/openshift/origin/issues/14659 image = " " is special token in pod template
- // others may have done similar
- if len(ctr.Image) == 0 {
- allErrs = append(allErrs, field.Required(idxPath.Child("image"), ""))
- }
- if ctr.Lifecycle != nil {
- allErrs = append(allErrs, validateLifecycle(ctr.Lifecycle, idxPath.Child("lifecycle"))...)
- }
- allErrs = append(allErrs, validateProbe(ctr.LivenessProbe, idxPath.Child("livenessProbe"))...)
- // Liveness-specific validation
- if ctr.LivenessProbe != nil && ctr.LivenessProbe.SuccessThreshold != 1 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("livenessProbe", "successThreshold"), ctr.LivenessProbe.SuccessThreshold, "must be 1"))
- }
-
- switch ctr.TerminationMessagePolicy {
- case core.TerminationMessageReadFile, core.TerminationMessageFallbackToLogsOnError:
- case "":
- allErrs = append(allErrs, field.Required(idxPath.Child("terminationMessagePolicy"), "must be 'File' or 'FallbackToLogsOnError'"))
- default:
- allErrs = append(allErrs, field.Invalid(idxPath.Child("terminationMessagePolicy"), ctr.TerminationMessagePolicy, "must be 'File' or 'FallbackToLogsOnError'"))
- }
-
- allErrs = append(allErrs, validateProbe(ctr.ReadinessProbe, idxPath.Child("readinessProbe"))...)
- allErrs = append(allErrs, validateContainerPorts(ctr.Ports, idxPath.Child("ports"))...)
- allErrs = append(allErrs, ValidateEnv(ctr.Env, idxPath.Child("env"))...)
- allErrs = append(allErrs, ValidateEnvFrom(ctr.EnvFrom, idxPath.Child("envFrom"))...)
- allErrs = append(allErrs, ValidateVolumeMounts(ctr.VolumeMounts, volDevices, volumes, &ctr, idxPath.Child("volumeMounts"))...)
- allErrs = append(allErrs, ValidateVolumeDevices(ctr.VolumeDevices, volMounts, volumes, idxPath.Child("volumeDevices"))...)
- allErrs = append(allErrs, validatePullPolicy(ctr.ImagePullPolicy, idxPath.Child("imagePullPolicy"))...)
- allErrs = append(allErrs, ValidateResourceRequirements(&ctr.Resources, idxPath.Child("resources"))...)
- allErrs = append(allErrs, ValidateSecurityContext(ctr.SecurityContext, idxPath.Child("securityContext"))...)
- }
- // Check for colliding ports across all containers.
- allErrs = append(allErrs, checkHostPortConflicts(containers, fldPath)...)
-
- return allErrs
-}
-
-func validateRestartPolicy(restartPolicy *core.RestartPolicy, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- switch *restartPolicy {
- case core.RestartPolicyAlways, core.RestartPolicyOnFailure, core.RestartPolicyNever:
- break
- case "":
- allErrors = append(allErrors, field.Required(fldPath, ""))
- default:
- validValues := []string{string(core.RestartPolicyAlways), string(core.RestartPolicyOnFailure), string(core.RestartPolicyNever)}
- allErrors = append(allErrors, field.NotSupported(fldPath, *restartPolicy, validValues))
- }
-
- return allErrors
-}
-
-func validateDNSPolicy(dnsPolicy *core.DNSPolicy, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- switch *dnsPolicy {
- case core.DNSClusterFirstWithHostNet, core.DNSClusterFirst, core.DNSDefault:
- case core.DNSNone:
- if !utilfeature.DefaultFeatureGate.Enabled(features.CustomPodDNS) {
- allErrors = append(allErrors, field.Invalid(fldPath, dnsPolicy, "DNSPolicy: can not use 'None', custom pod DNS is disabled by feature gate"))
- }
- case "":
- allErrors = append(allErrors, field.Required(fldPath, ""))
- default:
- validValues := []string{string(core.DNSClusterFirstWithHostNet), string(core.DNSClusterFirst), string(core.DNSDefault)}
- if utilfeature.DefaultFeatureGate.Enabled(features.CustomPodDNS) {
- validValues = append(validValues, string(core.DNSNone))
- }
- allErrors = append(allErrors, field.NotSupported(fldPath, dnsPolicy, validValues))
- }
- return allErrors
-}
-
-const (
- // Limits on various DNS parameters. These are derived from
- // restrictions in Linux libc name resolution handling.
- // Max number of DNS name servers.
- MaxDNSNameservers = 3
- // Max number of domains in search path.
- MaxDNSSearchPaths = 6
- // Max number of characters in search path.
- MaxDNSSearchListChars = 256
-)
-
-func validatePodDNSConfig(dnsConfig *core.PodDNSConfig, dnsPolicy *core.DNSPolicy, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- // Validate DNSNone case. Must provide at least one DNS name server.
- if utilfeature.DefaultFeatureGate.Enabled(features.CustomPodDNS) && dnsPolicy != nil && *dnsPolicy == core.DNSNone {
- if dnsConfig == nil {
- return append(allErrs, field.Required(fldPath, fmt.Sprintf("must provide `dnsConfig` when `dnsPolicy` is %s", core.DNSNone)))
- }
- if len(dnsConfig.Nameservers) == 0 {
- return append(allErrs, field.Required(fldPath.Child("nameservers"), fmt.Sprintf("must provide at least one DNS nameserver when `dnsPolicy` is %s", core.DNSNone)))
- }
- }
-
- if dnsConfig != nil {
- if !utilfeature.DefaultFeatureGate.Enabled(features.CustomPodDNS) {
- return append(allErrs, field.Forbidden(fldPath, "DNSConfig: custom pod DNS is disabled by feature gate"))
- }
-
- // Validate nameservers.
- if len(dnsConfig.Nameservers) > MaxDNSNameservers {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("nameservers"), dnsConfig.Nameservers, fmt.Sprintf("must not have more than %v nameservers", MaxDNSNameservers)))
- }
- for i, ns := range dnsConfig.Nameservers {
- if ip := net.ParseIP(ns); ip == nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("nameservers").Index(i), ns, "must be valid IP address"))
- }
- }
- // Validate searches.
- if len(dnsConfig.Searches) > MaxDNSSearchPaths {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("searches"), dnsConfig.Searches, fmt.Sprintf("must not have more than %v search paths", MaxDNSSearchPaths)))
- }
- // Include the space between search paths.
- if len(strings.Join(dnsConfig.Searches, " ")) > MaxDNSSearchListChars {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("searches"), dnsConfig.Searches, "must not have more than 256 characters (including spaces) in the search list"))
- }
- for i, search := range dnsConfig.Searches {
- allErrs = append(allErrs, ValidateDNS1123Subdomain(search, fldPath.Child("searches").Index(i))...)
- }
- // Validate options.
- for i, option := range dnsConfig.Options {
- if len(option.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("options").Index(i), "must not be empty"))
- }
- }
- }
- return allErrs
-}
-
-func validateHostNetwork(hostNetwork bool, containers []core.Container, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- if hostNetwork {
- for i, container := range containers {
- portsPath := fldPath.Index(i).Child("ports")
- for i, port := range container.Ports {
- idxPath := portsPath.Index(i)
- if port.HostPort != port.ContainerPort {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("containerPort"), port.ContainerPort, "must match `hostPort` when `hostNetwork` is true"))
- }
- }
- }
- }
- return allErrors
-}
-
-// validateImagePullSecrets checks to make sure the pull secrets are well
-// formed. Right now, we only expect name to be set (it's the only field). If
-// this ever changes and someone decides to set those fields, we'd like to
-// know.
-func validateImagePullSecrets(imagePullSecrets []core.LocalObjectReference, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- for i, currPullSecret := range imagePullSecrets {
- idxPath := fldPath.Index(i)
- strippedRef := core.LocalObjectReference{Name: currPullSecret.Name}
- if !reflect.DeepEqual(strippedRef, currPullSecret) {
- allErrors = append(allErrors, field.Invalid(idxPath, currPullSecret, "only name may be set"))
- }
- }
- return allErrors
-}
-
-// validateAffinity checks if given affinities are valid
-func validateAffinity(affinity *core.Affinity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if affinity != nil {
- if affinity.NodeAffinity != nil {
- allErrs = append(allErrs, validateNodeAffinity(affinity.NodeAffinity, fldPath.Child("nodeAffinity"))...)
- }
- if affinity.PodAffinity != nil {
- allErrs = append(allErrs, validatePodAffinity(affinity.PodAffinity, fldPath.Child("podAffinity"))...)
- }
- if affinity.PodAntiAffinity != nil {
- allErrs = append(allErrs, validatePodAntiAffinity(affinity.PodAntiAffinity, fldPath.Child("podAntiAffinity"))...)
- }
- }
-
- return allErrs
-}
-
-func validateTaintEffect(effect *core.TaintEffect, allowEmpty bool, fldPath *field.Path) field.ErrorList {
- if !allowEmpty && len(*effect) == 0 {
- return field.ErrorList{field.Required(fldPath, "")}
- }
-
- allErrors := field.ErrorList{}
- switch *effect {
- // TODO: Replace next line with subsequent commented-out line when implement TaintEffectNoScheduleNoAdmit.
- case core.TaintEffectNoSchedule, core.TaintEffectPreferNoSchedule, core.TaintEffectNoExecute:
- // case core.TaintEffectNoSchedule, core.TaintEffectPreferNoSchedule, core.TaintEffectNoScheduleNoAdmit, core.TaintEffectNoExecute:
- default:
- validValues := []string{
- string(core.TaintEffectNoSchedule),
- string(core.TaintEffectPreferNoSchedule),
- string(core.TaintEffectNoExecute),
- // TODO: Uncomment this block when implement TaintEffectNoScheduleNoAdmit.
- // string(core.TaintEffectNoScheduleNoAdmit),
- }
- allErrors = append(allErrors, field.NotSupported(fldPath, effect, validValues))
- }
- return allErrors
-}
-
-// validateOnlyAddedTolerations validates updated pod tolerations.
-func validateOnlyAddedTolerations(newTolerations []core.Toleration, oldTolerations []core.Toleration, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, old := range oldTolerations {
- found := false
- old.TolerationSeconds = nil
- for _, new := range newTolerations {
- new.TolerationSeconds = nil
- if reflect.DeepEqual(old, new) {
- found = true
- break
- }
- }
- if !found {
- allErrs = append(allErrs, field.Forbidden(fldPath, "existing toleration can not be modified except its tolerationSeconds"))
- return allErrs
- }
- }
-
- allErrs = append(allErrs, ValidateTolerations(newTolerations, fldPath)...)
- return allErrs
-}
-
-func ValidateHostAliases(hostAliases []core.HostAlias, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, hostAlias := range hostAliases {
- if ip := net.ParseIP(hostAlias.IP); ip == nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("ip"), hostAlias.IP, "must be valid IP address"))
- }
- for _, hostname := range hostAlias.Hostnames {
- allErrs = append(allErrs, ValidateDNS1123Subdomain(hostname, fldPath.Child("hostnames"))...)
- }
- }
- return allErrs
-}
-
-// ValidateTolerations tests if given tolerations have valid data.
-func ValidateTolerations(tolerations []core.Toleration, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- for i, toleration := range tolerations {
- idxPath := fldPath.Index(i)
- // validate the toleration key
- if len(toleration.Key) > 0 {
- allErrors = append(allErrors, unversionedvalidation.ValidateLabelName(toleration.Key, idxPath.Child("key"))...)
- }
-
- // empty toleration key with Exists operator and empty value means match all taints
- if len(toleration.Key) == 0 && toleration.Operator != core.TolerationOpExists {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration.Operator,
- "operator must be Exists when `key` is empty, which means \"match all values and all keys\""))
- }
-
- if toleration.TolerationSeconds != nil && toleration.Effect != core.TaintEffectNoExecute {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("effect"), toleration.Effect,
- "effect must be 'NoExecute' when `tolerationSeconds` is set"))
- }
-
- // validate toleration operator and value
- switch toleration.Operator {
- // empty operator means Equal
- case core.TolerationOpEqual, "":
- if errs := validation.IsValidLabelValue(toleration.Value); len(errs) != 0 {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration.Value, strings.Join(errs, ";")))
- }
- case core.TolerationOpExists:
- if len(toleration.Value) > 0 {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration, "value must be empty when `operator` is 'Exists'"))
- }
- default:
- validValues := []string{string(core.TolerationOpEqual), string(core.TolerationOpExists)}
- allErrors = append(allErrors, field.NotSupported(idxPath.Child("operator"), toleration.Operator, validValues))
- }
-
- // validate toleration effect, empty toleration effect means match all taint effects
- if len(toleration.Effect) > 0 {
- allErrors = append(allErrors, validateTaintEffect(&toleration.Effect, true, idxPath.Child("effect"))...)
- }
- }
- return allErrors
-}
-
-func toResourceNames(resources core.ResourceList) []core.ResourceName {
- result := []core.ResourceName{}
- for resourceName := range resources {
- result = append(result, resourceName)
- }
- return result
-}
-
-func toSet(resourceNames []core.ResourceName) sets.String {
- result := sets.NewString()
- for _, resourceName := range resourceNames {
- result.Insert(string(resourceName))
- }
- return result
-}
-
-func toContainerResourcesSet(ctr *core.Container) sets.String {
- resourceNames := toResourceNames(ctr.Resources.Requests)
- resourceNames = append(resourceNames, toResourceNames(ctr.Resources.Limits)...)
- return toSet(resourceNames)
-}
-
-// validateContainersOnlyForPod does additional validation for containers on a pod versus a pod template
-// it only does additive validation of fields not covered in validateContainers
-func validateContainersOnlyForPod(containers []core.Container, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i, ctr := range containers {
- idxPath := fldPath.Index(i)
- if len(ctr.Image) != len(strings.TrimSpace(ctr.Image)) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("image"), ctr.Image, "must not have leading or trailing whitespace"))
- }
- }
- return allErrs
-}
-
-// ValidatePod tests if required fields in the pod are set.
-func ValidatePod(pod *core.Pod) field.ErrorList {
- fldPath := field.NewPath("metadata")
- allErrs := ValidateObjectMeta(&pod.ObjectMeta, true, ValidatePodName, fldPath)
- allErrs = append(allErrs, ValidatePodSpecificAnnotations(pod.ObjectMeta.Annotations, &pod.Spec, fldPath.Child("annotations"))...)
- allErrs = append(allErrs, ValidatePodSpec(&pod.Spec, field.NewPath("spec"))...)
-
- // we do additional validation only pertinent for pods and not pod templates
- // this was done to preserve backwards compatibility
- specPath := field.NewPath("spec")
-
- allErrs = append(allErrs, validateContainersOnlyForPod(pod.Spec.Containers, specPath.Child("containers"))...)
- allErrs = append(allErrs, validateContainersOnlyForPod(pod.Spec.InitContainers, specPath.Child("initContainers"))...)
-
- if utilfeature.DefaultFeatureGate.Enabled(features.HugePages) {
- hugePageResources := sets.NewString()
- for i := range pod.Spec.Containers {
- resourceSet := toContainerResourcesSet(&pod.Spec.Containers[i])
- for resourceStr := range resourceSet {
- if v1helper.IsHugePageResourceName(v1.ResourceName(resourceStr)) {
- hugePageResources.Insert(resourceStr)
- }
- }
- }
- if len(hugePageResources) > 1 {
- allErrs = append(allErrs, field.Invalid(specPath, hugePageResources, "must use a single hugepage size in a pod spec"))
- }
- }
-
- return allErrs
-}
-
-// ValidatePodSpec tests that the specified PodSpec has valid data.
-// This includes checking formatting and uniqueness. It also canonicalizes the
-// structure by setting default values and implementing any backwards-compatibility
-// tricks.
-func ValidatePodSpec(spec *core.PodSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- vols, vErrs := ValidateVolumes(spec.Volumes, fldPath.Child("volumes"))
- allErrs = append(allErrs, vErrs...)
- allErrs = append(allErrs, validateContainers(spec.Containers, vols, fldPath.Child("containers"))...)
- allErrs = append(allErrs, validateInitContainers(spec.InitContainers, spec.Containers, vols, fldPath.Child("initContainers"))...)
- allErrs = append(allErrs, validateRestartPolicy(&spec.RestartPolicy, fldPath.Child("restartPolicy"))...)
- allErrs = append(allErrs, validateDNSPolicy(&spec.DNSPolicy, fldPath.Child("dnsPolicy"))...)
- allErrs = append(allErrs, unversionedvalidation.ValidateLabels(spec.NodeSelector, fldPath.Child("nodeSelector"))...)
- allErrs = append(allErrs, ValidatePodSecurityContext(spec.SecurityContext, spec, fldPath, fldPath.Child("securityContext"))...)
- allErrs = append(allErrs, validateImagePullSecrets(spec.ImagePullSecrets, fldPath.Child("imagePullSecrets"))...)
- allErrs = append(allErrs, validateAffinity(spec.Affinity, fldPath.Child("affinity"))...)
- allErrs = append(allErrs, validatePodDNSConfig(spec.DNSConfig, &spec.DNSPolicy, fldPath.Child("dnsConfig"))...)
- if len(spec.ServiceAccountName) > 0 {
- for _, msg := range ValidateServiceAccountName(spec.ServiceAccountName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("serviceAccountName"), spec.ServiceAccountName, msg))
- }
- }
-
- if len(spec.NodeName) > 0 {
- for _, msg := range ValidateNodeName(spec.NodeName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("nodeName"), spec.NodeName, msg))
- }
- }
-
- if spec.ActiveDeadlineSeconds != nil {
- value := *spec.ActiveDeadlineSeconds
- if value < 1 || value > math.MaxInt32 {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("activeDeadlineSeconds"), value, validation.InclusiveRangeError(1, math.MaxInt32)))
- }
- }
-
- if len(spec.Hostname) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(spec.Hostname, fldPath.Child("hostname"))...)
- }
-
- if len(spec.Subdomain) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(spec.Subdomain, fldPath.Child("subdomain"))...)
- }
-
- if len(spec.Tolerations) > 0 {
- allErrs = append(allErrs, ValidateTolerations(spec.Tolerations, fldPath.Child("tolerations"))...)
- }
-
- if len(spec.HostAliases) > 0 {
- allErrs = append(allErrs, ValidateHostAliases(spec.HostAliases, fldPath.Child("hostAliases"))...)
- }
-
- if len(spec.PriorityClassName) > 0 {
- if utilfeature.DefaultFeatureGate.Enabled(features.PodPriority) {
- for _, msg := range ValidatePriorityClassName(spec.PriorityClassName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("priorityClassName"), spec.PriorityClassName, msg))
- }
- }
- }
-
- return allErrs
-}
-
-// ValidateNodeSelectorRequirement tests that the specified NodeSelectorRequirement fields has valid data
-func ValidateNodeSelectorRequirement(rq core.NodeSelectorRequirement, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- switch rq.Operator {
- case core.NodeSelectorOpIn, core.NodeSelectorOpNotIn:
- if len(rq.Values) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("values"), "must be specified when `operator` is 'In' or 'NotIn'"))
- }
- case core.NodeSelectorOpExists, core.NodeSelectorOpDoesNotExist:
- if len(rq.Values) > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("values"), "may not be specified when `operator` is 'Exists' or 'DoesNotExist'"))
- }
-
- case core.NodeSelectorOpGt, core.NodeSelectorOpLt:
- if len(rq.Values) != 1 {
- allErrs = append(allErrs, field.Required(fldPath.Child("values"), "must be specified single value when `operator` is 'Lt' or 'Gt'"))
- }
- default:
- allErrs = append(allErrs, field.Invalid(fldPath.Child("operator"), rq.Operator, "not a valid selector operator"))
- }
- allErrs = append(allErrs, unversionedvalidation.ValidateLabelName(rq.Key, fldPath.Child("key"))...)
- return allErrs
-}
-
-// ValidateNodeSelectorTerm tests that the specified node selector term has valid data
-func ValidateNodeSelectorTerm(term core.NodeSelectorTerm, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if len(term.MatchExpressions) == 0 {
- return append(allErrs, field.Required(fldPath.Child("matchExpressions"), "must have at least one node selector requirement"))
- }
- for j, req := range term.MatchExpressions {
- allErrs = append(allErrs, ValidateNodeSelectorRequirement(req, fldPath.Child("matchExpressions").Index(j))...)
- }
- return allErrs
-}
-
-// ValidateNodeSelector tests that the specified nodeSelector fields has valid data
-func ValidateNodeSelector(nodeSelector *core.NodeSelector, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- termFldPath := fldPath.Child("nodeSelectorTerms")
- if len(nodeSelector.NodeSelectorTerms) == 0 {
- return append(allErrs, field.Required(termFldPath, "must have at least one node selector term"))
- }
-
- for i, term := range nodeSelector.NodeSelectorTerms {
- allErrs = append(allErrs, ValidateNodeSelectorTerm(term, termFldPath.Index(i))...)
- }
-
- return allErrs
-}
-
-// ValidateAvoidPodsInNodeAnnotations tests that the serialized AvoidPods in Node.Annotations has valid data
-func ValidateAvoidPodsInNodeAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- v1Avoids, err := v1helper.GetAvoidPodsFromNodeAnnotations(annotations)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("AvoidPods"), core.PreferAvoidPodsAnnotationKey, err.Error()))
- return allErrs
- }
- var avoids core.AvoidPods
- if err := corev1.Convert_v1_AvoidPods_To_core_AvoidPods(&v1Avoids, &avoids, nil); err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("AvoidPods"), core.PreferAvoidPodsAnnotationKey, err.Error()))
- return allErrs
- }
-
- if len(avoids.PreferAvoidPods) != 0 {
- for i, pa := range avoids.PreferAvoidPods {
- idxPath := fldPath.Child(core.PreferAvoidPodsAnnotationKey).Index(i)
- allErrs = append(allErrs, validatePreferAvoidPodsEntry(pa, idxPath)...)
- }
- }
-
- return allErrs
-}
-
-// validatePreferAvoidPodsEntry tests if given PreferAvoidPodsEntry has valid data.
-func validatePreferAvoidPodsEntry(avoidPodEntry core.PreferAvoidPodsEntry, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
- if avoidPodEntry.PodSignature.PodController == nil {
- allErrors = append(allErrors, field.Required(fldPath.Child("PodSignature"), ""))
- } else {
- if *(avoidPodEntry.PodSignature.PodController.Controller) != true {
- allErrors = append(allErrors,
- field.Invalid(fldPath.Child("PodSignature").Child("PodController").Child("Controller"),
- *(avoidPodEntry.PodSignature.PodController.Controller), "must point to a controller"))
- }
- }
- return allErrors
-}
-
-// ValidatePreferredSchedulingTerms tests that the specified SoftNodeAffinity fields has valid data
-func ValidatePreferredSchedulingTerms(terms []core.PreferredSchedulingTerm, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- for i, term := range terms {
- if term.Weight <= 0 || term.Weight > 100 {
- allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("weight"), term.Weight, "must be in the range 1-100"))
- }
-
- allErrs = append(allErrs, ValidateNodeSelectorTerm(term.Preference, fldPath.Index(i).Child("preference"))...)
- }
- return allErrs
-}
-
-// validatePodAffinityTerm tests that the specified podAffinityTerm fields have valid data
-func validatePodAffinityTerm(podAffinityTerm core.PodAffinityTerm, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- allErrs = append(allErrs, unversionedvalidation.ValidateLabelSelector(podAffinityTerm.LabelSelector, fldPath.Child("matchExpressions"))...)
- for _, name := range podAffinityTerm.Namespaces {
- for _, msg := range ValidateNamespaceName(name, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("namespace"), name, msg))
- }
- }
- if len(podAffinityTerm.TopologyKey) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("topologyKey"), "can not be empty"))
- }
- return append(allErrs, unversionedvalidation.ValidateLabelName(podAffinityTerm.TopologyKey, fldPath.Child("topologyKey"))...)
-}
-
-// validatePodAffinityTerms tests that the specified podAffinityTerms fields have valid data
-func validatePodAffinityTerms(podAffinityTerms []core.PodAffinityTerm, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i, podAffinityTerm := range podAffinityTerms {
- allErrs = append(allErrs, validatePodAffinityTerm(podAffinityTerm, fldPath.Index(i))...)
- }
- return allErrs
-}
-
-// validateWeightedPodAffinityTerms tests that the specified weightedPodAffinityTerms fields have valid data
-func validateWeightedPodAffinityTerms(weightedPodAffinityTerms []core.WeightedPodAffinityTerm, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for j, weightedTerm := range weightedPodAffinityTerms {
- if weightedTerm.Weight <= 0 || weightedTerm.Weight > 100 {
- allErrs = append(allErrs, field.Invalid(fldPath.Index(j).Child("weight"), weightedTerm.Weight, "must be in the range 1-100"))
- }
- allErrs = append(allErrs, validatePodAffinityTerm(weightedTerm.PodAffinityTerm, fldPath.Index(j).Child("podAffinityTerm"))...)
- }
- return allErrs
-}
-
-// validatePodAntiAffinity tests that the specified podAntiAffinity fields have valid data
-func validatePodAntiAffinity(podAntiAffinity *core.PodAntiAffinity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- // TODO:Uncomment below code once RequiredDuringSchedulingRequiredDuringExecution is implemented.
- // if podAntiAffinity.RequiredDuringSchedulingRequiredDuringExecution != nil {
- // allErrs = append(allErrs, validatePodAffinityTerms(podAntiAffinity.RequiredDuringSchedulingRequiredDuringExecution, false,
- // fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
- //}
- if podAntiAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, validatePodAffinityTerms(podAntiAffinity.RequiredDuringSchedulingIgnoredDuringExecution,
- fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
- }
- if podAntiAffinity.PreferredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, validateWeightedPodAffinityTerms(podAntiAffinity.PreferredDuringSchedulingIgnoredDuringExecution,
- fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
- }
- return allErrs
-}
-
-// validateNodeAffinity tests that the specified nodeAffinity fields have valid data
-func validateNodeAffinity(na *core.NodeAffinity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- // TODO: Uncomment the next three lines once RequiredDuringSchedulingRequiredDuringExecution is implemented.
- // if na.RequiredDuringSchedulingRequiredDuringExecution != nil {
- // allErrs = append(allErrs, ValidateNodeSelector(na.RequiredDuringSchedulingRequiredDuringExecution, fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
- // }
- if na.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, ValidateNodeSelector(na.RequiredDuringSchedulingIgnoredDuringExecution, fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
- }
- if len(na.PreferredDuringSchedulingIgnoredDuringExecution) > 0 {
- allErrs = append(allErrs, ValidatePreferredSchedulingTerms(na.PreferredDuringSchedulingIgnoredDuringExecution, fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
- }
- return allErrs
-}
-
-// validatePodAffinity tests that the specified podAffinity fields have valid data
-func validatePodAffinity(podAffinity *core.PodAffinity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- // TODO:Uncomment below code once RequiredDuringSchedulingRequiredDuringExecution is implemented.
- // if podAffinity.RequiredDuringSchedulingRequiredDuringExecution != nil {
- // allErrs = append(allErrs, validatePodAffinityTerms(podAffinity.RequiredDuringSchedulingRequiredDuringExecution, false,
- // fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
- //}
- if podAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, validatePodAffinityTerms(podAffinity.RequiredDuringSchedulingIgnoredDuringExecution,
- fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
- }
- if podAffinity.PreferredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, validateWeightedPodAffinityTerms(podAffinity.PreferredDuringSchedulingIgnoredDuringExecution,
- fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
- }
- return allErrs
-}
-
-func ValidateSeccompProfile(p string, fldPath *field.Path) field.ErrorList {
- if p == "docker/default" {
- return nil
- }
- if p == "unconfined" {
- return nil
- }
- if strings.HasPrefix(p, "localhost/") {
- return validateLocalDescendingPath(strings.TrimPrefix(p, "localhost/"), fldPath)
- }
- return field.ErrorList{field.Invalid(fldPath, p, "must be a valid seccomp profile")}
-}
-
-func ValidateSeccompPodAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if p, exists := annotations[core.SeccompPodAnnotationKey]; exists {
- allErrs = append(allErrs, ValidateSeccompProfile(p, fldPath.Child(core.SeccompPodAnnotationKey))...)
- }
- for k, p := range annotations {
- if strings.HasPrefix(k, core.SeccompContainerAnnotationKeyPrefix) {
- allErrs = append(allErrs, ValidateSeccompProfile(p, fldPath.Child(k))...)
- }
- }
-
- return allErrs
-}
-
-func ValidateAppArmorPodAnnotations(annotations map[string]string, spec *core.PodSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for k, p := range annotations {
- if !strings.HasPrefix(k, apparmor.ContainerAnnotationKeyPrefix) {
- continue
- }
- // TODO: this belongs to admission, not general pod validation:
- if !utilfeature.DefaultFeatureGate.Enabled(features.AppArmor) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "AppArmor is disabled by feature-gate"))
- continue
- }
- containerName := strings.TrimPrefix(k, apparmor.ContainerAnnotationKeyPrefix)
- if !podSpecHasContainer(spec, containerName) {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(k), containerName, "container not found"))
- }
-
- if err := apparmor.ValidateProfileFormat(p); err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Key(k), p, err.Error()))
- }
- }
-
- return allErrs
-}
-
-func podSpecHasContainer(spec *core.PodSpec, containerName string) bool {
- for _, c := range spec.InitContainers {
- if c.Name == containerName {
- return true
- }
- }
- for _, c := range spec.Containers {
- if c.Name == containerName {
- return true
- }
- }
- return false
-}
-
-const (
- // a sysctl segment regex, concatenated with dots to form a sysctl name
- SysctlSegmentFmt string = "[a-z0-9]([-_a-z0-9]*[a-z0-9])?"
-
- // a sysctl name regex
- SysctlFmt string = "(" + SysctlSegmentFmt + "\\.)*" + SysctlSegmentFmt
-
- // the maximal length of a sysctl name
- SysctlMaxLength int = 253
-)
-
-var sysctlRegexp = regexp.MustCompile("^" + SysctlFmt + "$")
-
-// IsValidSysctlName checks that the given string is a valid sysctl name,
-// i.e. matches SysctlFmt.
-func IsValidSysctlName(name string) bool {
- if len(name) > SysctlMaxLength {
- return false
- }
- return sysctlRegexp.MatchString(name)
-}
-
-func validateSysctls(sysctls []core.Sysctl, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i, s := range sysctls {
- if len(s.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Index(i).Child("name"), ""))
- } else if !IsValidSysctlName(s.Name) {
- allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("name"), s.Name, fmt.Sprintf("must have at most %d characters and match regex %s", SysctlMaxLength, SysctlFmt)))
- }
- }
- return allErrs
-}
-
-// ValidatePodSecurityContext test that the specified PodSecurityContext has valid data.
-func ValidatePodSecurityContext(securityContext *core.PodSecurityContext, spec *core.PodSpec, specPath, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if securityContext != nil {
- allErrs = append(allErrs, validateHostNetwork(securityContext.HostNetwork, spec.Containers, specPath.Child("containers"))...)
- if securityContext.FSGroup != nil {
- for _, msg := range validation.IsValidGroupID(*securityContext.FSGroup) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("fsGroup"), *(securityContext.FSGroup), msg))
- }
- }
- if securityContext.RunAsUser != nil {
- for _, msg := range validation.IsValidUserID(*securityContext.RunAsUser) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsUser"), *(securityContext.RunAsUser), msg))
- }
- }
- for g, gid := range securityContext.SupplementalGroups {
- for _, msg := range validation.IsValidGroupID(gid) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("supplementalGroups").Index(g), gid, msg))
- }
- }
- if securityContext.ShareProcessNamespace != nil {
- if !utilfeature.DefaultFeatureGate.Enabled(features.PodShareProcessNamespace) {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("shareProcessNamespace"), "Process Namespace Sharing is disabled by PodShareProcessNamespace feature-gate"))
- } else if securityContext.HostPID && *securityContext.ShareProcessNamespace {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("shareProcessNamespace"), *securityContext.ShareProcessNamespace, "ShareProcessNamespace and HostPID cannot both be enabled"))
- }
- }
- }
-
- return allErrs
-}
-
-func ValidateContainerUpdates(newContainers, oldContainers []core.Container, fldPath *field.Path) (allErrs field.ErrorList, stop bool) {
- allErrs = field.ErrorList{}
- if len(newContainers) != len(oldContainers) {
- //TODO: Pinpoint the specific container that causes the invalid error after we have strategic merge diff
- allErrs = append(allErrs, field.Forbidden(fldPath, "pod updates may not add or remove containers"))
- return allErrs, true
- }
-
- // validate updated container images
- for i, ctr := range newContainers {
- if len(ctr.Image) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Index(i).Child("image"), ""))
- }
- // this is only called from ValidatePodUpdate so its safe to check leading/trailing whitespace.
- if len(strings.TrimSpace(ctr.Image)) != len(ctr.Image) {
- allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("image"), ctr.Image, "must not have leading or trailing whitespace"))
- }
- }
- return allErrs, false
-}
-
-// ValidatePodUpdate tests to see if the update is legal for an end user to make. newPod is updated with fields
-// that cannot be changed.
-func ValidatePodUpdate(newPod, oldPod *core.Pod) field.ErrorList {
- fldPath := field.NewPath("metadata")
- allErrs := ValidateObjectMetaUpdate(&newPod.ObjectMeta, &oldPod.ObjectMeta, fldPath)
- allErrs = append(allErrs, ValidatePodSpecificAnnotationUpdates(newPod, oldPod, fldPath.Child("annotations"))...)
- specPath := field.NewPath("spec")
-
- // validate updateable fields:
- // 1. spec.containers[*].image
- // 2. spec.initContainers[*].image
- // 3. spec.activeDeadlineSeconds
-
- containerErrs, stop := ValidateContainerUpdates(newPod.Spec.Containers, oldPod.Spec.Containers, specPath.Child("containers"))
- allErrs = append(allErrs, containerErrs...)
- if stop {
- return allErrs
- }
- containerErrs, stop = ValidateContainerUpdates(newPod.Spec.InitContainers, oldPod.Spec.InitContainers, specPath.Child("initContainers"))
- allErrs = append(allErrs, containerErrs...)
- if stop {
- return allErrs
- }
-
- // validate updated spec.activeDeadlineSeconds. two types of updates are allowed:
- // 1. from nil to a positive value
- // 2. from a positive value to a lesser, non-negative value
- if newPod.Spec.ActiveDeadlineSeconds != nil {
- newActiveDeadlineSeconds := *newPod.Spec.ActiveDeadlineSeconds
- if newActiveDeadlineSeconds < 0 || newActiveDeadlineSeconds > math.MaxInt32 {
- allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newActiveDeadlineSeconds, validation.InclusiveRangeError(0, math.MaxInt32)))
- return allErrs
- }
- if oldPod.Spec.ActiveDeadlineSeconds != nil {
- oldActiveDeadlineSeconds := *oldPod.Spec.ActiveDeadlineSeconds
- if oldActiveDeadlineSeconds < newActiveDeadlineSeconds {
- allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newActiveDeadlineSeconds, "must be less than or equal to previous value"))
- return allErrs
- }
- }
- } else if oldPod.Spec.ActiveDeadlineSeconds != nil {
- allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newPod.Spec.ActiveDeadlineSeconds, "must not update from a positive integer to nil value"))
- }
-
- // handle updateable fields by munging those fields prior to deep equal comparison.
- mungedPod := *newPod
- // munge spec.containers[*].image
- var newContainers []core.Container
- for ix, container := range mungedPod.Spec.Containers {
- container.Image = oldPod.Spec.Containers[ix].Image
- newContainers = append(newContainers, container)
- }
- mungedPod.Spec.Containers = newContainers
- // munge spec.initContainers[*].image
- var newInitContainers []core.Container
- for ix, container := range mungedPod.Spec.InitContainers {
- container.Image = oldPod.Spec.InitContainers[ix].Image
- newInitContainers = append(newInitContainers, container)
- }
- mungedPod.Spec.InitContainers = newInitContainers
- // munge spec.activeDeadlineSeconds
- mungedPod.Spec.ActiveDeadlineSeconds = nil
- if oldPod.Spec.ActiveDeadlineSeconds != nil {
- activeDeadlineSeconds := *oldPod.Spec.ActiveDeadlineSeconds
- mungedPod.Spec.ActiveDeadlineSeconds = &activeDeadlineSeconds
- }
-
- // Allow only additions to tolerations updates.
- mungedPod.Spec.Tolerations = oldPod.Spec.Tolerations
- allErrs = append(allErrs, validateOnlyAddedTolerations(newPod.Spec.Tolerations, oldPod.Spec.Tolerations, specPath.Child("tolerations"))...)
-
- if !apiequality.Semantic.DeepEqual(mungedPod.Spec, oldPod.Spec) {
- // This diff isn't perfect, but it's a helluva lot better an "I'm not going to tell you what the difference is".
- //TODO: Pinpoint the specific field that causes the invalid error after we have strategic merge diff
- specDiff := diff.ObjectDiff(mungedPod.Spec, oldPod.Spec)
- allErrs = append(allErrs, field.Forbidden(specPath, fmt.Sprintf("pod updates may not change fields other than `spec.containers[*].image`, `spec.initContainers[*].image`, `spec.activeDeadlineSeconds` or `spec.tolerations` (only additions to existing tolerations)\n%v", specDiff)))
- }
-
- return allErrs
-}
-
-// ValidateContainerStateTransition test to if any illegal container state transitions are being attempted
-func ValidateContainerStateTransition(newStatuses, oldStatuses []core.ContainerStatus, fldpath *field.Path, restartPolicy core.RestartPolicy) field.ErrorList {
- allErrs := field.ErrorList{}
- // If we should always restart, containers are allowed to leave the terminated state
- if restartPolicy == core.RestartPolicyAlways {
- return allErrs
- }
- for i, oldStatus := range oldStatuses {
- // Skip any container that is not terminated
- if oldStatus.State.Terminated == nil {
- continue
- }
- // Skip any container that failed but is allowed to restart
- if oldStatus.State.Terminated.ExitCode != 0 && restartPolicy == core.RestartPolicyOnFailure {
- continue
- }
- for _, newStatus := range newStatuses {
- if oldStatus.Name == newStatus.Name && newStatus.State.Terminated == nil {
- allErrs = append(allErrs, field.Forbidden(fldpath.Index(i).Child("state"), "may not be transitioned to non-terminated state"))
- }
- }
- }
- return allErrs
-}
-
-// ValidatePodStatusUpdate tests to see if the update is legal for an end user to make. newPod is updated with fields
-// that cannot be changed.
-func ValidatePodStatusUpdate(newPod, oldPod *core.Pod) field.ErrorList {
- fldPath := field.NewPath("metadata")
- allErrs := ValidateObjectMetaUpdate(&newPod.ObjectMeta, &oldPod.ObjectMeta, fldPath)
- allErrs = append(allErrs, ValidatePodSpecificAnnotationUpdates(newPod, oldPod, fldPath.Child("annotations"))...)
-
- fldPath = field.NewPath("status")
- if newPod.Spec.NodeName != oldPod.Spec.NodeName {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("nodeName"), "may not be changed directly"))
- }
-
- if newPod.Status.NominatedNodeName != oldPod.Status.NominatedNodeName && len(newPod.Status.NominatedNodeName) > 0 {
- for _, msg := range ValidateNodeName(newPod.Status.NominatedNodeName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("nominatedNodeName"), newPod.Status.NominatedNodeName, msg))
- }
- }
-
- // If pod should not restart, make sure the status update does not transition
- // any terminated containers to a non-terminated state.
- allErrs = append(allErrs, ValidateContainerStateTransition(newPod.Status.ContainerStatuses, oldPod.Status.ContainerStatuses, fldPath.Child("containerStatuses"), oldPod.Spec.RestartPolicy)...)
- allErrs = append(allErrs, ValidateContainerStateTransition(newPod.Status.InitContainerStatuses, oldPod.Status.InitContainerStatuses, fldPath.Child("initContainerStatuses"), oldPod.Spec.RestartPolicy)...)
-
- // For status update we ignore changes to pod spec.
- newPod.Spec = oldPod.Spec
-
- return allErrs
-}
-
-// ValidatePodBinding tests if required fields in the pod binding are legal.
-func ValidatePodBinding(binding *core.Binding) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if len(binding.Target.Kind) != 0 && binding.Target.Kind != "Node" {
- // TODO: When validation becomes versioned, this gets more complicated.
- allErrs = append(allErrs, field.NotSupported(field.NewPath("target", "kind"), binding.Target.Kind, []string{"Node", "<empty>"}))
- }
- if len(binding.Target.Name) == 0 {
- // TODO: When validation becomes versioned, this gets more complicated.
- allErrs = append(allErrs, field.Required(field.NewPath("target", "name"), ""))
- }
-
- return allErrs
-}
-
-// ValidatePodTemplate tests if required fields in the pod template are set.
-func ValidatePodTemplate(pod *core.PodTemplate) field.ErrorList {
- allErrs := ValidateObjectMeta(&pod.ObjectMeta, true, ValidatePodName, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidatePodTemplateSpec(&pod.Template, field.NewPath("template"))...)
- return allErrs
-}
-
-// ValidatePodTemplateUpdate tests to see if the update is legal for an end user to make. newPod is updated with fields
-// that cannot be changed.
-func ValidatePodTemplateUpdate(newPod, oldPod *core.PodTemplate) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&oldPod.ObjectMeta, &newPod.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidatePodTemplateSpec(&newPod.Template, field.NewPath("template"))...)
- return allErrs
-}
-
-var supportedSessionAffinityType = sets.NewString(string(core.ServiceAffinityClientIP), string(core.ServiceAffinityNone))
-var supportedServiceType = sets.NewString(string(core.ServiceTypeClusterIP), string(core.ServiceTypeNodePort),
- string(core.ServiceTypeLoadBalancer), string(core.ServiceTypeExternalName))
-
-// ValidateService tests if required fields/annotations of a Service are valid.
-func ValidateService(service *core.Service) field.ErrorList {
- allErrs := ValidateObjectMeta(&service.ObjectMeta, true, ValidateServiceName, field.NewPath("metadata"))
-
- specPath := field.NewPath("spec")
- isHeadlessService := service.Spec.ClusterIP == core.ClusterIPNone
- if len(service.Spec.Ports) == 0 && !isHeadlessService && service.Spec.Type != core.ServiceTypeExternalName {
- allErrs = append(allErrs, field.Required(specPath.Child("ports"), ""))
- }
- switch service.Spec.Type {
- case core.ServiceTypeLoadBalancer:
- for ix := range service.Spec.Ports {
- port := &service.Spec.Ports[ix]
- // This is a workaround for broken cloud environments that
- // over-open firewalls. Hopefully it can go away when more clouds
- // understand containers better.
- if port.Port == ports.KubeletPort {
- portPath := specPath.Child("ports").Index(ix)
- allErrs = append(allErrs, field.Invalid(portPath, port.Port, fmt.Sprintf("may not expose port %v externally since it is used by kubelet", ports.KubeletPort)))
- }
- }
- if service.Spec.ClusterIP == "None" {
- allErrs = append(allErrs, field.Invalid(specPath.Child("clusterIP"), service.Spec.ClusterIP, "may not be set to 'None' for LoadBalancer services"))
- }
- case core.ServiceTypeNodePort:
- if service.Spec.ClusterIP == "None" {
- allErrs = append(allErrs, field.Invalid(specPath.Child("clusterIP"), service.Spec.ClusterIP, "may not be set to 'None' for NodePort services"))
- }
- case core.ServiceTypeExternalName:
- if service.Spec.ClusterIP != "" {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("clusterIP"), "must be empty for ExternalName services"))
- }
- if len(service.Spec.ExternalName) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Subdomain(service.Spec.ExternalName, specPath.Child("externalName"))...)
- } else {
- allErrs = append(allErrs, field.Required(specPath.Child("externalName"), ""))
- }
- }
-
- allPortNames := sets.String{}
- portsPath := specPath.Child("ports")
- for i := range service.Spec.Ports {
- portPath := portsPath.Index(i)
- allErrs = append(allErrs, validateServicePort(&service.Spec.Ports[i], len(service.Spec.Ports) > 1, isHeadlessService, &allPortNames, portPath)...)
- }
-
- if service.Spec.Selector != nil {
- allErrs = append(allErrs, unversionedvalidation.ValidateLabels(service.Spec.Selector, specPath.Child("selector"))...)
- }
-
- if len(service.Spec.SessionAffinity) == 0 {
- allErrs = append(allErrs, field.Required(specPath.Child("sessionAffinity"), ""))
- } else if !supportedSessionAffinityType.Has(string(service.Spec.SessionAffinity)) {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("sessionAffinity"), service.Spec.SessionAffinity, supportedSessionAffinityType.List()))
- }
-
- if service.Spec.SessionAffinity == core.ServiceAffinityClientIP {
- allErrs = append(allErrs, validateClientIPAffinityConfig(service.Spec.SessionAffinityConfig, specPath.Child("sessionAffinityConfig"))...)
- } else if service.Spec.SessionAffinity == core.ServiceAffinityNone {
- if service.Spec.SessionAffinityConfig != nil {
- allErrs = append(allErrs, field.Forbidden(specPath.Child("sessionAffinityConfig"), fmt.Sprintf("must not be set when session affinity is %s", string(core.ServiceAffinityNone))))
- }
- }
-
- if helper.IsServiceIPSet(service) {
- if ip := net.ParseIP(service.Spec.ClusterIP); ip == nil {
- allErrs = append(allErrs, field.Invalid(specPath.Child("clusterIP"), service.Spec.ClusterIP, "must be empty, 'None', or a valid IP address"))
- }
- }
-
- ipPath := specPath.Child("externalIPs")
- for i, ip := range service.Spec.ExternalIPs {
- idxPath := ipPath.Index(i)
- if msgs := validation.IsValidIP(ip); len(msgs) != 0 {
- for i := range msgs {
- allErrs = append(allErrs, field.Invalid(idxPath, ip, msgs[i]))
- }
- } else {
- allErrs = append(allErrs, validateNonSpecialIP(ip, idxPath)...)
- }
- }
-
- if len(service.Spec.Type) == 0 {
- allErrs = append(allErrs, field.Required(specPath.Child("type"), ""))
- } else if !supportedServiceType.Has(string(service.Spec.Type)) {
- allErrs = append(allErrs, field.NotSupported(specPath.Child("type"), service.Spec.Type, supportedServiceType.List()))
- }
-
- if service.Spec.Type == core.ServiceTypeLoadBalancer {
- portsPath := specPath.Child("ports")
- includeProtocols := sets.NewString()
- for i := range service.Spec.Ports {
- portPath := portsPath.Index(i)
- if !supportedPortProtocols.Has(string(service.Spec.Ports[i].Protocol)) {
- allErrs = append(allErrs, field.Invalid(portPath.Child("protocol"), service.Spec.Ports[i].Protocol, "cannot create an external load balancer with non-TCP/UDP ports"))
- } else {
- includeProtocols.Insert(string(service.Spec.Ports[i].Protocol))
- }
- }
- if includeProtocols.Len() > 1 {
- allErrs = append(allErrs, field.Invalid(portsPath, service.Spec.Ports, "cannot create an external load balancer with mix protocols"))
- }
- }
-
- if service.Spec.Type == core.ServiceTypeClusterIP {
- portsPath := specPath.Child("ports")
- for i := range service.Spec.Ports {
- portPath := portsPath.Index(i)
- if service.Spec.Ports[i].NodePort != 0 {
- allErrs = append(allErrs, field.Forbidden(portPath.Child("nodePort"), "may not be used when `type` is 'ClusterIP'"))
- }
- }
- }
-
- // Check for duplicate NodePorts, considering (protocol,port) pairs
- portsPath = specPath.Child("ports")
- nodePorts := make(map[core.ServicePort]bool)
- for i := range service.Spec.Ports {
- port := &service.Spec.Ports[i]
- if port.NodePort == 0 {
- continue
- }
- portPath := portsPath.Index(i)
- var key core.ServicePort
- key.Protocol = port.Protocol
- key.NodePort = port.NodePort
- _, found := nodePorts[key]
- if found {
- allErrs = append(allErrs, field.Duplicate(portPath.Child("nodePort"), port.NodePort))
- }
- nodePorts[key] = true
- }
-
- // Check for duplicate Ports, considering (protocol,port) pairs
- portsPath = specPath.Child("ports")
- ports := make(map[core.ServicePort]bool)
- for i, port := range service.Spec.Ports {
- portPath := portsPath.Index(i)
- key := core.ServicePort{Protocol: port.Protocol, Port: port.Port}
- _, found := ports[key]
- if found {
- allErrs = append(allErrs, field.Duplicate(portPath, key))
- }
- ports[key] = true
- }
-
- // Validate SourceRange field and annotation
- _, ok := service.Annotations[core.AnnotationLoadBalancerSourceRangesKey]
- if len(service.Spec.LoadBalancerSourceRanges) > 0 || ok {
- var fieldPath *field.Path
- var val string
- if len(service.Spec.LoadBalancerSourceRanges) > 0 {
- fieldPath = specPath.Child("LoadBalancerSourceRanges")
- val = fmt.Sprintf("%v", service.Spec.LoadBalancerSourceRanges)
- } else {
- fieldPath = field.NewPath("metadata", "annotations").Key(core.AnnotationLoadBalancerSourceRangesKey)
- val = service.Annotations[core.AnnotationLoadBalancerSourceRangesKey]
- }
- if service.Spec.Type != core.ServiceTypeLoadBalancer {
- allErrs = append(allErrs, field.Forbidden(fieldPath, "may only be used when `type` is 'LoadBalancer'"))
- }
- _, err := apiservice.GetLoadBalancerSourceRanges(service)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fieldPath, val, "must be a list of IP ranges. For example, 10.240.0.0/24,10.250.0.0/24 "))
- }
- }
-
- allErrs = append(allErrs, validateServiceExternalTrafficFieldsValue(service)...)
-
- return allErrs
-}
-
-func validateServicePort(sp *core.ServicePort, requireName, isHeadlessService bool, allNames *sets.String, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if requireName && len(sp.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- } else if len(sp.Name) != 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(sp.Name, fldPath.Child("name"))...)
- if allNames.Has(sp.Name) {
- allErrs = append(allErrs, field.Duplicate(fldPath.Child("name"), sp.Name))
- } else {
- allNames.Insert(sp.Name)
- }
- }
-
- for _, msg := range validation.IsValidPortNum(int(sp.Port)) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("port"), sp.Port, msg))
- }
-
- if len(sp.Protocol) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("protocol"), ""))
- } else if !supportedPortProtocols.Has(string(sp.Protocol)) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("protocol"), sp.Protocol, supportedPortProtocols.List()))
- }
-
- allErrs = append(allErrs, ValidatePortNumOrName(sp.TargetPort, fldPath.Child("targetPort"))...)
-
- // in the v1 API, targetPorts on headless services were tolerated.
- // once we have version-specific validation, we can reject this on newer API versions, but until then, we have to tolerate it for compatibility.
- //
- // if isHeadlessService {
- // if sp.TargetPort.Type == intstr.String || (sp.TargetPort.Type == intstr.Int && sp.Port != sp.TargetPort.IntValue()) {
- // allErrs = append(allErrs, field.Invalid(fldPath.Child("targetPort"), sp.TargetPort, "must be equal to the value of 'port' when clusterIP = None"))
- // }
- // }
-
- return allErrs
-}
-
-// validateServiceExternalTrafficFieldsValue validates ExternalTraffic related annotations
-// have legal value.
-func validateServiceExternalTrafficFieldsValue(service *core.Service) field.ErrorList {
- allErrs := field.ErrorList{}
-
- // Check first class fields.
- if service.Spec.ExternalTrafficPolicy != "" &&
- service.Spec.ExternalTrafficPolicy != core.ServiceExternalTrafficPolicyTypeCluster &&
- service.Spec.ExternalTrafficPolicy != core.ServiceExternalTrafficPolicyTypeLocal {
- allErrs = append(allErrs, field.Invalid(field.NewPath("spec").Child("externalTrafficPolicy"), service.Spec.ExternalTrafficPolicy,
- fmt.Sprintf("ExternalTrafficPolicy must be empty, %v or %v", core.ServiceExternalTrafficPolicyTypeCluster, core.ServiceExternalTrafficPolicyTypeLocal)))
- }
- if service.Spec.HealthCheckNodePort < 0 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("spec").Child("healthCheckNodePort"), service.Spec.HealthCheckNodePort,
- "HealthCheckNodePort must be not less than 0"))
- }
-
- return allErrs
-}
-
-// ValidateServiceExternalTrafficFieldsCombination validates if ExternalTrafficPolicy,
-// HealthCheckNodePort and Type combination are legal. For update, it should be called
-// after clearing externalTraffic related fields for the ease of transitioning between
-// different service types.
-func ValidateServiceExternalTrafficFieldsCombination(service *core.Service) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if service.Spec.Type != core.ServiceTypeLoadBalancer &&
- service.Spec.Type != core.ServiceTypeNodePort &&
- service.Spec.ExternalTrafficPolicy != "" {
- allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "externalTrafficPolicy"), service.Spec.ExternalTrafficPolicy,
- "ExternalTrafficPolicy can only be set on NodePort and LoadBalancer service"))
- }
-
- if !apiservice.NeedsHealthCheck(service) &&
- service.Spec.HealthCheckNodePort != 0 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "healthCheckNodePort"), service.Spec.HealthCheckNodePort,
- "HealthCheckNodePort can only be set on LoadBalancer service with ExternalTrafficPolicy=Local"))
- }
-
- return allErrs
-}
-
-// ValidateServiceUpdate tests if required fields in the service are set during an update
-func ValidateServiceUpdate(service, oldService *core.Service) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&service.ObjectMeta, &oldService.ObjectMeta, field.NewPath("metadata"))
-
- // ClusterIP should be immutable for services using it (every type other than ExternalName)
- // which do not have ClusterIP assigned yet (empty string value)
- if service.Spec.Type != core.ServiceTypeExternalName {
- if oldService.Spec.Type != core.ServiceTypeExternalName && oldService.Spec.ClusterIP != "" {
- allErrs = append(allErrs, ValidateImmutableField(service.Spec.ClusterIP, oldService.Spec.ClusterIP, field.NewPath("spec", "clusterIP"))...)
- }
- }
-
- allErrs = append(allErrs, ValidateService(service)...)
- return allErrs
-}
-
-// ValidateServiceStatusUpdate tests if required fields in the Service are set when updating status.
-func ValidateServiceStatusUpdate(service, oldService *core.Service) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&service.ObjectMeta, &oldService.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateLoadBalancerStatus(&service.Status.LoadBalancer, field.NewPath("status", "loadBalancer"))...)
- return allErrs
-}
-
-// ValidateReplicationController tests if required fields in the replication controller are set.
-func ValidateReplicationController(controller *core.ReplicationController) field.ErrorList {
- allErrs := ValidateObjectMeta(&controller.ObjectMeta, true, ValidateReplicationControllerName, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateReplicationControllerSpec(&controller.Spec, field.NewPath("spec"))...)
- return allErrs
-}
-
-// ValidateReplicationControllerUpdate tests if required fields in the replication controller are set.
-func ValidateReplicationControllerUpdate(controller, oldController *core.ReplicationController) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&controller.ObjectMeta, &oldController.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateReplicationControllerSpec(&controller.Spec, field.NewPath("spec"))...)
- return allErrs
-}
-
-// ValidateReplicationControllerStatusUpdate tests if required fields in the replication controller are set.
-func ValidateReplicationControllerStatusUpdate(controller, oldController *core.ReplicationController) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&controller.ObjectMeta, &oldController.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateReplicationControllerStatus(controller.Status, field.NewPath("status"))...)
- return allErrs
-}
-
-func ValidateReplicationControllerStatus(status core.ReplicationControllerStatus, statusPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, ValidateNonnegativeField(int64(status.Replicas), statusPath.Child("replicas"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(status.FullyLabeledReplicas), statusPath.Child("fullyLabeledReplicas"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(status.ReadyReplicas), statusPath.Child("readyReplicas"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(status.AvailableReplicas), statusPath.Child("availableReplicas"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(status.ObservedGeneration), statusPath.Child("observedGeneration"))...)
- msg := "cannot be greater than status.replicas"
- if status.FullyLabeledReplicas > status.Replicas {
- allErrs = append(allErrs, field.Invalid(statusPath.Child("fullyLabeledReplicas"), status.FullyLabeledReplicas, msg))
- }
- if status.ReadyReplicas > status.Replicas {
- allErrs = append(allErrs, field.Invalid(statusPath.Child("readyReplicas"), status.ReadyReplicas, msg))
- }
- if status.AvailableReplicas > status.Replicas {
- allErrs = append(allErrs, field.Invalid(statusPath.Child("availableReplicas"), status.AvailableReplicas, msg))
- }
- if status.AvailableReplicas > status.ReadyReplicas {
- allErrs = append(allErrs, field.Invalid(statusPath.Child("availableReplicas"), status.AvailableReplicas, "cannot be greater than readyReplicas"))
- }
- return allErrs
-}
-
-// Validates that the given selector is non-empty.
-func ValidateNonEmptySelector(selectorMap map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- selector := labels.Set(selectorMap).AsSelector()
- if selector.Empty() {
- allErrs = append(allErrs, field.Required(fldPath, ""))
- }
- return allErrs
-}
-
-// Validates the given template and ensures that it is in accordance with the desired selector and replicas.
-func ValidatePodTemplateSpecForRC(template *core.PodTemplateSpec, selectorMap map[string]string, replicas int32, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if template == nil {
- allErrs = append(allErrs, field.Required(fldPath, ""))
- } else {
- selector := labels.Set(selectorMap).AsSelector()
- if !selector.Empty() {
- // Verify that the RC selector matches the labels in template.
- labels := labels.Set(template.Labels)
- if !selector.Matches(labels) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("metadata", "labels"), template.Labels, "`selector` does not match template `labels`"))
- }
- }
- allErrs = append(allErrs, ValidatePodTemplateSpec(template, fldPath)...)
- if replicas > 1 {
- allErrs = append(allErrs, ValidateReadOnlyPersistentDisks(template.Spec.Volumes, fldPath.Child("spec", "volumes"))...)
- }
- // RestartPolicy has already been first-order validated as per ValidatePodTemplateSpec().
- if template.Spec.RestartPolicy != core.RestartPolicyAlways {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("spec", "restartPolicy"), template.Spec.RestartPolicy, []string{string(core.RestartPolicyAlways)}))
- }
- if template.Spec.ActiveDeadlineSeconds != nil {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("spec", "activeDeadlineSeconds"), template.Spec.ActiveDeadlineSeconds, "must not be specified"))
- }
- }
- return allErrs
-}
-
-// ValidateReplicationControllerSpec tests if required fields in the replication controller spec are set.
-func ValidateReplicationControllerSpec(spec *core.ReplicationControllerSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, ValidateNonnegativeField(int64(spec.MinReadySeconds), fldPath.Child("minReadySeconds"))...)
- allErrs = append(allErrs, ValidateNonEmptySelector(spec.Selector, fldPath.Child("selector"))...)
- allErrs = append(allErrs, ValidateNonnegativeField(int64(spec.Replicas), fldPath.Child("replicas"))...)
- allErrs = append(allErrs, ValidatePodTemplateSpecForRC(spec.Template, spec.Selector, spec.Replicas, fldPath.Child("template"))...)
- return allErrs
-}
-
-// ValidatePodTemplateSpec validates the spec of a pod template
-func ValidatePodTemplateSpec(spec *core.PodTemplateSpec, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, unversionedvalidation.ValidateLabels(spec.Labels, fldPath.Child("labels"))...)
- allErrs = append(allErrs, ValidateAnnotations(spec.Annotations, fldPath.Child("annotations"))...)
- allErrs = append(allErrs, ValidatePodSpecificAnnotations(spec.Annotations, &spec.Spec, fldPath.Child("annotations"))...)
- allErrs = append(allErrs, ValidatePodSpec(&spec.Spec, fldPath.Child("spec"))...)
- return allErrs
-}
-
-func ValidateReadOnlyPersistentDisks(volumes []core.Volume, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i := range volumes {
- vol := &volumes[i]
- idxPath := fldPath.Index(i)
- if vol.GCEPersistentDisk != nil {
- if vol.GCEPersistentDisk.ReadOnly == false {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("gcePersistentDisk", "readOnly"), false, "must be true for replicated pods > 1; GCE PD can only be mounted on multiple machines if it is read-only"))
- }
- }
- // TODO: What to do for AWS? It doesn't support replicas
- }
- return allErrs
-}
-
-// ValidateTaintsInNodeAnnotations tests that the serialized taints in Node.Annotations has valid data
-func ValidateTaintsInNodeAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- taints, err := helper.GetTaintsFromNodeAnnotations(annotations)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath, core.TaintsAnnotationKey, err.Error()))
- return allErrs
- }
-
- if len(taints) > 0 {
- allErrs = append(allErrs, validateNodeTaints(taints, fldPath.Child(core.TaintsAnnotationKey))...)
- }
-
- return allErrs
-}
-
-// validateNodeTaints tests if given taints have valid data.
-func validateNodeTaints(taints []core.Taint, fldPath *field.Path) field.ErrorList {
- allErrors := field.ErrorList{}
-
- uniqueTaints := map[core.TaintEffect]sets.String{}
-
- for i, currTaint := range taints {
- idxPath := fldPath.Index(i)
- // validate the taint key
- allErrors = append(allErrors, unversionedvalidation.ValidateLabelName(currTaint.Key, idxPath.Child("key"))...)
- // validate the taint value
- if errs := validation.IsValidLabelValue(currTaint.Value); len(errs) != 0 {
- allErrors = append(allErrors, field.Invalid(idxPath.Child("value"), currTaint.Value, strings.Join(errs, ";")))
- }
- // validate the taint effect
- allErrors = append(allErrors, validateTaintEffect(&currTaint.Effect, false, idxPath.Child("effect"))...)
-
- // validate if taint is unique by <key, effect>
- if len(uniqueTaints[currTaint.Effect]) > 0 && uniqueTaints[currTaint.Effect].Has(currTaint.Key) {
- duplicatedError := field.Duplicate(idxPath, currTaint)
- duplicatedError.Detail = "taints must be unique by key and effect pair"
- allErrors = append(allErrors, duplicatedError)
- continue
- }
-
- // add taint to existingTaints for uniqueness check
- if len(uniqueTaints[currTaint.Effect]) == 0 {
- uniqueTaints[currTaint.Effect] = sets.String{}
- }
- uniqueTaints[currTaint.Effect].Insert(currTaint.Key)
- }
- return allErrors
-}
-
-func ValidateNodeSpecificAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- if annotations[core.TaintsAnnotationKey] != "" {
- allErrs = append(allErrs, ValidateTaintsInNodeAnnotations(annotations, fldPath)...)
- }
-
- if annotations[core.PreferAvoidPodsAnnotationKey] != "" {
- allErrs = append(allErrs, ValidateAvoidPodsInNodeAnnotations(annotations, fldPath)...)
- }
- return allErrs
-}
-
-// ValidateNode tests if required fields in the node are set.
-func ValidateNode(node *core.Node) field.ErrorList {
- fldPath := field.NewPath("metadata")
- allErrs := ValidateObjectMeta(&node.ObjectMeta, false, ValidateNodeName, fldPath)
- allErrs = append(allErrs, ValidateNodeSpecificAnnotations(node.ObjectMeta.Annotations, fldPath.Child("annotations"))...)
- if len(node.Spec.Taints) > 0 {
- allErrs = append(allErrs, validateNodeTaints(node.Spec.Taints, fldPath.Child("taints"))...)
- }
-
- // Only validate spec.
- // All status fields are optional and can be updated later.
- // That said, if specified, we need to ensure they are valid.
- allErrs = append(allErrs, ValidateNodeResources(node)...)
-
- // external ID is required.
- if len(node.Spec.ExternalID) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("spec", "externalID"), ""))
- }
-
- // Only allow Node.Spec.ConfigSource to be set if the DynamicKubeletConfig feature gate is enabled
- if node.Spec.ConfigSource != nil && !utilfeature.DefaultFeatureGate.Enabled(features.DynamicKubeletConfig) {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "configSource"), "configSource may only be set if the DynamicKubeletConfig feature gate is enabled)"))
- }
-
- if len(node.Spec.PodCIDR) != 0 {
- _, err := ValidateCIDR(node.Spec.PodCIDR)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "podCIDR"), node.Spec.PodCIDR, "not a valid CIDR"))
- }
- }
- return allErrs
-}
-
-// ValidateNodeResources is used to make sure a node has valid capacity and allocatable values.
-func ValidateNodeResources(node *core.Node) field.ErrorList {
- allErrs := field.ErrorList{}
- // Validate resource quantities in capacity.
- hugePageSizes := sets.NewString()
- for k, v := range node.Status.Capacity {
- resPath := field.NewPath("status", "capacity", string(k))
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- // track any huge page size that has a positive value
- if helper.IsHugePageResourceName(k) && v.Value() > int64(0) {
- hugePageSizes.Insert(string(k))
- }
- if len(hugePageSizes) > 1 {
- allErrs = append(allErrs, field.Invalid(resPath, v, "may not have pre-allocated hugepages for multiple page sizes"))
- }
- }
- // Validate resource quantities in allocatable.
- hugePageSizes = sets.NewString()
- for k, v := range node.Status.Allocatable {
- resPath := field.NewPath("status", "allocatable", string(k))
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- // track any huge page size that has a positive value
- if helper.IsHugePageResourceName(k) && v.Value() > int64(0) {
- hugePageSizes.Insert(string(k))
- }
- if len(hugePageSizes) > 1 {
- allErrs = append(allErrs, field.Invalid(resPath, v, "may not have pre-allocated hugepages for multiple page sizes"))
- }
- }
- return allErrs
-}
-
-// ValidateNodeUpdate tests to make sure a node update can be applied. Modifies oldNode.
-func ValidateNodeUpdate(node, oldNode *core.Node) field.ErrorList {
- fldPath := field.NewPath("metadata")
- allErrs := ValidateObjectMetaUpdate(&node.ObjectMeta, &oldNode.ObjectMeta, fldPath)
- allErrs = append(allErrs, ValidateNodeSpecificAnnotations(node.ObjectMeta.Annotations, fldPath.Child("annotations"))...)
-
- // TODO: Enable the code once we have better core object.status update model. Currently,
- // anyone can update node status.
- // if !apiequality.Semantic.DeepEqual(node.Status, core.NodeStatus{}) {
- // allErrs = append(allErrs, field.Invalid("status", node.Status, "must be empty"))
- // }
-
- allErrs = append(allErrs, ValidateNodeResources(node)...)
-
- // Validate no duplicate addresses in node status.
- addresses := make(map[core.NodeAddress]bool)
- for i, address := range node.Status.Addresses {
- if _, ok := addresses[address]; ok {
- allErrs = append(allErrs, field.Duplicate(field.NewPath("status", "addresses").Index(i), address))
- }
- addresses[address] = true
- }
-
- if len(oldNode.Spec.PodCIDR) == 0 {
- // Allow the controller manager to assign a CIDR to a node if it doesn't have one.
- oldNode.Spec.PodCIDR = node.Spec.PodCIDR
- } else {
- if oldNode.Spec.PodCIDR != node.Spec.PodCIDR {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "podCIDR"), "node updates may not change podCIDR except from \"\" to valid"))
- }
- }
-
- // Allow controller manager updating provider ID when not set
- if len(oldNode.Spec.ProviderID) == 0 {
- oldNode.Spec.ProviderID = node.Spec.ProviderID
- } else {
- if oldNode.Spec.ProviderID != node.Spec.ProviderID {
- allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "providerID"), "node updates may not change providerID except from \"\" to valid"))
- }
- }
-
- // TODO: move reset function to its own location
- // Ignore metadata changes now that they have been tested
- oldNode.ObjectMeta = node.ObjectMeta
- // Allow users to update capacity
- oldNode.Status.Capacity = node.Status.Capacity
- // Allow users to unschedule node
- oldNode.Spec.Unschedulable = node.Spec.Unschedulable
- // Clear status
- oldNode.Status = node.Status
-
- // update taints
- if len(node.Spec.Taints) > 0 {
- allErrs = append(allErrs, validateNodeTaints(node.Spec.Taints, fldPath.Child("taints"))...)
- }
- oldNode.Spec.Taints = node.Spec.Taints
-
- // Allow updates to Node.Spec.ConfigSource if DynamicKubeletConfig feature gate is enabled
- if utilfeature.DefaultFeatureGate.Enabled(features.DynamicKubeletConfig) {
- oldNode.Spec.ConfigSource = node.Spec.ConfigSource
- }
-
- // We made allowed changes to oldNode, and now we compare oldNode to node. Any remaining differences indicate changes to protected fields.
- // TODO: Add a 'real' error type for this error and provide print actual diffs.
- if !apiequality.Semantic.DeepEqual(oldNode, node) {
- glog.V(4).Infof("Update failed validation %#v vs %#v", oldNode, node)
- allErrs = append(allErrs, field.Forbidden(field.NewPath(""), "node updates may only change labels, taints, or capacity (or configSource, if the DynamicKubeletConfig feature gate is enabled)"))
- }
-
- return allErrs
-}
-
-// Validate compute resource typename.
-// Refer to docs/design/resources.md for more details.
-func validateResourceName(value string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, msg := range validation.IsQualifiedName(value) {
- allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
- }
- if len(allErrs) != 0 {
- return allErrs
- }
-
- if len(strings.Split(value, "/")) == 1 {
- if !helper.IsStandardResourceName(value) {
- return append(allErrs, field.Invalid(fldPath, value, "must be a standard resource type or fully qualified"))
- }
- }
-
- return allErrs
-}
-
-// Validate container resource name
-// Refer to docs/design/resources.md for more details.
-func validateContainerResourceName(value string, fldPath *field.Path) field.ErrorList {
- allErrs := validateResourceName(value, fldPath)
-
- if len(strings.Split(value, "/")) == 1 {
- if !helper.IsStandardContainerResourceName(value) {
- return append(allErrs, field.Invalid(fldPath, value, "must be a standard resource for containers"))
- }
- } else if !helper.IsDefaultNamespaceResource(core.ResourceName(value)) {
- if !helper.IsExtendedResourceName(core.ResourceName(value)) {
- return append(allErrs, field.Invalid(fldPath, value, "doesn't follow extended resource name standard"))
- }
- }
- return allErrs
-}
-
-// isLocalStorageResource checks whether the resource is local ephemeral storage
-func isLocalStorageResource(name string) bool {
- if name == string(core.ResourceEphemeralStorage) || name == string(core.ResourceRequestsEphemeralStorage) ||
- name == string(core.ResourceLimitsEphemeralStorage) {
- return true
- } else {
- return false
- }
-}
-
-// Validate resource names that can go in a resource quota
-// Refer to docs/design/resources.md for more details.
-func ValidateResourceQuotaResourceName(value string, fldPath *field.Path) field.ErrorList {
- allErrs := validateResourceName(value, fldPath)
- if isLocalStorageResource(value) && !utilfeature.DefaultFeatureGate.Enabled(features.LocalStorageCapacityIsolation) {
- return append(allErrs, field.Forbidden(fldPath, "ResourceEphemeralStorage field disabled by feature-gate for ResourceQuota"))
- }
- if len(strings.Split(value, "/")) == 1 {
- if !helper.IsStandardQuotaResourceName(value) {
- return append(allErrs, field.Invalid(fldPath, value, isInvalidQuotaResource))
- }
- }
- return allErrs
-}
-
-// Validate limit range types
-func validateLimitRangeTypeName(value string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, msg := range validation.IsQualifiedName(value) {
- allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
- }
- if len(allErrs) != 0 {
- return allErrs
- }
-
- if len(strings.Split(value, "/")) == 1 {
- if !helper.IsStandardLimitRangeType(value) {
- return append(allErrs, field.Invalid(fldPath, value, "must be a standard limit type or fully qualified"))
- }
- }
-
- return allErrs
-}
-
-// Validate limit range resource name
-// limit types (other than Pod/Container) could contain storage not just cpu or memory
-func validateLimitRangeResourceName(limitType core.LimitType, value string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if value == string(core.ResourceEphemeralStorage) && !utilfeature.DefaultFeatureGate.Enabled(features.LocalStorageCapacityIsolation) {
- return append(allErrs, field.Forbidden(fldPath, "ResourceEphemeralStorage field disabled by feature-gate for Resource LimitRange"))
- }
- switch limitType {
- case core.LimitTypePod, core.LimitTypeContainer:
- return validateContainerResourceName(value, fldPath)
- default:
- return validateResourceName(value, fldPath)
- }
-}
-
-// ValidateLimitRange tests if required fields in the LimitRange are set.
-func ValidateLimitRange(limitRange *core.LimitRange) field.ErrorList {
- allErrs := ValidateObjectMeta(&limitRange.ObjectMeta, true, ValidateLimitRangeName, field.NewPath("metadata"))
-
- // ensure resource names are properly qualified per docs/design/resources.md
- limitTypeSet := map[core.LimitType]bool{}
- fldPath := field.NewPath("spec", "limits")
- for i := range limitRange.Spec.Limits {
- idxPath := fldPath.Index(i)
- limit := &limitRange.Spec.Limits[i]
- allErrs = append(allErrs, validateLimitRangeTypeName(string(limit.Type), idxPath.Child("type"))...)
-
- _, found := limitTypeSet[limit.Type]
- if found {
- allErrs = append(allErrs, field.Duplicate(idxPath.Child("type"), limit.Type))
- }
- limitTypeSet[limit.Type] = true
-
- keys := sets.String{}
- min := map[string]resource.Quantity{}
- max := map[string]resource.Quantity{}
- defaults := map[string]resource.Quantity{}
- defaultRequests := map[string]resource.Quantity{}
- maxLimitRequestRatios := map[string]resource.Quantity{}
-
- for k, q := range limit.Max {
- allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("max").Key(string(k)))...)
- keys.Insert(string(k))
- max[string(k)] = q
- }
- for k, q := range limit.Min {
- allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("min").Key(string(k)))...)
- keys.Insert(string(k))
- min[string(k)] = q
- }
-
- if limit.Type == core.LimitTypePod {
- if len(limit.Default) > 0 {
- allErrs = append(allErrs, field.Forbidden(idxPath.Child("default"), "may not be specified when `type` is 'Pod'"))
- }
- if len(limit.DefaultRequest) > 0 {
- allErrs = append(allErrs, field.Forbidden(idxPath.Child("defaultRequest"), "may not be specified when `type` is 'Pod'"))
- }
- } else {
- for k, q := range limit.Default {
- allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("default").Key(string(k)))...)
- keys.Insert(string(k))
- defaults[string(k)] = q
- }
- for k, q := range limit.DefaultRequest {
- allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("defaultRequest").Key(string(k)))...)
- keys.Insert(string(k))
- defaultRequests[string(k)] = q
- }
- }
-
- if limit.Type == core.LimitTypePersistentVolumeClaim {
- _, minQuantityFound := limit.Min[core.ResourceStorage]
- _, maxQuantityFound := limit.Max[core.ResourceStorage]
- if !minQuantityFound && !maxQuantityFound {
- allErrs = append(allErrs, field.Required(idxPath.Child("limits"), "either minimum or maximum storage value is required, but neither was provided"))
- }
- }
-
- for k, q := range limit.MaxLimitRequestRatio {
- allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("maxLimitRequestRatio").Key(string(k)))...)
- keys.Insert(string(k))
- maxLimitRequestRatios[string(k)] = q
- }
-
- for k := range keys {
- minQuantity, minQuantityFound := min[k]
- maxQuantity, maxQuantityFound := max[k]
- defaultQuantity, defaultQuantityFound := defaults[k]
- defaultRequestQuantity, defaultRequestQuantityFound := defaultRequests[k]
- maxRatio, maxRatioFound := maxLimitRequestRatios[k]
-
- if minQuantityFound && maxQuantityFound && minQuantity.Cmp(maxQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("min").Key(string(k)), minQuantity, fmt.Sprintf("min value %s is greater than max value %s", minQuantity.String(), maxQuantity.String())))
- }
-
- if defaultRequestQuantityFound && minQuantityFound && minQuantity.Cmp(defaultRequestQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("min value %s is greater than default request value %s", minQuantity.String(), defaultRequestQuantity.String())))
- }
-
- if defaultRequestQuantityFound && maxQuantityFound && defaultRequestQuantity.Cmp(maxQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default request value %s is greater than max value %s", defaultRequestQuantity.String(), maxQuantity.String())))
- }
-
- if defaultRequestQuantityFound && defaultQuantityFound && defaultRequestQuantity.Cmp(defaultQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default request value %s is greater than default limit value %s", defaultRequestQuantity.String(), defaultQuantity.String())))
- }
-
- if defaultQuantityFound && minQuantityFound && minQuantity.Cmp(defaultQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("default").Key(string(k)), minQuantity, fmt.Sprintf("min value %s is greater than default value %s", minQuantity.String(), defaultQuantity.String())))
- }
-
- if defaultQuantityFound && maxQuantityFound && defaultQuantity.Cmp(maxQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("default").Key(string(k)), maxQuantity, fmt.Sprintf("default value %s is greater than max value %s", defaultQuantity.String(), maxQuantity.String())))
- }
- if maxRatioFound && maxRatio.Cmp(*resource.NewQuantity(1, resource.DecimalSI)) < 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("maxLimitRequestRatio").Key(string(k)), maxRatio, fmt.Sprintf("ratio %s is less than 1", maxRatio.String())))
- }
- if maxRatioFound && minQuantityFound && maxQuantityFound {
- maxRatioValue := float64(maxRatio.Value())
- minQuantityValue := minQuantity.Value()
- maxQuantityValue := maxQuantity.Value()
- if maxRatio.Value() < resource.MaxMilliValue && minQuantityValue < resource.MaxMilliValue && maxQuantityValue < resource.MaxMilliValue {
- maxRatioValue = float64(maxRatio.MilliValue()) / 1000
- minQuantityValue = minQuantity.MilliValue()
- maxQuantityValue = maxQuantity.MilliValue()
- }
- maxRatioLimit := float64(maxQuantityValue) / float64(minQuantityValue)
- if maxRatioValue > maxRatioLimit {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("maxLimitRequestRatio").Key(string(k)), maxRatio, fmt.Sprintf("ratio %s is greater than max/min = %f", maxRatio.String(), maxRatioLimit)))
- }
- }
-
- // for GPU, hugepages and other resources that are not allowed to overcommit,
- // the default value and defaultRequest value must match if both are specified
- if !helper.IsOvercommitAllowed(core.ResourceName(k)) && defaultQuantityFound && defaultRequestQuantityFound && defaultQuantity.Cmp(defaultRequestQuantity) != 0 {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default value %s must equal to defaultRequest value %s in %s", defaultQuantity.String(), defaultRequestQuantity.String(), k)))
- }
- }
- }
-
- return allErrs
-}
-
-// ValidateServiceAccount tests if required fields in the ServiceAccount are set.
-func ValidateServiceAccount(serviceAccount *core.ServiceAccount) field.ErrorList {
- allErrs := ValidateObjectMeta(&serviceAccount.ObjectMeta, true, ValidateServiceAccountName, field.NewPath("metadata"))
- return allErrs
-}
-
-// ValidateServiceAccountUpdate tests if required fields in the ServiceAccount are set.
-func ValidateServiceAccountUpdate(newServiceAccount, oldServiceAccount *core.ServiceAccount) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newServiceAccount.ObjectMeta, &oldServiceAccount.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateServiceAccount(newServiceAccount)...)
- return allErrs
-}
-
-// ValidateSecret tests if required fields in the Secret are set.
-func ValidateSecret(secret *core.Secret) field.ErrorList {
- allErrs := ValidateObjectMeta(&secret.ObjectMeta, true, ValidateSecretName, field.NewPath("metadata"))
-
- dataPath := field.NewPath("data")
- totalSize := 0
- for key, value := range secret.Data {
- for _, msg := range validation.IsConfigMapKey(key) {
- allErrs = append(allErrs, field.Invalid(dataPath.Key(key), key, msg))
- }
- totalSize += len(value)
- }
- if totalSize > core.MaxSecretSize {
- allErrs = append(allErrs, field.TooLong(dataPath, "", core.MaxSecretSize))
- }
-
- switch secret.Type {
- case core.SecretTypeServiceAccountToken:
- // Only require Annotations[kubernetes.io/service-account.name]
- // Additional fields (like Annotations[kubernetes.io/service-account.uid] and Data[token]) might be contributed later by a controller loop
- if value := secret.Annotations[core.ServiceAccountNameKey]; len(value) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("metadata", "annotations").Key(core.ServiceAccountNameKey), ""))
- }
- case core.SecretTypeOpaque, "":
- // no-op
- case core.SecretTypeDockercfg:
- dockercfgBytes, exists := secret.Data[core.DockerConfigKey]
- if !exists {
- allErrs = append(allErrs, field.Required(dataPath.Key(core.DockerConfigKey), ""))
- break
- }
-
- // make sure that the content is well-formed json.
- if err := json.Unmarshal(dockercfgBytes, &map[string]interface{}{}); err != nil {
- allErrs = append(allErrs, field.Invalid(dataPath.Key(core.DockerConfigKey), "<secret contents redacted>", err.Error()))
- }
- case core.SecretTypeDockerConfigJson:
- dockerConfigJsonBytes, exists := secret.Data[core.DockerConfigJsonKey]
- if !exists {
- allErrs = append(allErrs, field.Required(dataPath.Key(core.DockerConfigJsonKey), ""))
- break
- }
-
- // make sure that the content is well-formed json.
- if err := json.Unmarshal(dockerConfigJsonBytes, &map[string]interface{}{}); err != nil {
- allErrs = append(allErrs, field.Invalid(dataPath.Key(core.DockerConfigJsonKey), "<secret contents redacted>", err.Error()))
- }
- case core.SecretTypeBasicAuth:
- _, usernameFieldExists := secret.Data[core.BasicAuthUsernameKey]
- _, passwordFieldExists := secret.Data[core.BasicAuthPasswordKey]
-
- // username or password might be empty, but the field must be present
- if !usernameFieldExists && !passwordFieldExists {
- allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.BasicAuthUsernameKey), ""))
- allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.BasicAuthPasswordKey), ""))
- break
- }
- case core.SecretTypeSSHAuth:
- if len(secret.Data[core.SSHAuthPrivateKey]) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.SSHAuthPrivateKey), ""))
- break
- }
-
- case core.SecretTypeTLS:
- if _, exists := secret.Data[core.TLSCertKey]; !exists {
- allErrs = append(allErrs, field.Required(dataPath.Key(core.TLSCertKey), ""))
- }
- if _, exists := secret.Data[core.TLSPrivateKeyKey]; !exists {
- allErrs = append(allErrs, field.Required(dataPath.Key(core.TLSPrivateKeyKey), ""))
- }
- // TODO: Verify that the key matches the cert.
- default:
- // no-op
- }
-
- return allErrs
-}
-
-// ValidateSecretUpdate tests if required fields in the Secret are set.
-func ValidateSecretUpdate(newSecret, oldSecret *core.Secret) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newSecret.ObjectMeta, &oldSecret.ObjectMeta, field.NewPath("metadata"))
-
- if len(newSecret.Type) == 0 {
- newSecret.Type = oldSecret.Type
- }
-
- allErrs = append(allErrs, ValidateImmutableField(newSecret.Type, oldSecret.Type, field.NewPath("type"))...)
-
- allErrs = append(allErrs, ValidateSecret(newSecret)...)
- return allErrs
-}
-
-// ValidateConfigMapName can be used to check whether the given ConfigMap name is valid.
-// Prefix indicates this name will be used as part of generation, in which case
-// trailing dashes are allowed.
-var ValidateConfigMapName = NameIsDNSSubdomain
-
-// ValidateConfigMap tests whether required fields in the ConfigMap are set.
-func ValidateConfigMap(cfg *core.ConfigMap) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, ValidateObjectMeta(&cfg.ObjectMeta, true, ValidateConfigMapName, field.NewPath("metadata"))...)
-
- totalSize := 0
-
- for key, value := range cfg.Data {
- for _, msg := range validation.IsConfigMapKey(key) {
- allErrs = append(allErrs, field.Invalid(field.NewPath("data").Key(key), key, msg))
- }
- // check if we have a duplicate key in the other bag
- if _, isValue := cfg.BinaryData[key]; isValue {
- msg := "duplicate of key present in binaryData"
- allErrs = append(allErrs, field.Invalid(field.NewPath("data").Key(key), key, msg))
- }
- totalSize += len(value)
- }
- for key, value := range cfg.BinaryData {
- for _, msg := range validation.IsConfigMapKey(key) {
- allErrs = append(allErrs, field.Invalid(field.NewPath("binaryData").Key(key), key, msg))
- }
- totalSize += len(value)
- }
- if totalSize > core.MaxSecretSize {
- // pass back "" to indicate that the error refers to the whole object.
- allErrs = append(allErrs, field.TooLong(field.NewPath(""), cfg, core.MaxSecretSize))
- }
-
- return allErrs
-}
-
-// ValidateConfigMapUpdate tests if required fields in the ConfigMap are set.
-func ValidateConfigMapUpdate(newCfg, oldCfg *core.ConfigMap) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, ValidateObjectMetaUpdate(&newCfg.ObjectMeta, &oldCfg.ObjectMeta, field.NewPath("metadata"))...)
- allErrs = append(allErrs, ValidateConfigMap(newCfg)...)
-
- return allErrs
-}
-
-func validateBasicResource(quantity resource.Quantity, fldPath *field.Path) field.ErrorList {
- if quantity.Value() < 0 {
- return field.ErrorList{field.Invalid(fldPath, quantity.Value(), "must be a valid resource quantity")}
- }
- return field.ErrorList{}
-}
-
-// Validates resource requirement spec.
-func ValidateResourceRequirements(requirements *core.ResourceRequirements, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- limPath := fldPath.Child("limits")
- reqPath := fldPath.Child("requests")
- limContainsCpuOrMemory := false
- reqContainsCpuOrMemory := false
- limContainsHugePages := false
- reqContainsHugePages := false
- supportedQoSComputeResources := sets.NewString(string(core.ResourceCPU), string(core.ResourceMemory))
- for resourceName, quantity := range requirements.Limits {
-
- fldPath := limPath.Key(string(resourceName))
- // Validate resource name.
- allErrs = append(allErrs, validateContainerResourceName(string(resourceName), fldPath)...)
-
- // Validate resource quantity.
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(resourceName), quantity, fldPath)...)
-
- if resourceName == core.ResourceEphemeralStorage && !utilfeature.DefaultFeatureGate.Enabled(features.LocalStorageCapacityIsolation) {
- allErrs = append(allErrs, field.Forbidden(limPath, "ResourceEphemeralStorage field disabled by feature-gate for ResourceRequirements"))
- }
- if helper.IsHugePageResourceName(resourceName) {
- if !utilfeature.DefaultFeatureGate.Enabled(features.HugePages) {
- allErrs = append(allErrs, field.Forbidden(limPath, fmt.Sprintf("%s field disabled by feature-gate for ResourceRequirements", resourceName)))
- } else {
- limContainsHugePages = true
- }
- }
-
- if supportedQoSComputeResources.Has(string(resourceName)) {
- limContainsCpuOrMemory = true
- }
- }
- for resourceName, quantity := range requirements.Requests {
- fldPath := reqPath.Key(string(resourceName))
- // Validate resource name.
- allErrs = append(allErrs, validateContainerResourceName(string(resourceName), fldPath)...)
- // Validate resource quantity.
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(resourceName), quantity, fldPath)...)
-
- // Check that request <= limit.
- limitQuantity, exists := requirements.Limits[resourceName]
- if exists {
- // For non overcommitable resources, not only requests can't exceed limits, they also can't be lower, i.e. must be equal.
- if quantity.Cmp(limitQuantity) != 0 && !helper.IsOvercommitAllowed(resourceName) {
- allErrs = append(allErrs, field.Invalid(reqPath, quantity.String(), fmt.Sprintf("must be equal to %s limit", resourceName)))
- } else if quantity.Cmp(limitQuantity) > 0 {
- allErrs = append(allErrs, field.Invalid(reqPath, quantity.String(), fmt.Sprintf("must be less than or equal to %s limit", resourceName)))
- }
- } else if !helper.IsOvercommitAllowed(resourceName) {
- allErrs = append(allErrs, field.Required(limPath, "Limit must be set for non overcommitable resources"))
- }
- if helper.IsHugePageResourceName(resourceName) {
- reqContainsHugePages = true
- }
- if supportedQoSComputeResources.Has(string(resourceName)) {
- reqContainsCpuOrMemory = true
- }
-
- }
- if !limContainsCpuOrMemory && !reqContainsCpuOrMemory && (reqContainsHugePages || limContainsHugePages) {
- allErrs = append(allErrs, field.Forbidden(fldPath, fmt.Sprintf("HugePages require cpu or memory")))
- }
-
- return allErrs
-}
-
-// validateResourceQuotaScopes ensures that each enumerated hard resource constraint is valid for set of scopes
-func validateResourceQuotaScopes(resourceQuotaSpec *core.ResourceQuotaSpec, fld *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(resourceQuotaSpec.Scopes) == 0 {
- return allErrs
- }
- hardLimits := sets.NewString()
- for k := range resourceQuotaSpec.Hard {
- hardLimits.Insert(string(k))
- }
- fldPath := fld.Child("scopes")
- scopeSet := sets.NewString()
- for _, scope := range resourceQuotaSpec.Scopes {
- if !helper.IsStandardResourceQuotaScope(string(scope)) {
- allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "unsupported scope"))
- }
- for _, k := range hardLimits.List() {
- if helper.IsStandardQuotaResourceName(k) && !helper.IsResourceQuotaScopeValidForResource(scope, k) {
- allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "unsupported scope applied to resource"))
- }
- }
- scopeSet.Insert(string(scope))
- }
- invalidScopePairs := []sets.String{
- sets.NewString(string(core.ResourceQuotaScopeBestEffort), string(core.ResourceQuotaScopeNotBestEffort)),
- sets.NewString(string(core.ResourceQuotaScopeTerminating), string(core.ResourceQuotaScopeNotTerminating)),
- }
- for _, invalidScopePair := range invalidScopePairs {
- if scopeSet.HasAll(invalidScopePair.List()...) {
- allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "conflicting scopes"))
- }
- }
- return allErrs
-}
-
-// ValidateResourceQuota tests if required fields in the ResourceQuota are set.
-func ValidateResourceQuota(resourceQuota *core.ResourceQuota) field.ErrorList {
- allErrs := ValidateObjectMeta(&resourceQuota.ObjectMeta, true, ValidateResourceQuotaName, field.NewPath("metadata"))
-
- allErrs = append(allErrs, ValidateResourceQuotaSpec(&resourceQuota.Spec, field.NewPath("spec"))...)
- allErrs = append(allErrs, ValidateResourceQuotaStatus(&resourceQuota.Status, field.NewPath("status"))...)
-
- return allErrs
-}
-
-func ValidateResourceQuotaStatus(status *core.ResourceQuotaStatus, fld *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- fldPath := fld.Child("hard")
- for k, v := range status.Hard {
- resPath := fldPath.Key(string(k))
- allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- }
- fldPath = fld.Child("used")
- for k, v := range status.Used {
- resPath := fldPath.Key(string(k))
- allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- }
-
- return allErrs
-}
-
-func ValidateResourceQuotaSpec(resourceQuotaSpec *core.ResourceQuotaSpec, fld *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
-
- fldPath := fld.Child("hard")
- for k, v := range resourceQuotaSpec.Hard {
- resPath := fldPath.Key(string(k))
- allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- }
- allErrs = append(allErrs, validateResourceQuotaScopes(resourceQuotaSpec, fld)...)
-
- return allErrs
-}
-
-// ValidateResourceQuantityValue enforces that specified quantity is valid for specified resource
-func ValidateResourceQuantityValue(resource string, value resource.Quantity, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- allErrs = append(allErrs, ValidateNonnegativeQuantity(value, fldPath)...)
- if helper.IsIntegerResourceName(resource) {
- if value.MilliValue()%int64(1000) != int64(0) {
- allErrs = append(allErrs, field.Invalid(fldPath, value, isNotIntegerErrorMsg))
- }
- }
- return allErrs
-}
-
-// ValidateResourceQuotaUpdate tests to see if the update is legal for an end user to make.
-// newResourceQuota is updated with fields that cannot be changed.
-func ValidateResourceQuotaUpdate(newResourceQuota, oldResourceQuota *core.ResourceQuota) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newResourceQuota.ObjectMeta, &oldResourceQuota.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateResourceQuotaSpec(&newResourceQuota.Spec, field.NewPath("spec"))...)
-
- // ensure scopes cannot change, and that resources are still valid for scope
- fldPath := field.NewPath("spec", "scopes")
- oldScopes := sets.NewString()
- newScopes := sets.NewString()
- for _, scope := range newResourceQuota.Spec.Scopes {
- newScopes.Insert(string(scope))
- }
- for _, scope := range oldResourceQuota.Spec.Scopes {
- oldScopes.Insert(string(scope))
- }
- if !oldScopes.Equal(newScopes) {
- allErrs = append(allErrs, field.Invalid(fldPath, newResourceQuota.Spec.Scopes, fieldImmutableErrorMsg))
- }
-
- newResourceQuota.Status = oldResourceQuota.Status
- return allErrs
-}
-
-// ValidateResourceQuotaStatusUpdate tests to see if the status update is legal for an end user to make.
-// newResourceQuota is updated with fields that cannot be changed.
-func ValidateResourceQuotaStatusUpdate(newResourceQuota, oldResourceQuota *core.ResourceQuota) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newResourceQuota.ObjectMeta, &oldResourceQuota.ObjectMeta, field.NewPath("metadata"))
- if len(newResourceQuota.ResourceVersion) == 0 {
- allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
- }
- fldPath := field.NewPath("status", "hard")
- for k, v := range newResourceQuota.Status.Hard {
- resPath := fldPath.Key(string(k))
- allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- }
- fldPath = field.NewPath("status", "used")
- for k, v := range newResourceQuota.Status.Used {
- resPath := fldPath.Key(string(k))
- allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
- allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
- }
- newResourceQuota.Spec = oldResourceQuota.Spec
- return allErrs
-}
-
-// ValidateNamespace tests if required fields are set.
-func ValidateNamespace(namespace *core.Namespace) field.ErrorList {
- allErrs := ValidateObjectMeta(&namespace.ObjectMeta, false, ValidateNamespaceName, field.NewPath("metadata"))
- for i := range namespace.Spec.Finalizers {
- allErrs = append(allErrs, validateFinalizerName(string(namespace.Spec.Finalizers[i]), field.NewPath("spec", "finalizers"))...)
- }
- return allErrs
-}
-
-// Validate finalizer names
-func validateFinalizerName(stringValue string, fldPath *field.Path) field.ErrorList {
- allErrs := apimachineryvalidation.ValidateFinalizerName(stringValue, fldPath)
- for _, err := range validateKubeFinalizerName(stringValue, fldPath) {
- allErrs = append(allErrs, err)
- }
-
- return allErrs
-}
-
-// validateKubeFinalizerName checks for "standard" names of legacy finalizer
-func validateKubeFinalizerName(stringValue string, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if len(strings.Split(stringValue, "/")) == 1 {
- if !helper.IsStandardFinalizerName(stringValue) {
- return append(allErrs, field.Invalid(fldPath, stringValue, "name is neither a standard finalizer name nor is it fully qualified"))
- }
- }
-
- return allErrs
-}
-
-// ValidateNamespaceUpdate tests to make sure a namespace update can be applied.
-// newNamespace is updated with fields that cannot be changed
-func ValidateNamespaceUpdate(newNamespace *core.Namespace, oldNamespace *core.Namespace) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
- newNamespace.Spec.Finalizers = oldNamespace.Spec.Finalizers
- newNamespace.Status = oldNamespace.Status
- return allErrs
-}
-
-// ValidateNamespaceStatusUpdate tests to see if the update is legal for an end user to make. newNamespace is updated with fields
-// that cannot be changed.
-func ValidateNamespaceStatusUpdate(newNamespace, oldNamespace *core.Namespace) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
- newNamespace.Spec = oldNamespace.Spec
- if newNamespace.DeletionTimestamp.IsZero() {
- if newNamespace.Status.Phase != core.NamespaceActive {
- allErrs = append(allErrs, field.Invalid(field.NewPath("status", "Phase"), newNamespace.Status.Phase, "may only be 'Active' if `deletionTimestamp` is empty"))
- }
- } else {
- if newNamespace.Status.Phase != core.NamespaceTerminating {
- allErrs = append(allErrs, field.Invalid(field.NewPath("status", "Phase"), newNamespace.Status.Phase, "may only be 'Terminating' if `deletionTimestamp` is not empty"))
- }
- }
- return allErrs
-}
-
-// ValidateNamespaceFinalizeUpdate tests to see if the update is legal for an end user to make.
-// newNamespace is updated with fields that cannot be changed.
-func ValidateNamespaceFinalizeUpdate(newNamespace, oldNamespace *core.Namespace) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
-
- fldPath := field.NewPath("spec", "finalizers")
- for i := range newNamespace.Spec.Finalizers {
- idxPath := fldPath.Index(i)
- allErrs = append(allErrs, validateFinalizerName(string(newNamespace.Spec.Finalizers[i]), idxPath)...)
- }
- newNamespace.Status = oldNamespace.Status
- return allErrs
-}
-
-// Construct lookup map of old subset IPs to NodeNames.
-func updateEpAddrToNodeNameMap(ipToNodeName map[string]string, addresses []core.EndpointAddress) {
- for n := range addresses {
- if addresses[n].NodeName == nil {
- continue
- }
- ipToNodeName[addresses[n].IP] = *addresses[n].NodeName
- }
-}
-
-// Build a map across all subsets of IP -> NodeName
-func buildEndpointAddressNodeNameMap(subsets []core.EndpointSubset) map[string]string {
- ipToNodeName := make(map[string]string)
- for i := range subsets {
- updateEpAddrToNodeNameMap(ipToNodeName, subsets[i].Addresses)
- updateEpAddrToNodeNameMap(ipToNodeName, subsets[i].NotReadyAddresses)
- }
- return ipToNodeName
-}
-
-func validateEpAddrNodeNameTransition(addr *core.EndpointAddress, ipToNodeName map[string]string, fldPath *field.Path) field.ErrorList {
- errList := field.ErrorList{}
- existingNodeName, found := ipToNodeName[addr.IP]
- if !found {
- return errList
- }
- if addr.NodeName == nil || *addr.NodeName == existingNodeName {
- return errList
- }
- // NodeName entry found for this endpoint IP, but user is attempting to change NodeName
- return append(errList, field.Forbidden(fldPath, fmt.Sprintf("Cannot change NodeName for %s to %s", addr.IP, *addr.NodeName)))
-}
-
-// ValidateEndpoints tests if required fields are set.
-func ValidateEndpoints(endpoints *core.Endpoints) field.ErrorList {
- allErrs := ValidateObjectMeta(&endpoints.ObjectMeta, true, ValidateEndpointsName, field.NewPath("metadata"))
- allErrs = append(allErrs, ValidateEndpointsSpecificAnnotations(endpoints.Annotations, field.NewPath("annotations"))...)
- allErrs = append(allErrs, validateEndpointSubsets(endpoints.Subsets, []core.EndpointSubset{}, field.NewPath("subsets"))...)
- return allErrs
-}
-
-func validateEndpointSubsets(subsets []core.EndpointSubset, oldSubsets []core.EndpointSubset, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- ipToNodeName := buildEndpointAddressNodeNameMap(oldSubsets)
- for i := range subsets {
- ss := &subsets[i]
- idxPath := fldPath.Index(i)
-
- // EndpointSubsets must include endpoint address. For headless service, we allow its endpoints not to have ports.
- if len(ss.Addresses) == 0 && len(ss.NotReadyAddresses) == 0 {
- //TODO: consider adding a RequiredOneOf() error for this and similar cases
- allErrs = append(allErrs, field.Required(idxPath, "must specify `addresses` or `notReadyAddresses`"))
- }
- for addr := range ss.Addresses {
- allErrs = append(allErrs, validateEndpointAddress(&ss.Addresses[addr], idxPath.Child("addresses").Index(addr), ipToNodeName)...)
- }
- for addr := range ss.NotReadyAddresses {
- allErrs = append(allErrs, validateEndpointAddress(&ss.NotReadyAddresses[addr], idxPath.Child("notReadyAddresses").Index(addr), ipToNodeName)...)
- }
- for port := range ss.Ports {
- allErrs = append(allErrs, validateEndpointPort(&ss.Ports[port], len(ss.Ports) > 1, idxPath.Child("ports").Index(port))...)
- }
- }
-
- return allErrs
-}
-
-func validateEndpointAddress(address *core.EndpointAddress, fldPath *field.Path, ipToNodeName map[string]string) field.ErrorList {
- allErrs := field.ErrorList{}
- for _, msg := range validation.IsValidIP(address.IP) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("ip"), address.IP, msg))
- }
- if len(address.Hostname) > 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(address.Hostname, fldPath.Child("hostname"))...)
- }
- // During endpoint update, verify that NodeName is a DNS subdomain and transition rules allow the update
- if address.NodeName != nil {
- for _, msg := range ValidateNodeName(*address.NodeName, false) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("nodeName"), *address.NodeName, msg))
- }
- }
- allErrs = append(allErrs, validateEpAddrNodeNameTransition(address, ipToNodeName, fldPath.Child("nodeName"))...)
- if len(allErrs) > 0 {
- return allErrs
- }
- allErrs = append(allErrs, validateNonSpecialIP(address.IP, fldPath.Child("ip"))...)
- return allErrs
-}
-
-func validateNonSpecialIP(ipAddress string, fldPath *field.Path) field.ErrorList {
- // We disallow some IPs as endpoints or external-ips. Specifically,
- // unspecified and loopback addresses are nonsensical and link-local
- // addresses tend to be used for node-centric purposes (e.g. metadata
- // service).
- allErrs := field.ErrorList{}
- ip := net.ParseIP(ipAddress)
- if ip == nil {
- allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "must be a valid IP address"))
- return allErrs
- }
- if ip.IsUnspecified() {
- allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be unspecified (0.0.0.0)"))
- }
- if ip.IsLoopback() {
- allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the loopback range (127.0.0.0/8)"))
- }
- if ip.IsLinkLocalUnicast() {
- allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the link-local range (169.254.0.0/16)"))
- }
- if ip.IsLinkLocalMulticast() {
- allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the link-local multicast range (224.0.0.0/24)"))
- }
- return allErrs
-}
-
-func validateEndpointPort(port *core.EndpointPort, requireName bool, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- if requireName && len(port.Name) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
- } else if len(port.Name) != 0 {
- allErrs = append(allErrs, ValidateDNS1123Label(port.Name, fldPath.Child("name"))...)
- }
- for _, msg := range validation.IsValidPortNum(int(port.Port)) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("port"), port.Port, msg))
- }
- if len(port.Protocol) == 0 {
- allErrs = append(allErrs, field.Required(fldPath.Child("protocol"), ""))
- } else if !supportedPortProtocols.Has(string(port.Protocol)) {
- allErrs = append(allErrs, field.NotSupported(fldPath.Child("protocol"), port.Protocol, supportedPortProtocols.List()))
- }
- return allErrs
-}
-
-// ValidateEndpointsUpdate tests to make sure an endpoints update can be applied.
-func ValidateEndpointsUpdate(newEndpoints, oldEndpoints *core.Endpoints) field.ErrorList {
- allErrs := ValidateObjectMetaUpdate(&newEndpoints.ObjectMeta, &oldEndpoints.ObjectMeta, field.NewPath("metadata"))
- allErrs = append(allErrs, validateEndpointSubsets(newEndpoints.Subsets, oldEndpoints.Subsets, field.NewPath("subsets"))...)
- allErrs = append(allErrs, ValidateEndpointsSpecificAnnotations(newEndpoints.Annotations, field.NewPath("annotations"))...)
- return allErrs
-}
-
-// ValidateSecurityContext ensure the security context contains valid settings
-func ValidateSecurityContext(sc *core.SecurityContext, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- //this should only be true for testing since SecurityContext is defaulted by the core
- if sc == nil {
- return allErrs
- }
-
- if sc.Privileged != nil {
- if *sc.Privileged && !capabilities.Get().AllowPrivileged {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("privileged"), "disallowed by cluster policy"))
- }
- }
-
- if sc.RunAsUser != nil {
- for _, msg := range validation.IsValidUserID(*sc.RunAsUser) {
- allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsUser"), *sc.RunAsUser, msg))
- }
- }
-
- if sc.AllowPrivilegeEscalation != nil && !*sc.AllowPrivilegeEscalation {
- if sc.Privileged != nil && *sc.Privileged {
- allErrs = append(allErrs, field.Invalid(fldPath, sc, "cannot set `allowPrivilegeEscalation` to false and `privileged` to true"))
- }
-
- if sc.Capabilities != nil {
- for _, cap := range sc.Capabilities.Add {
- if string(cap) == "CAP_SYS_ADMIN" {
- allErrs = append(allErrs, field.Invalid(fldPath, sc, "cannot set `allowPrivilegeEscalation` to false and `capabilities.Add` CAP_SYS_ADMIN"))
- }
- }
- }
- }
-
- return allErrs
-}
-
-func ValidatePodLogOptions(opts *core.PodLogOptions) field.ErrorList {
- allErrs := field.ErrorList{}
- if opts.TailLines != nil && *opts.TailLines < 0 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("tailLines"), *opts.TailLines, isNegativeErrorMsg))
- }
- if opts.LimitBytes != nil && *opts.LimitBytes < 1 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("limitBytes"), *opts.LimitBytes, "must be greater than 0"))
- }
- switch {
- case opts.SinceSeconds != nil && opts.SinceTime != nil:
- allErrs = append(allErrs, field.Forbidden(field.NewPath(""), "at most one of `sinceTime` or `sinceSeconds` may be specified"))
- case opts.SinceSeconds != nil:
- if *opts.SinceSeconds < 1 {
- allErrs = append(allErrs, field.Invalid(field.NewPath("sinceSeconds"), *opts.SinceSeconds, "must be greater than 0"))
- }
- }
- return allErrs
-}
-
-// ValidateLoadBalancerStatus validates required fields on a LoadBalancerStatus
-func ValidateLoadBalancerStatus(status *core.LoadBalancerStatus, fldPath *field.Path) field.ErrorList {
- allErrs := field.ErrorList{}
- for i, ingress := range status.Ingress {
- idxPath := fldPath.Child("ingress").Index(i)
- if len(ingress.IP) > 0 {
- if isIP := (net.ParseIP(ingress.IP) != nil); !isIP {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("ip"), ingress.IP, "must be a valid IP address"))
- }
- }
- if len(ingress.Hostname) > 0 {
- for _, msg := range validation.IsDNS1123Subdomain(ingress.Hostname) {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("hostname"), ingress.Hostname, msg))
- }
- if isIP := (net.ParseIP(ingress.Hostname) != nil); isIP {
- allErrs = append(allErrs, field.Invalid(idxPath.Child("hostname"), ingress.Hostname, "must be a DNS name, not an IP address"))
- }
- }
- }
- return allErrs
-}
-
-func sysctlIntersection(a []core.Sysctl, b []core.Sysctl) []string {
- lookup := make(map[string]struct{}, len(a))
- result := []string{}
- for i := range a {
- lookup[a[i].Name] = struct{}{}
- }
- for i := range b {
- if _, found := lookup[b[i].Name]; found {
- result = append(result, b[i].Name)
- }
- }
- return result
-}
-
-// validateStorageNodeAffinityAnnotation tests that the serialized TopologyConstraints in PersistentVolume.Annotations has valid data
-func validateStorageNodeAffinityAnnotation(annotations map[string]string, fldPath *field.Path) (bool, field.ErrorList) {
- allErrs := field.ErrorList{}
-
- na, err := helper.GetStorageNodeAffinityFromAnnotation(annotations)
- if err != nil {
- allErrs = append(allErrs, field.Invalid(fldPath, core.AlphaStorageNodeAffinityAnnotation, err.Error()))
- return false, allErrs
- }
- if na == nil {
- return false, allErrs
- }
-
- if !utilfeature.DefaultFeatureGate.Enabled(features.VolumeScheduling) {
- allErrs = append(allErrs, field.Forbidden(fldPath, "Storage node affinity is disabled by feature-gate"))
- }
-
- policySpecified := false
- if na.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- allErrs = append(allErrs, ValidateNodeSelector(na.RequiredDuringSchedulingIgnoredDuringExecution, fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
- policySpecified = true
- }
-
- if len(na.PreferredDuringSchedulingIgnoredDuringExecution) > 0 {
- allErrs = append(allErrs, field.Forbidden(fldPath.Child("preferredDuringSchedulingIgnoredDuringExection"), "Storage node affinity does not support preferredDuringSchedulingIgnoredDuringExecution"))
- }
- return policySpecified, allErrs
-}
-
-// validateVolumeNodeAffinity tests that the PersistentVolume.NodeAffinity has valid data
-// returns:
-// - true if volumeNodeAffinity is set
-// - errorList if there are validation errors
-func validateVolumeNodeAffinity(nodeAffinity *core.VolumeNodeAffinity, fldPath *field.Path) (bool, field.ErrorList) {
- allErrs := field.ErrorList{}
-
- if nodeAffinity == nil {
- return false, allErrs
- }
-
- if !utilfeature.DefaultFeatureGate.Enabled(features.VolumeScheduling) {
- allErrs = append(allErrs, field.Forbidden(fldPath, "Volume node affinity is disabled by feature-gate"))
- }
-
- if nodeAffinity.Required != nil {
- allErrs = append(allErrs, ValidateNodeSelector(nodeAffinity.Required, fldPath.Child("required"))...)
- } else {
- allErrs = append(allErrs, field.Required(fldPath.Child("required"), "must specify required node constraints"))
- }
-
- return true, allErrs
-}
-
-// ValidateCIDR validates whether a CIDR matches the conventions expected by net.ParseCIDR
-func ValidateCIDR(cidr string) (*net.IPNet, error) {
- _, net, err := net.ParseCIDR(cidr)
- if err != nil {
- return nil, err
- }
- return net, nil
-}
-
-func IsDecremented(update, old *int32) bool {
- if update == nil && old != nil {
- return true
- }
- if update == nil || old == nil {
- return false
- }
- return *update < *old
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/core/zz_generated.deepcopy.go b/vendor/k8s.io/kubernetes/pkg/apis/core/zz_generated.deepcopy.go
deleted file mode 100644
index f0aa3b426..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/core/zz_generated.deepcopy.go
+++ /dev/null
@@ -1,5952 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by deepcopy-gen. DO NOT EDIT.
-
-package core
-
-import (
- v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
- types "k8s.io/apimachinery/pkg/types"
-)
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AWSElasticBlockStoreVolumeSource) DeepCopyInto(out *AWSElasticBlockStoreVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AWSElasticBlockStoreVolumeSource.
-func (in *AWSElasticBlockStoreVolumeSource) DeepCopy() *AWSElasticBlockStoreVolumeSource {
- if in == nil {
- return nil
- }
- out := new(AWSElasticBlockStoreVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Affinity) DeepCopyInto(out *Affinity) {
- *out = *in
- if in.NodeAffinity != nil {
- in, out := &in.NodeAffinity, &out.NodeAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(NodeAffinity)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PodAffinity != nil {
- in, out := &in.PodAffinity, &out.PodAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodAffinity)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PodAntiAffinity != nil {
- in, out := &in.PodAntiAffinity, &out.PodAntiAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodAntiAffinity)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Affinity.
-func (in *Affinity) DeepCopy() *Affinity {
- if in == nil {
- return nil
- }
- out := new(Affinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AttachedVolume) DeepCopyInto(out *AttachedVolume) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AttachedVolume.
-func (in *AttachedVolume) DeepCopy() *AttachedVolume {
- if in == nil {
- return nil
- }
- out := new(AttachedVolume)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AvoidPods) DeepCopyInto(out *AvoidPods) {
- *out = *in
- if in.PreferAvoidPods != nil {
- in, out := &in.PreferAvoidPods, &out.PreferAvoidPods
- *out = make([]PreferAvoidPodsEntry, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AvoidPods.
-func (in *AvoidPods) DeepCopy() *AvoidPods {
- if in == nil {
- return nil
- }
- out := new(AvoidPods)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AzureDiskVolumeSource) DeepCopyInto(out *AzureDiskVolumeSource) {
- *out = *in
- if in.CachingMode != nil {
- in, out := &in.CachingMode, &out.CachingMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureDataDiskCachingMode)
- **out = **in
- }
- }
- if in.FSType != nil {
- in, out := &in.FSType, &out.FSType
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- if in.ReadOnly != nil {
- in, out := &in.ReadOnly, &out.ReadOnly
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.Kind != nil {
- in, out := &in.Kind, &out.Kind
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureDataDiskKind)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AzureDiskVolumeSource.
-func (in *AzureDiskVolumeSource) DeepCopy() *AzureDiskVolumeSource {
- if in == nil {
- return nil
- }
- out := new(AzureDiskVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AzureFilePersistentVolumeSource) DeepCopyInto(out *AzureFilePersistentVolumeSource) {
- *out = *in
- if in.SecretNamespace != nil {
- in, out := &in.SecretNamespace, &out.SecretNamespace
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AzureFilePersistentVolumeSource.
-func (in *AzureFilePersistentVolumeSource) DeepCopy() *AzureFilePersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(AzureFilePersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AzureFileVolumeSource) DeepCopyInto(out *AzureFileVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AzureFileVolumeSource.
-func (in *AzureFileVolumeSource) DeepCopy() *AzureFileVolumeSource {
- if in == nil {
- return nil
- }
- out := new(AzureFileVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Binding) DeepCopyInto(out *Binding) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- out.Target = in.Target
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Binding.
-func (in *Binding) DeepCopy() *Binding {
- if in == nil {
- return nil
- }
- out := new(Binding)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Binding) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CSIPersistentVolumeSource) DeepCopyInto(out *CSIPersistentVolumeSource) {
- *out = *in
- if in.VolumeAttributes != nil {
- in, out := &in.VolumeAttributes, &out.VolumeAttributes
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.ControllerPublishSecretRef != nil {
- in, out := &in.ControllerPublishSecretRef, &out.ControllerPublishSecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- if in.NodeStageSecretRef != nil {
- in, out := &in.NodeStageSecretRef, &out.NodeStageSecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- if in.NodePublishSecretRef != nil {
- in, out := &in.NodePublishSecretRef, &out.NodePublishSecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIPersistentVolumeSource.
-func (in *CSIPersistentVolumeSource) DeepCopy() *CSIPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(CSIPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Capabilities) DeepCopyInto(out *Capabilities) {
- *out = *in
- if in.Add != nil {
- in, out := &in.Add, &out.Add
- *out = make([]Capability, len(*in))
- copy(*out, *in)
- }
- if in.Drop != nil {
- in, out := &in.Drop, &out.Drop
- *out = make([]Capability, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Capabilities.
-func (in *Capabilities) DeepCopy() *Capabilities {
- if in == nil {
- return nil
- }
- out := new(Capabilities)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CephFSPersistentVolumeSource) DeepCopyInto(out *CephFSPersistentVolumeSource) {
- *out = *in
- if in.Monitors != nil {
- in, out := &in.Monitors, &out.Monitors
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CephFSPersistentVolumeSource.
-func (in *CephFSPersistentVolumeSource) DeepCopy() *CephFSPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(CephFSPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CephFSVolumeSource) DeepCopyInto(out *CephFSVolumeSource) {
- *out = *in
- if in.Monitors != nil {
- in, out := &in.Monitors, &out.Monitors
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CephFSVolumeSource.
-func (in *CephFSVolumeSource) DeepCopy() *CephFSVolumeSource {
- if in == nil {
- return nil
- }
- out := new(CephFSVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CinderVolumeSource) DeepCopyInto(out *CinderVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CinderVolumeSource.
-func (in *CinderVolumeSource) DeepCopy() *CinderVolumeSource {
- if in == nil {
- return nil
- }
- out := new(CinderVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ClientIPConfig) DeepCopyInto(out *ClientIPConfig) {
- *out = *in
- if in.TimeoutSeconds != nil {
- in, out := &in.TimeoutSeconds, &out.TimeoutSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ClientIPConfig.
-func (in *ClientIPConfig) DeepCopy() *ClientIPConfig {
- if in == nil {
- return nil
- }
- out := new(ClientIPConfig)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ComponentCondition) DeepCopyInto(out *ComponentCondition) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ComponentCondition.
-func (in *ComponentCondition) DeepCopy() *ComponentCondition {
- if in == nil {
- return nil
- }
- out := new(ComponentCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ComponentStatus) DeepCopyInto(out *ComponentStatus) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]ComponentCondition, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ComponentStatus.
-func (in *ComponentStatus) DeepCopy() *ComponentStatus {
- if in == nil {
- return nil
- }
- out := new(ComponentStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ComponentStatus) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ComponentStatusList) DeepCopyInto(out *ComponentStatusList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ComponentStatus, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ComponentStatusList.
-func (in *ComponentStatusList) DeepCopy() *ComponentStatusList {
- if in == nil {
- return nil
- }
- out := new(ComponentStatusList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ComponentStatusList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMap) DeepCopyInto(out *ConfigMap) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Data != nil {
- in, out := &in.Data, &out.Data
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.BinaryData != nil {
- in, out := &in.BinaryData, &out.BinaryData
- *out = make(map[string][]byte, len(*in))
- for key, val := range *in {
- if val == nil {
- (*out)[key] = nil
- } else {
- (*out)[key] = make([]byte, len(val))
- copy((*out)[key], val)
- }
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMap.
-func (in *ConfigMap) DeepCopy() *ConfigMap {
- if in == nil {
- return nil
- }
- out := new(ConfigMap)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ConfigMap) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMapEnvSource) DeepCopyInto(out *ConfigMapEnvSource) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMapEnvSource.
-func (in *ConfigMapEnvSource) DeepCopy() *ConfigMapEnvSource {
- if in == nil {
- return nil
- }
- out := new(ConfigMapEnvSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMapKeySelector) DeepCopyInto(out *ConfigMapKeySelector) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMapKeySelector.
-func (in *ConfigMapKeySelector) DeepCopy() *ConfigMapKeySelector {
- if in == nil {
- return nil
- }
- out := new(ConfigMapKeySelector)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMapList) DeepCopyInto(out *ConfigMapList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ConfigMap, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMapList.
-func (in *ConfigMapList) DeepCopy() *ConfigMapList {
- if in == nil {
- return nil
- }
- out := new(ConfigMapList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ConfigMapList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMapProjection) DeepCopyInto(out *ConfigMapProjection) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]KeyToPath, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMapProjection.
-func (in *ConfigMapProjection) DeepCopy() *ConfigMapProjection {
- if in == nil {
- return nil
- }
- out := new(ConfigMapProjection)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ConfigMapVolumeSource) DeepCopyInto(out *ConfigMapVolumeSource) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]KeyToPath, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.DefaultMode != nil {
- in, out := &in.DefaultMode, &out.DefaultMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ConfigMapVolumeSource.
-func (in *ConfigMapVolumeSource) DeepCopy() *ConfigMapVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ConfigMapVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Container) DeepCopyInto(out *Container) {
- *out = *in
- if in.Command != nil {
- in, out := &in.Command, &out.Command
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.Args != nil {
- in, out := &in.Args, &out.Args
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]ContainerPort, len(*in))
- copy(*out, *in)
- }
- if in.EnvFrom != nil {
- in, out := &in.EnvFrom, &out.EnvFrom
- *out = make([]EnvFromSource, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Env != nil {
- in, out := &in.Env, &out.Env
- *out = make([]EnvVar, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- in.Resources.DeepCopyInto(&out.Resources)
- if in.VolumeMounts != nil {
- in, out := &in.VolumeMounts, &out.VolumeMounts
- *out = make([]VolumeMount, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.VolumeDevices != nil {
- in, out := &in.VolumeDevices, &out.VolumeDevices
- *out = make([]VolumeDevice, len(*in))
- copy(*out, *in)
- }
- if in.LivenessProbe != nil {
- in, out := &in.LivenessProbe, &out.LivenessProbe
- if *in == nil {
- *out = nil
- } else {
- *out = new(Probe)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.ReadinessProbe != nil {
- in, out := &in.ReadinessProbe, &out.ReadinessProbe
- if *in == nil {
- *out = nil
- } else {
- *out = new(Probe)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Lifecycle != nil {
- in, out := &in.Lifecycle, &out.Lifecycle
- if *in == nil {
- *out = nil
- } else {
- *out = new(Lifecycle)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecurityContext)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Container.
-func (in *Container) DeepCopy() *Container {
- if in == nil {
- return nil
- }
- out := new(Container)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerImage) DeepCopyInto(out *ContainerImage) {
- *out = *in
- if in.Names != nil {
- in, out := &in.Names, &out.Names
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerImage.
-func (in *ContainerImage) DeepCopy() *ContainerImage {
- if in == nil {
- return nil
- }
- out := new(ContainerImage)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerPort) DeepCopyInto(out *ContainerPort) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerPort.
-func (in *ContainerPort) DeepCopy() *ContainerPort {
- if in == nil {
- return nil
- }
- out := new(ContainerPort)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerState) DeepCopyInto(out *ContainerState) {
- *out = *in
- if in.Waiting != nil {
- in, out := &in.Waiting, &out.Waiting
- if *in == nil {
- *out = nil
- } else {
- *out = new(ContainerStateWaiting)
- **out = **in
- }
- }
- if in.Running != nil {
- in, out := &in.Running, &out.Running
- if *in == nil {
- *out = nil
- } else {
- *out = new(ContainerStateRunning)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Terminated != nil {
- in, out := &in.Terminated, &out.Terminated
- if *in == nil {
- *out = nil
- } else {
- *out = new(ContainerStateTerminated)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerState.
-func (in *ContainerState) DeepCopy() *ContainerState {
- if in == nil {
- return nil
- }
- out := new(ContainerState)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerStateRunning) DeepCopyInto(out *ContainerStateRunning) {
- *out = *in
- in.StartedAt.DeepCopyInto(&out.StartedAt)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerStateRunning.
-func (in *ContainerStateRunning) DeepCopy() *ContainerStateRunning {
- if in == nil {
- return nil
- }
- out := new(ContainerStateRunning)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerStateTerminated) DeepCopyInto(out *ContainerStateTerminated) {
- *out = *in
- in.StartedAt.DeepCopyInto(&out.StartedAt)
- in.FinishedAt.DeepCopyInto(&out.FinishedAt)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerStateTerminated.
-func (in *ContainerStateTerminated) DeepCopy() *ContainerStateTerminated {
- if in == nil {
- return nil
- }
- out := new(ContainerStateTerminated)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerStateWaiting) DeepCopyInto(out *ContainerStateWaiting) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerStateWaiting.
-func (in *ContainerStateWaiting) DeepCopy() *ContainerStateWaiting {
- if in == nil {
- return nil
- }
- out := new(ContainerStateWaiting)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ContainerStatus) DeepCopyInto(out *ContainerStatus) {
- *out = *in
- in.State.DeepCopyInto(&out.State)
- in.LastTerminationState.DeepCopyInto(&out.LastTerminationState)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ContainerStatus.
-func (in *ContainerStatus) DeepCopy() *ContainerStatus {
- if in == nil {
- return nil
- }
- out := new(ContainerStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonEndpoint) DeepCopyInto(out *DaemonEndpoint) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonEndpoint.
-func (in *DaemonEndpoint) DeepCopy() *DaemonEndpoint {
- if in == nil {
- return nil
- }
- out := new(DaemonEndpoint)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeleteOptions) DeepCopyInto(out *DeleteOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.GracePeriodSeconds != nil {
- in, out := &in.GracePeriodSeconds, &out.GracePeriodSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.Preconditions != nil {
- in, out := &in.Preconditions, &out.Preconditions
- if *in == nil {
- *out = nil
- } else {
- *out = new(Preconditions)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.OrphanDependents != nil {
- in, out := &in.OrphanDependents, &out.OrphanDependents
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.PropagationPolicy != nil {
- in, out := &in.PropagationPolicy, &out.PropagationPolicy
- if *in == nil {
- *out = nil
- } else {
- *out = new(DeletionPropagation)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeleteOptions.
-func (in *DeleteOptions) DeepCopy() *DeleteOptions {
- if in == nil {
- return nil
- }
- out := new(DeleteOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *DeleteOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DownwardAPIProjection) DeepCopyInto(out *DownwardAPIProjection) {
- *out = *in
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]DownwardAPIVolumeFile, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DownwardAPIProjection.
-func (in *DownwardAPIProjection) DeepCopy() *DownwardAPIProjection {
- if in == nil {
- return nil
- }
- out := new(DownwardAPIProjection)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DownwardAPIVolumeFile) DeepCopyInto(out *DownwardAPIVolumeFile) {
- *out = *in
- if in.FieldRef != nil {
- in, out := &in.FieldRef, &out.FieldRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectFieldSelector)
- **out = **in
- }
- }
- if in.ResourceFieldRef != nil {
- in, out := &in.ResourceFieldRef, &out.ResourceFieldRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ResourceFieldSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Mode != nil {
- in, out := &in.Mode, &out.Mode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DownwardAPIVolumeFile.
-func (in *DownwardAPIVolumeFile) DeepCopy() *DownwardAPIVolumeFile {
- if in == nil {
- return nil
- }
- out := new(DownwardAPIVolumeFile)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DownwardAPIVolumeSource) DeepCopyInto(out *DownwardAPIVolumeSource) {
- *out = *in
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]DownwardAPIVolumeFile, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.DefaultMode != nil {
- in, out := &in.DefaultMode, &out.DefaultMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DownwardAPIVolumeSource.
-func (in *DownwardAPIVolumeSource) DeepCopy() *DownwardAPIVolumeSource {
- if in == nil {
- return nil
- }
- out := new(DownwardAPIVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EmptyDirVolumeSource) DeepCopyInto(out *EmptyDirVolumeSource) {
- *out = *in
- if in.SizeLimit != nil {
- in, out := &in.SizeLimit, &out.SizeLimit
- if *in == nil {
- *out = nil
- } else {
- x := (*in).DeepCopy()
- *out = &x
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EmptyDirVolumeSource.
-func (in *EmptyDirVolumeSource) DeepCopy() *EmptyDirVolumeSource {
- if in == nil {
- return nil
- }
- out := new(EmptyDirVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EndpointAddress) DeepCopyInto(out *EndpointAddress) {
- *out = *in
- if in.NodeName != nil {
- in, out := &in.NodeName, &out.NodeName
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- if in.TargetRef != nil {
- in, out := &in.TargetRef, &out.TargetRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EndpointAddress.
-func (in *EndpointAddress) DeepCopy() *EndpointAddress {
- if in == nil {
- return nil
- }
- out := new(EndpointAddress)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EndpointPort) DeepCopyInto(out *EndpointPort) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EndpointPort.
-func (in *EndpointPort) DeepCopy() *EndpointPort {
- if in == nil {
- return nil
- }
- out := new(EndpointPort)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EndpointSubset) DeepCopyInto(out *EndpointSubset) {
- *out = *in
- if in.Addresses != nil {
- in, out := &in.Addresses, &out.Addresses
- *out = make([]EndpointAddress, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.NotReadyAddresses != nil {
- in, out := &in.NotReadyAddresses, &out.NotReadyAddresses
- *out = make([]EndpointAddress, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]EndpointPort, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EndpointSubset.
-func (in *EndpointSubset) DeepCopy() *EndpointSubset {
- if in == nil {
- return nil
- }
- out := new(EndpointSubset)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Endpoints) DeepCopyInto(out *Endpoints) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Subsets != nil {
- in, out := &in.Subsets, &out.Subsets
- *out = make([]EndpointSubset, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Endpoints.
-func (in *Endpoints) DeepCopy() *Endpoints {
- if in == nil {
- return nil
- }
- out := new(Endpoints)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Endpoints) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EndpointsList) DeepCopyInto(out *EndpointsList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Endpoints, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EndpointsList.
-func (in *EndpointsList) DeepCopy() *EndpointsList {
- if in == nil {
- return nil
- }
- out := new(EndpointsList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *EndpointsList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EnvFromSource) DeepCopyInto(out *EnvFromSource) {
- *out = *in
- if in.ConfigMapRef != nil {
- in, out := &in.ConfigMapRef, &out.ConfigMapRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ConfigMapEnvSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretEnvSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnvFromSource.
-func (in *EnvFromSource) DeepCopy() *EnvFromSource {
- if in == nil {
- return nil
- }
- out := new(EnvFromSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EnvVar) DeepCopyInto(out *EnvVar) {
- *out = *in
- if in.ValueFrom != nil {
- in, out := &in.ValueFrom, &out.ValueFrom
- if *in == nil {
- *out = nil
- } else {
- *out = new(EnvVarSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnvVar.
-func (in *EnvVar) DeepCopy() *EnvVar {
- if in == nil {
- return nil
- }
- out := new(EnvVar)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EnvVarSource) DeepCopyInto(out *EnvVarSource) {
- *out = *in
- if in.FieldRef != nil {
- in, out := &in.FieldRef, &out.FieldRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectFieldSelector)
- **out = **in
- }
- }
- if in.ResourceFieldRef != nil {
- in, out := &in.ResourceFieldRef, &out.ResourceFieldRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ResourceFieldSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.ConfigMapKeyRef != nil {
- in, out := &in.ConfigMapKeyRef, &out.ConfigMapKeyRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ConfigMapKeySelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.SecretKeyRef != nil {
- in, out := &in.SecretKeyRef, &out.SecretKeyRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretKeySelector)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EnvVarSource.
-func (in *EnvVarSource) DeepCopy() *EnvVarSource {
- if in == nil {
- return nil
- }
- out := new(EnvVarSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Event) DeepCopyInto(out *Event) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- out.InvolvedObject = in.InvolvedObject
- out.Source = in.Source
- in.FirstTimestamp.DeepCopyInto(&out.FirstTimestamp)
- in.LastTimestamp.DeepCopyInto(&out.LastTimestamp)
- in.EventTime.DeepCopyInto(&out.EventTime)
- if in.Series != nil {
- in, out := &in.Series, &out.Series
- if *in == nil {
- *out = nil
- } else {
- *out = new(EventSeries)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Related != nil {
- in, out := &in.Related, &out.Related
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Event.
-func (in *Event) DeepCopy() *Event {
- if in == nil {
- return nil
- }
- out := new(Event)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Event) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EventList) DeepCopyInto(out *EventList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Event, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventList.
-func (in *EventList) DeepCopy() *EventList {
- if in == nil {
- return nil
- }
- out := new(EventList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *EventList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EventSeries) DeepCopyInto(out *EventSeries) {
- *out = *in
- in.LastObservedTime.DeepCopyInto(&out.LastObservedTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventSeries.
-func (in *EventSeries) DeepCopy() *EventSeries {
- if in == nil {
- return nil
- }
- out := new(EventSeries)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *EventSource) DeepCopyInto(out *EventSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventSource.
-func (in *EventSource) DeepCopy() *EventSource {
- if in == nil {
- return nil
- }
- out := new(EventSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExecAction) DeepCopyInto(out *ExecAction) {
- *out = *in
- if in.Command != nil {
- in, out := &in.Command, &out.Command
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExecAction.
-func (in *ExecAction) DeepCopy() *ExecAction {
- if in == nil {
- return nil
- }
- out := new(ExecAction)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *FCVolumeSource) DeepCopyInto(out *FCVolumeSource) {
- *out = *in
- if in.TargetWWNs != nil {
- in, out := &in.TargetWWNs, &out.TargetWWNs
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.Lun != nil {
- in, out := &in.Lun, &out.Lun
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.WWIDs != nil {
- in, out := &in.WWIDs, &out.WWIDs
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FCVolumeSource.
-func (in *FCVolumeSource) DeepCopy() *FCVolumeSource {
- if in == nil {
- return nil
- }
- out := new(FCVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *FlexPersistentVolumeSource) DeepCopyInto(out *FlexPersistentVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- if in.Options != nil {
- in, out := &in.Options, &out.Options
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FlexPersistentVolumeSource.
-func (in *FlexPersistentVolumeSource) DeepCopy() *FlexPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(FlexPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *FlexVolumeSource) DeepCopyInto(out *FlexVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- if in.Options != nil {
- in, out := &in.Options, &out.Options
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FlexVolumeSource.
-func (in *FlexVolumeSource) DeepCopy() *FlexVolumeSource {
- if in == nil {
- return nil
- }
- out := new(FlexVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *FlockerVolumeSource) DeepCopyInto(out *FlockerVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FlockerVolumeSource.
-func (in *FlockerVolumeSource) DeepCopy() *FlockerVolumeSource {
- if in == nil {
- return nil
- }
- out := new(FlockerVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *GCEPersistentDiskVolumeSource) DeepCopyInto(out *GCEPersistentDiskVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GCEPersistentDiskVolumeSource.
-func (in *GCEPersistentDiskVolumeSource) DeepCopy() *GCEPersistentDiskVolumeSource {
- if in == nil {
- return nil
- }
- out := new(GCEPersistentDiskVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *GitRepoVolumeSource) DeepCopyInto(out *GitRepoVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GitRepoVolumeSource.
-func (in *GitRepoVolumeSource) DeepCopy() *GitRepoVolumeSource {
- if in == nil {
- return nil
- }
- out := new(GitRepoVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *GlusterfsVolumeSource) DeepCopyInto(out *GlusterfsVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GlusterfsVolumeSource.
-func (in *GlusterfsVolumeSource) DeepCopy() *GlusterfsVolumeSource {
- if in == nil {
- return nil
- }
- out := new(GlusterfsVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HTTPGetAction) DeepCopyInto(out *HTTPGetAction) {
- *out = *in
- out.Port = in.Port
- if in.HTTPHeaders != nil {
- in, out := &in.HTTPHeaders, &out.HTTPHeaders
- *out = make([]HTTPHeader, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPGetAction.
-func (in *HTTPGetAction) DeepCopy() *HTTPGetAction {
- if in == nil {
- return nil
- }
- out := new(HTTPGetAction)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HTTPHeader) DeepCopyInto(out *HTTPHeader) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPHeader.
-func (in *HTTPHeader) DeepCopy() *HTTPHeader {
- if in == nil {
- return nil
- }
- out := new(HTTPHeader)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Handler) DeepCopyInto(out *Handler) {
- *out = *in
- if in.Exec != nil {
- in, out := &in.Exec, &out.Exec
- if *in == nil {
- *out = nil
- } else {
- *out = new(ExecAction)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.HTTPGet != nil {
- in, out := &in.HTTPGet, &out.HTTPGet
- if *in == nil {
- *out = nil
- } else {
- *out = new(HTTPGetAction)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.TCPSocket != nil {
- in, out := &in.TCPSocket, &out.TCPSocket
- if *in == nil {
- *out = nil
- } else {
- *out = new(TCPSocketAction)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Handler.
-func (in *Handler) DeepCopy() *Handler {
- if in == nil {
- return nil
- }
- out := new(Handler)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HostAlias) DeepCopyInto(out *HostAlias) {
- *out = *in
- if in.Hostnames != nil {
- in, out := &in.Hostnames, &out.Hostnames
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HostAlias.
-func (in *HostAlias) DeepCopy() *HostAlias {
- if in == nil {
- return nil
- }
- out := new(HostAlias)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HostPathVolumeSource) DeepCopyInto(out *HostPathVolumeSource) {
- *out = *in
- if in.Type != nil {
- in, out := &in.Type, &out.Type
- if *in == nil {
- *out = nil
- } else {
- *out = new(HostPathType)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HostPathVolumeSource.
-func (in *HostPathVolumeSource) DeepCopy() *HostPathVolumeSource {
- if in == nil {
- return nil
- }
- out := new(HostPathVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ISCSIPersistentVolumeSource) DeepCopyInto(out *ISCSIPersistentVolumeSource) {
- *out = *in
- if in.Portals != nil {
- in, out := &in.Portals, &out.Portals
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- if in.InitiatorName != nil {
- in, out := &in.InitiatorName, &out.InitiatorName
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ISCSIPersistentVolumeSource.
-func (in *ISCSIPersistentVolumeSource) DeepCopy() *ISCSIPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ISCSIPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ISCSIVolumeSource) DeepCopyInto(out *ISCSIVolumeSource) {
- *out = *in
- if in.Portals != nil {
- in, out := &in.Portals, &out.Portals
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- if in.InitiatorName != nil {
- in, out := &in.InitiatorName, &out.InitiatorName
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ISCSIVolumeSource.
-func (in *ISCSIVolumeSource) DeepCopy() *ISCSIVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ISCSIVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *KeyToPath) DeepCopyInto(out *KeyToPath) {
- *out = *in
- if in.Mode != nil {
- in, out := &in.Mode, &out.Mode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new KeyToPath.
-func (in *KeyToPath) DeepCopy() *KeyToPath {
- if in == nil {
- return nil
- }
- out := new(KeyToPath)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Lifecycle) DeepCopyInto(out *Lifecycle) {
- *out = *in
- if in.PostStart != nil {
- in, out := &in.PostStart, &out.PostStart
- if *in == nil {
- *out = nil
- } else {
- *out = new(Handler)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PreStop != nil {
- in, out := &in.PreStop, &out.PreStop
- if *in == nil {
- *out = nil
- } else {
- *out = new(Handler)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Lifecycle.
-func (in *Lifecycle) DeepCopy() *Lifecycle {
- if in == nil {
- return nil
- }
- out := new(Lifecycle)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LimitRange) DeepCopyInto(out *LimitRange) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LimitRange.
-func (in *LimitRange) DeepCopy() *LimitRange {
- if in == nil {
- return nil
- }
- out := new(LimitRange)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *LimitRange) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LimitRangeItem) DeepCopyInto(out *LimitRangeItem) {
- *out = *in
- if in.Max != nil {
- in, out := &in.Max, &out.Max
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Min != nil {
- in, out := &in.Min, &out.Min
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Default != nil {
- in, out := &in.Default, &out.Default
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.DefaultRequest != nil {
- in, out := &in.DefaultRequest, &out.DefaultRequest
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.MaxLimitRequestRatio != nil {
- in, out := &in.MaxLimitRequestRatio, &out.MaxLimitRequestRatio
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LimitRangeItem.
-func (in *LimitRangeItem) DeepCopy() *LimitRangeItem {
- if in == nil {
- return nil
- }
- out := new(LimitRangeItem)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LimitRangeList) DeepCopyInto(out *LimitRangeList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]LimitRange, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LimitRangeList.
-func (in *LimitRangeList) DeepCopy() *LimitRangeList {
- if in == nil {
- return nil
- }
- out := new(LimitRangeList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *LimitRangeList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LimitRangeSpec) DeepCopyInto(out *LimitRangeSpec) {
- *out = *in
- if in.Limits != nil {
- in, out := &in.Limits, &out.Limits
- *out = make([]LimitRangeItem, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LimitRangeSpec.
-func (in *LimitRangeSpec) DeepCopy() *LimitRangeSpec {
- if in == nil {
- return nil
- }
- out := new(LimitRangeSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *List) DeepCopyInto(out *List) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]runtime.Object, len(*in))
- for i := range *in {
- if (*in)[i] == nil {
- (*out)[i] = nil
- } else {
- (*out)[i] = (*in)[i].DeepCopyObject()
- }
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new List.
-func (in *List) DeepCopy() *List {
- if in == nil {
- return nil
- }
- out := new(List)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *List) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ListOptions) DeepCopyInto(out *ListOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.LabelSelector == nil {
- out.LabelSelector = nil
- } else {
- out.LabelSelector = in.LabelSelector.DeepCopySelector()
- }
- if in.FieldSelector == nil {
- out.FieldSelector = nil
- } else {
- out.FieldSelector = in.FieldSelector.DeepCopySelector()
- }
- if in.TimeoutSeconds != nil {
- in, out := &in.TimeoutSeconds, &out.TimeoutSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ListOptions.
-func (in *ListOptions) DeepCopy() *ListOptions {
- if in == nil {
- return nil
- }
- out := new(ListOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ListOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LoadBalancerIngress) DeepCopyInto(out *LoadBalancerIngress) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LoadBalancerIngress.
-func (in *LoadBalancerIngress) DeepCopy() *LoadBalancerIngress {
- if in == nil {
- return nil
- }
- out := new(LoadBalancerIngress)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LoadBalancerStatus) DeepCopyInto(out *LoadBalancerStatus) {
- *out = *in
- if in.Ingress != nil {
- in, out := &in.Ingress, &out.Ingress
- *out = make([]LoadBalancerIngress, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LoadBalancerStatus.
-func (in *LoadBalancerStatus) DeepCopy() *LoadBalancerStatus {
- if in == nil {
- return nil
- }
- out := new(LoadBalancerStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LocalObjectReference) DeepCopyInto(out *LocalObjectReference) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LocalObjectReference.
-func (in *LocalObjectReference) DeepCopy() *LocalObjectReference {
- if in == nil {
- return nil
- }
- out := new(LocalObjectReference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LocalVolumeSource) DeepCopyInto(out *LocalVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LocalVolumeSource.
-func (in *LocalVolumeSource) DeepCopy() *LocalVolumeSource {
- if in == nil {
- return nil
- }
- out := new(LocalVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NFSVolumeSource) DeepCopyInto(out *NFSVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NFSVolumeSource.
-func (in *NFSVolumeSource) DeepCopy() *NFSVolumeSource {
- if in == nil {
- return nil
- }
- out := new(NFSVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Namespace) DeepCopyInto(out *Namespace) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- out.Status = in.Status
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Namespace.
-func (in *Namespace) DeepCopy() *Namespace {
- if in == nil {
- return nil
- }
- out := new(Namespace)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Namespace) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NamespaceList) DeepCopyInto(out *NamespaceList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Namespace, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NamespaceList.
-func (in *NamespaceList) DeepCopy() *NamespaceList {
- if in == nil {
- return nil
- }
- out := new(NamespaceList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NamespaceList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NamespaceSpec) DeepCopyInto(out *NamespaceSpec) {
- *out = *in
- if in.Finalizers != nil {
- in, out := &in.Finalizers, &out.Finalizers
- *out = make([]FinalizerName, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NamespaceSpec.
-func (in *NamespaceSpec) DeepCopy() *NamespaceSpec {
- if in == nil {
- return nil
- }
- out := new(NamespaceSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NamespaceStatus) DeepCopyInto(out *NamespaceStatus) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NamespaceStatus.
-func (in *NamespaceStatus) DeepCopy() *NamespaceStatus {
- if in == nil {
- return nil
- }
- out := new(NamespaceStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Node) DeepCopyInto(out *Node) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Node.
-func (in *Node) DeepCopy() *Node {
- if in == nil {
- return nil
- }
- out := new(Node)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Node) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeAddress) DeepCopyInto(out *NodeAddress) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeAddress.
-func (in *NodeAddress) DeepCopy() *NodeAddress {
- if in == nil {
- return nil
- }
- out := new(NodeAddress)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeAffinity) DeepCopyInto(out *NodeAffinity) {
- *out = *in
- if in.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.RequiredDuringSchedulingIgnoredDuringExecution, &out.RequiredDuringSchedulingIgnoredDuringExecution
- if *in == nil {
- *out = nil
- } else {
- *out = new(NodeSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PreferredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.PreferredDuringSchedulingIgnoredDuringExecution, &out.PreferredDuringSchedulingIgnoredDuringExecution
- *out = make([]PreferredSchedulingTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeAffinity.
-func (in *NodeAffinity) DeepCopy() *NodeAffinity {
- if in == nil {
- return nil
- }
- out := new(NodeAffinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeCondition) DeepCopyInto(out *NodeCondition) {
- *out = *in
- in.LastHeartbeatTime.DeepCopyInto(&out.LastHeartbeatTime)
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeCondition.
-func (in *NodeCondition) DeepCopy() *NodeCondition {
- if in == nil {
- return nil
- }
- out := new(NodeCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeConfigSource) DeepCopyInto(out *NodeConfigSource) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.ConfigMapRef != nil {
- in, out := &in.ConfigMapRef, &out.ConfigMapRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeConfigSource.
-func (in *NodeConfigSource) DeepCopy() *NodeConfigSource {
- if in == nil {
- return nil
- }
- out := new(NodeConfigSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NodeConfigSource) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeDaemonEndpoints) DeepCopyInto(out *NodeDaemonEndpoints) {
- *out = *in
- out.KubeletEndpoint = in.KubeletEndpoint
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeDaemonEndpoints.
-func (in *NodeDaemonEndpoints) DeepCopy() *NodeDaemonEndpoints {
- if in == nil {
- return nil
- }
- out := new(NodeDaemonEndpoints)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeList) DeepCopyInto(out *NodeList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Node, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeList.
-func (in *NodeList) DeepCopy() *NodeList {
- if in == nil {
- return nil
- }
- out := new(NodeList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NodeList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeProxyOptions) DeepCopyInto(out *NodeProxyOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeProxyOptions.
-func (in *NodeProxyOptions) DeepCopy() *NodeProxyOptions {
- if in == nil {
- return nil
- }
- out := new(NodeProxyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NodeProxyOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeResources) DeepCopyInto(out *NodeResources) {
- *out = *in
- if in.Capacity != nil {
- in, out := &in.Capacity, &out.Capacity
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeResources.
-func (in *NodeResources) DeepCopy() *NodeResources {
- if in == nil {
- return nil
- }
- out := new(NodeResources)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeSelector) DeepCopyInto(out *NodeSelector) {
- *out = *in
- if in.NodeSelectorTerms != nil {
- in, out := &in.NodeSelectorTerms, &out.NodeSelectorTerms
- *out = make([]NodeSelectorTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeSelector.
-func (in *NodeSelector) DeepCopy() *NodeSelector {
- if in == nil {
- return nil
- }
- out := new(NodeSelector)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeSelectorRequirement) DeepCopyInto(out *NodeSelectorRequirement) {
- *out = *in
- if in.Values != nil {
- in, out := &in.Values, &out.Values
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeSelectorRequirement.
-func (in *NodeSelectorRequirement) DeepCopy() *NodeSelectorRequirement {
- if in == nil {
- return nil
- }
- out := new(NodeSelectorRequirement)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeSelectorTerm) DeepCopyInto(out *NodeSelectorTerm) {
- *out = *in
- if in.MatchExpressions != nil {
- in, out := &in.MatchExpressions, &out.MatchExpressions
- *out = make([]NodeSelectorRequirement, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeSelectorTerm.
-func (in *NodeSelectorTerm) DeepCopy() *NodeSelectorTerm {
- if in == nil {
- return nil
- }
- out := new(NodeSelectorTerm)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeSpec) DeepCopyInto(out *NodeSpec) {
- *out = *in
- if in.Taints != nil {
- in, out := &in.Taints, &out.Taints
- *out = make([]Taint, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.ConfigSource != nil {
- in, out := &in.ConfigSource, &out.ConfigSource
- if *in == nil {
- *out = nil
- } else {
- *out = new(NodeConfigSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeSpec.
-func (in *NodeSpec) DeepCopy() *NodeSpec {
- if in == nil {
- return nil
- }
- out := new(NodeSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeStatus) DeepCopyInto(out *NodeStatus) {
- *out = *in
- if in.Capacity != nil {
- in, out := &in.Capacity, &out.Capacity
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Allocatable != nil {
- in, out := &in.Allocatable, &out.Allocatable
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]NodeCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Addresses != nil {
- in, out := &in.Addresses, &out.Addresses
- *out = make([]NodeAddress, len(*in))
- copy(*out, *in)
- }
- out.DaemonEndpoints = in.DaemonEndpoints
- out.NodeInfo = in.NodeInfo
- if in.Images != nil {
- in, out := &in.Images, &out.Images
- *out = make([]ContainerImage, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.VolumesInUse != nil {
- in, out := &in.VolumesInUse, &out.VolumesInUse
- *out = make([]UniqueVolumeName, len(*in))
- copy(*out, *in)
- }
- if in.VolumesAttached != nil {
- in, out := &in.VolumesAttached, &out.VolumesAttached
- *out = make([]AttachedVolume, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeStatus.
-func (in *NodeStatus) DeepCopy() *NodeStatus {
- if in == nil {
- return nil
- }
- out := new(NodeStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NodeSystemInfo) DeepCopyInto(out *NodeSystemInfo) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NodeSystemInfo.
-func (in *NodeSystemInfo) DeepCopy() *NodeSystemInfo {
- if in == nil {
- return nil
- }
- out := new(NodeSystemInfo)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ObjectFieldSelector) DeepCopyInto(out *ObjectFieldSelector) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ObjectFieldSelector.
-func (in *ObjectFieldSelector) DeepCopy() *ObjectFieldSelector {
- if in == nil {
- return nil
- }
- out := new(ObjectFieldSelector)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ObjectMeta) DeepCopyInto(out *ObjectMeta) {
- *out = *in
- in.CreationTimestamp.DeepCopyInto(&out.CreationTimestamp)
- if in.DeletionTimestamp != nil {
- in, out := &in.DeletionTimestamp, &out.DeletionTimestamp
- if *in == nil {
- *out = nil
- } else {
- *out = (*in).DeepCopy()
- }
- }
- if in.DeletionGracePeriodSeconds != nil {
- in, out := &in.DeletionGracePeriodSeconds, &out.DeletionGracePeriodSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.Labels != nil {
- in, out := &in.Labels, &out.Labels
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.Annotations != nil {
- in, out := &in.Annotations, &out.Annotations
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.OwnerReferences != nil {
- in, out := &in.OwnerReferences, &out.OwnerReferences
- *out = make([]v1.OwnerReference, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Initializers != nil {
- in, out := &in.Initializers, &out.Initializers
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.Initializers)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Finalizers != nil {
- in, out := &in.Finalizers, &out.Finalizers
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ObjectMeta.
-func (in *ObjectMeta) DeepCopy() *ObjectMeta {
- if in == nil {
- return nil
- }
- out := new(ObjectMeta)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ObjectReference) DeepCopyInto(out *ObjectReference) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ObjectReference.
-func (in *ObjectReference) DeepCopy() *ObjectReference {
- if in == nil {
- return nil
- }
- out := new(ObjectReference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ObjectReference) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolume) DeepCopyInto(out *PersistentVolume) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- out.Status = in.Status
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolume.
-func (in *PersistentVolume) DeepCopy() *PersistentVolume {
- if in == nil {
- return nil
- }
- out := new(PersistentVolume)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PersistentVolume) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaim) DeepCopyInto(out *PersistentVolumeClaim) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaim.
-func (in *PersistentVolumeClaim) DeepCopy() *PersistentVolumeClaim {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaim)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PersistentVolumeClaim) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaimCondition) DeepCopyInto(out *PersistentVolumeClaimCondition) {
- *out = *in
- in.LastProbeTime.DeepCopyInto(&out.LastProbeTime)
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimCondition.
-func (in *PersistentVolumeClaimCondition) DeepCopy() *PersistentVolumeClaimCondition {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaimCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaimList) DeepCopyInto(out *PersistentVolumeClaimList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]PersistentVolumeClaim, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimList.
-func (in *PersistentVolumeClaimList) DeepCopy() *PersistentVolumeClaimList {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaimList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PersistentVolumeClaimList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaimSpec) DeepCopyInto(out *PersistentVolumeClaimSpec) {
- *out = *in
- if in.AccessModes != nil {
- in, out := &in.AccessModes, &out.AccessModes
- *out = make([]PersistentVolumeAccessMode, len(*in))
- copy(*out, *in)
- }
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- in.Resources.DeepCopyInto(&out.Resources)
- if in.StorageClassName != nil {
- in, out := &in.StorageClassName, &out.StorageClassName
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- if in.VolumeMode != nil {
- in, out := &in.VolumeMode, &out.VolumeMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(PersistentVolumeMode)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimSpec.
-func (in *PersistentVolumeClaimSpec) DeepCopy() *PersistentVolumeClaimSpec {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaimSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaimStatus) DeepCopyInto(out *PersistentVolumeClaimStatus) {
- *out = *in
- if in.AccessModes != nil {
- in, out := &in.AccessModes, &out.AccessModes
- *out = make([]PersistentVolumeAccessMode, len(*in))
- copy(*out, *in)
- }
- if in.Capacity != nil {
- in, out := &in.Capacity, &out.Capacity
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]PersistentVolumeClaimCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimStatus.
-func (in *PersistentVolumeClaimStatus) DeepCopy() *PersistentVolumeClaimStatus {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaimStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeClaimVolumeSource) DeepCopyInto(out *PersistentVolumeClaimVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimVolumeSource.
-func (in *PersistentVolumeClaimVolumeSource) DeepCopy() *PersistentVolumeClaimVolumeSource {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeClaimVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeList) DeepCopyInto(out *PersistentVolumeList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]PersistentVolume, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeList.
-func (in *PersistentVolumeList) DeepCopy() *PersistentVolumeList {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PersistentVolumeList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeSource) DeepCopyInto(out *PersistentVolumeSource) {
- *out = *in
- if in.GCEPersistentDisk != nil {
- in, out := &in.GCEPersistentDisk, &out.GCEPersistentDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(GCEPersistentDiskVolumeSource)
- **out = **in
- }
- }
- if in.AWSElasticBlockStore != nil {
- in, out := &in.AWSElasticBlockStore, &out.AWSElasticBlockStore
- if *in == nil {
- *out = nil
- } else {
- *out = new(AWSElasticBlockStoreVolumeSource)
- **out = **in
- }
- }
- if in.HostPath != nil {
- in, out := &in.HostPath, &out.HostPath
- if *in == nil {
- *out = nil
- } else {
- *out = new(HostPathVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Glusterfs != nil {
- in, out := &in.Glusterfs, &out.Glusterfs
- if *in == nil {
- *out = nil
- } else {
- *out = new(GlusterfsVolumeSource)
- **out = **in
- }
- }
- if in.NFS != nil {
- in, out := &in.NFS, &out.NFS
- if *in == nil {
- *out = nil
- } else {
- *out = new(NFSVolumeSource)
- **out = **in
- }
- }
- if in.RBD != nil {
- in, out := &in.RBD, &out.RBD
- if *in == nil {
- *out = nil
- } else {
- *out = new(RBDPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Quobyte != nil {
- in, out := &in.Quobyte, &out.Quobyte
- if *in == nil {
- *out = nil
- } else {
- *out = new(QuobyteVolumeSource)
- **out = **in
- }
- }
- if in.ISCSI != nil {
- in, out := &in.ISCSI, &out.ISCSI
- if *in == nil {
- *out = nil
- } else {
- *out = new(ISCSIPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.FlexVolume != nil {
- in, out := &in.FlexVolume, &out.FlexVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(FlexPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Cinder != nil {
- in, out := &in.Cinder, &out.Cinder
- if *in == nil {
- *out = nil
- } else {
- *out = new(CinderVolumeSource)
- **out = **in
- }
- }
- if in.CephFS != nil {
- in, out := &in.CephFS, &out.CephFS
- if *in == nil {
- *out = nil
- } else {
- *out = new(CephFSPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.FC != nil {
- in, out := &in.FC, &out.FC
- if *in == nil {
- *out = nil
- } else {
- *out = new(FCVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Flocker != nil {
- in, out := &in.Flocker, &out.Flocker
- if *in == nil {
- *out = nil
- } else {
- *out = new(FlockerVolumeSource)
- **out = **in
- }
- }
- if in.AzureFile != nil {
- in, out := &in.AzureFile, &out.AzureFile
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureFilePersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.VsphereVolume != nil {
- in, out := &in.VsphereVolume, &out.VsphereVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(VsphereVirtualDiskVolumeSource)
- **out = **in
- }
- }
- if in.AzureDisk != nil {
- in, out := &in.AzureDisk, &out.AzureDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureDiskVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PhotonPersistentDisk != nil {
- in, out := &in.PhotonPersistentDisk, &out.PhotonPersistentDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(PhotonPersistentDiskVolumeSource)
- **out = **in
- }
- }
- if in.PortworxVolume != nil {
- in, out := &in.PortworxVolume, &out.PortworxVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(PortworxVolumeSource)
- **out = **in
- }
- }
- if in.ScaleIO != nil {
- in, out := &in.ScaleIO, &out.ScaleIO
- if *in == nil {
- *out = nil
- } else {
- *out = new(ScaleIOPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Local != nil {
- in, out := &in.Local, &out.Local
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalVolumeSource)
- **out = **in
- }
- }
- if in.StorageOS != nil {
- in, out := &in.StorageOS, &out.StorageOS
- if *in == nil {
- *out = nil
- } else {
- *out = new(StorageOSPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.CSI != nil {
- in, out := &in.CSI, &out.CSI
- if *in == nil {
- *out = nil
- } else {
- *out = new(CSIPersistentVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeSource.
-func (in *PersistentVolumeSource) DeepCopy() *PersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeSpec) DeepCopyInto(out *PersistentVolumeSpec) {
- *out = *in
- if in.Capacity != nil {
- in, out := &in.Capacity, &out.Capacity
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- in.PersistentVolumeSource.DeepCopyInto(&out.PersistentVolumeSource)
- if in.AccessModes != nil {
- in, out := &in.AccessModes, &out.AccessModes
- *out = make([]PersistentVolumeAccessMode, len(*in))
- copy(*out, *in)
- }
- if in.ClaimRef != nil {
- in, out := &in.ClaimRef, &out.ClaimRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectReference)
- **out = **in
- }
- }
- if in.MountOptions != nil {
- in, out := &in.MountOptions, &out.MountOptions
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.VolumeMode != nil {
- in, out := &in.VolumeMode, &out.VolumeMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(PersistentVolumeMode)
- **out = **in
- }
- }
- if in.NodeAffinity != nil {
- in, out := &in.NodeAffinity, &out.NodeAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(VolumeNodeAffinity)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeSpec.
-func (in *PersistentVolumeSpec) DeepCopy() *PersistentVolumeSpec {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PersistentVolumeStatus) DeepCopyInto(out *PersistentVolumeStatus) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeStatus.
-func (in *PersistentVolumeStatus) DeepCopy() *PersistentVolumeStatus {
- if in == nil {
- return nil
- }
- out := new(PersistentVolumeStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PhotonPersistentDiskVolumeSource) DeepCopyInto(out *PhotonPersistentDiskVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PhotonPersistentDiskVolumeSource.
-func (in *PhotonPersistentDiskVolumeSource) DeepCopy() *PhotonPersistentDiskVolumeSource {
- if in == nil {
- return nil
- }
- out := new(PhotonPersistentDiskVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Pod) DeepCopyInto(out *Pod) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Pod.
-func (in *Pod) DeepCopy() *Pod {
- if in == nil {
- return nil
- }
- out := new(Pod)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Pod) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodAffinity) DeepCopyInto(out *PodAffinity) {
- *out = *in
- if in.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.RequiredDuringSchedulingIgnoredDuringExecution, &out.RequiredDuringSchedulingIgnoredDuringExecution
- *out = make([]PodAffinityTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.PreferredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.PreferredDuringSchedulingIgnoredDuringExecution, &out.PreferredDuringSchedulingIgnoredDuringExecution
- *out = make([]WeightedPodAffinityTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodAffinity.
-func (in *PodAffinity) DeepCopy() *PodAffinity {
- if in == nil {
- return nil
- }
- out := new(PodAffinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodAffinityTerm) DeepCopyInto(out *PodAffinityTerm) {
- *out = *in
- if in.LabelSelector != nil {
- in, out := &in.LabelSelector, &out.LabelSelector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Namespaces != nil {
- in, out := &in.Namespaces, &out.Namespaces
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodAffinityTerm.
-func (in *PodAffinityTerm) DeepCopy() *PodAffinityTerm {
- if in == nil {
- return nil
- }
- out := new(PodAffinityTerm)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodAntiAffinity) DeepCopyInto(out *PodAntiAffinity) {
- *out = *in
- if in.RequiredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.RequiredDuringSchedulingIgnoredDuringExecution, &out.RequiredDuringSchedulingIgnoredDuringExecution
- *out = make([]PodAffinityTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.PreferredDuringSchedulingIgnoredDuringExecution != nil {
- in, out := &in.PreferredDuringSchedulingIgnoredDuringExecution, &out.PreferredDuringSchedulingIgnoredDuringExecution
- *out = make([]WeightedPodAffinityTerm, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodAntiAffinity.
-func (in *PodAntiAffinity) DeepCopy() *PodAntiAffinity {
- if in == nil {
- return nil
- }
- out := new(PodAntiAffinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodAttachOptions) DeepCopyInto(out *PodAttachOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodAttachOptions.
-func (in *PodAttachOptions) DeepCopy() *PodAttachOptions {
- if in == nil {
- return nil
- }
- out := new(PodAttachOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodAttachOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodCondition) DeepCopyInto(out *PodCondition) {
- *out = *in
- in.LastProbeTime.DeepCopyInto(&out.LastProbeTime)
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodCondition.
-func (in *PodCondition) DeepCopy() *PodCondition {
- if in == nil {
- return nil
- }
- out := new(PodCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodDNSConfig) DeepCopyInto(out *PodDNSConfig) {
- *out = *in
- if in.Nameservers != nil {
- in, out := &in.Nameservers, &out.Nameservers
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.Searches != nil {
- in, out := &in.Searches, &out.Searches
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.Options != nil {
- in, out := &in.Options, &out.Options
- *out = make([]PodDNSConfigOption, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodDNSConfig.
-func (in *PodDNSConfig) DeepCopy() *PodDNSConfig {
- if in == nil {
- return nil
- }
- out := new(PodDNSConfig)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodDNSConfigOption) DeepCopyInto(out *PodDNSConfigOption) {
- *out = *in
- if in.Value != nil {
- in, out := &in.Value, &out.Value
- if *in == nil {
- *out = nil
- } else {
- *out = new(string)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodDNSConfigOption.
-func (in *PodDNSConfigOption) DeepCopy() *PodDNSConfigOption {
- if in == nil {
- return nil
- }
- out := new(PodDNSConfigOption)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodExecOptions) DeepCopyInto(out *PodExecOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.Command != nil {
- in, out := &in.Command, &out.Command
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodExecOptions.
-func (in *PodExecOptions) DeepCopy() *PodExecOptions {
- if in == nil {
- return nil
- }
- out := new(PodExecOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodExecOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodList) DeepCopyInto(out *PodList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Pod, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodList.
-func (in *PodList) DeepCopy() *PodList {
- if in == nil {
- return nil
- }
- out := new(PodList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodLogOptions) DeepCopyInto(out *PodLogOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.SinceSeconds != nil {
- in, out := &in.SinceSeconds, &out.SinceSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.SinceTime != nil {
- in, out := &in.SinceTime, &out.SinceTime
- if *in == nil {
- *out = nil
- } else {
- *out = (*in).DeepCopy()
- }
- }
- if in.TailLines != nil {
- in, out := &in.TailLines, &out.TailLines
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.LimitBytes != nil {
- in, out := &in.LimitBytes, &out.LimitBytes
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodLogOptions.
-func (in *PodLogOptions) DeepCopy() *PodLogOptions {
- if in == nil {
- return nil
- }
- out := new(PodLogOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodLogOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodPortForwardOptions) DeepCopyInto(out *PodPortForwardOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]int32, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodPortForwardOptions.
-func (in *PodPortForwardOptions) DeepCopy() *PodPortForwardOptions {
- if in == nil {
- return nil
- }
- out := new(PodPortForwardOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodPortForwardOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodProxyOptions) DeepCopyInto(out *PodProxyOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodProxyOptions.
-func (in *PodProxyOptions) DeepCopy() *PodProxyOptions {
- if in == nil {
- return nil
- }
- out := new(PodProxyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodProxyOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSecurityContext) DeepCopyInto(out *PodSecurityContext) {
- *out = *in
- if in.ShareProcessNamespace != nil {
- in, out := &in.ShareProcessNamespace, &out.ShareProcessNamespace
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.SELinuxOptions != nil {
- in, out := &in.SELinuxOptions, &out.SELinuxOptions
- if *in == nil {
- *out = nil
- } else {
- *out = new(SELinuxOptions)
- **out = **in
- }
- }
- if in.RunAsUser != nil {
- in, out := &in.RunAsUser, &out.RunAsUser
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.RunAsNonRoot != nil {
- in, out := &in.RunAsNonRoot, &out.RunAsNonRoot
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.SupplementalGroups != nil {
- in, out := &in.SupplementalGroups, &out.SupplementalGroups
- *out = make([]int64, len(*in))
- copy(*out, *in)
- }
- if in.FSGroup != nil {
- in, out := &in.FSGroup, &out.FSGroup
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSecurityContext.
-func (in *PodSecurityContext) DeepCopy() *PodSecurityContext {
- if in == nil {
- return nil
- }
- out := new(PodSecurityContext)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSignature) DeepCopyInto(out *PodSignature) {
- *out = *in
- if in.PodController != nil {
- in, out := &in.PodController, &out.PodController
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.OwnerReference)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSignature.
-func (in *PodSignature) DeepCopy() *PodSignature {
- if in == nil {
- return nil
- }
- out := new(PodSignature)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSpec) DeepCopyInto(out *PodSpec) {
- *out = *in
- if in.Volumes != nil {
- in, out := &in.Volumes, &out.Volumes
- *out = make([]Volume, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.InitContainers != nil {
- in, out := &in.InitContainers, &out.InitContainers
- *out = make([]Container, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Containers != nil {
- in, out := &in.Containers, &out.Containers
- *out = make([]Container, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.TerminationGracePeriodSeconds != nil {
- in, out := &in.TerminationGracePeriodSeconds, &out.TerminationGracePeriodSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.ActiveDeadlineSeconds != nil {
- in, out := &in.ActiveDeadlineSeconds, &out.ActiveDeadlineSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.NodeSelector != nil {
- in, out := &in.NodeSelector, &out.NodeSelector
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.AutomountServiceAccountToken != nil {
- in, out := &in.AutomountServiceAccountToken, &out.AutomountServiceAccountToken
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.SecurityContext != nil {
- in, out := &in.SecurityContext, &out.SecurityContext
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodSecurityContext)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.ImagePullSecrets != nil {
- in, out := &in.ImagePullSecrets, &out.ImagePullSecrets
- *out = make([]LocalObjectReference, len(*in))
- copy(*out, *in)
- }
- if in.Affinity != nil {
- in, out := &in.Affinity, &out.Affinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(Affinity)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Tolerations != nil {
- in, out := &in.Tolerations, &out.Tolerations
- *out = make([]Toleration, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.HostAliases != nil {
- in, out := &in.HostAliases, &out.HostAliases
- *out = make([]HostAlias, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Priority != nil {
- in, out := &in.Priority, &out.Priority
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.DNSConfig != nil {
- in, out := &in.DNSConfig, &out.DNSConfig
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodDNSConfig)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSpec.
-func (in *PodSpec) DeepCopy() *PodSpec {
- if in == nil {
- return nil
- }
- out := new(PodSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodStatus) DeepCopyInto(out *PodStatus) {
- *out = *in
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]PodCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.StartTime != nil {
- in, out := &in.StartTime, &out.StartTime
- if *in == nil {
- *out = nil
- } else {
- *out = (*in).DeepCopy()
- }
- }
- if in.InitContainerStatuses != nil {
- in, out := &in.InitContainerStatuses, &out.InitContainerStatuses
- *out = make([]ContainerStatus, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.ContainerStatuses != nil {
- in, out := &in.ContainerStatuses, &out.ContainerStatuses
- *out = make([]ContainerStatus, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodStatus.
-func (in *PodStatus) DeepCopy() *PodStatus {
- if in == nil {
- return nil
- }
- out := new(PodStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodStatusResult) DeepCopyInto(out *PodStatusResult) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodStatusResult.
-func (in *PodStatusResult) DeepCopy() *PodStatusResult {
- if in == nil {
- return nil
- }
- out := new(PodStatusResult)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodStatusResult) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodTemplate) DeepCopyInto(out *PodTemplate) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Template.DeepCopyInto(&out.Template)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodTemplate.
-func (in *PodTemplate) DeepCopy() *PodTemplate {
- if in == nil {
- return nil
- }
- out := new(PodTemplate)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodTemplate) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodTemplateList) DeepCopyInto(out *PodTemplateList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]PodTemplate, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodTemplateList.
-func (in *PodTemplateList) DeepCopy() *PodTemplateList {
- if in == nil {
- return nil
- }
- out := new(PodTemplateList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodTemplateList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodTemplateSpec) DeepCopyInto(out *PodTemplateSpec) {
- *out = *in
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodTemplateSpec.
-func (in *PodTemplateSpec) DeepCopy() *PodTemplateSpec {
- if in == nil {
- return nil
- }
- out := new(PodTemplateSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PortworxVolumeSource) DeepCopyInto(out *PortworxVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PortworxVolumeSource.
-func (in *PortworxVolumeSource) DeepCopy() *PortworxVolumeSource {
- if in == nil {
- return nil
- }
- out := new(PortworxVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Preconditions) DeepCopyInto(out *Preconditions) {
- *out = *in
- if in.UID != nil {
- in, out := &in.UID, &out.UID
- if *in == nil {
- *out = nil
- } else {
- *out = new(types.UID)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Preconditions.
-func (in *Preconditions) DeepCopy() *Preconditions {
- if in == nil {
- return nil
- }
- out := new(Preconditions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PreferAvoidPodsEntry) DeepCopyInto(out *PreferAvoidPodsEntry) {
- *out = *in
- in.PodSignature.DeepCopyInto(&out.PodSignature)
- in.EvictionTime.DeepCopyInto(&out.EvictionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PreferAvoidPodsEntry.
-func (in *PreferAvoidPodsEntry) DeepCopy() *PreferAvoidPodsEntry {
- if in == nil {
- return nil
- }
- out := new(PreferAvoidPodsEntry)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PreferredSchedulingTerm) DeepCopyInto(out *PreferredSchedulingTerm) {
- *out = *in
- in.Preference.DeepCopyInto(&out.Preference)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PreferredSchedulingTerm.
-func (in *PreferredSchedulingTerm) DeepCopy() *PreferredSchedulingTerm {
- if in == nil {
- return nil
- }
- out := new(PreferredSchedulingTerm)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Probe) DeepCopyInto(out *Probe) {
- *out = *in
- in.Handler.DeepCopyInto(&out.Handler)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Probe.
-func (in *Probe) DeepCopy() *Probe {
- if in == nil {
- return nil
- }
- out := new(Probe)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ProjectedVolumeSource) DeepCopyInto(out *ProjectedVolumeSource) {
- *out = *in
- if in.Sources != nil {
- in, out := &in.Sources, &out.Sources
- *out = make([]VolumeProjection, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.DefaultMode != nil {
- in, out := &in.DefaultMode, &out.DefaultMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ProjectedVolumeSource.
-func (in *ProjectedVolumeSource) DeepCopy() *ProjectedVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ProjectedVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *QuobyteVolumeSource) DeepCopyInto(out *QuobyteVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new QuobyteVolumeSource.
-func (in *QuobyteVolumeSource) DeepCopy() *QuobyteVolumeSource {
- if in == nil {
- return nil
- }
- out := new(QuobyteVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RBDPersistentVolumeSource) DeepCopyInto(out *RBDPersistentVolumeSource) {
- *out = *in
- if in.CephMonitors != nil {
- in, out := &in.CephMonitors, &out.CephMonitors
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RBDPersistentVolumeSource.
-func (in *RBDPersistentVolumeSource) DeepCopy() *RBDPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(RBDPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RBDVolumeSource) DeepCopyInto(out *RBDVolumeSource) {
- *out = *in
- if in.CephMonitors != nil {
- in, out := &in.CephMonitors, &out.CephMonitors
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RBDVolumeSource.
-func (in *RBDVolumeSource) DeepCopy() *RBDVolumeSource {
- if in == nil {
- return nil
- }
- out := new(RBDVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RangeAllocation) DeepCopyInto(out *RangeAllocation) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Data != nil {
- in, out := &in.Data, &out.Data
- *out = make([]byte, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RangeAllocation.
-func (in *RangeAllocation) DeepCopy() *RangeAllocation {
- if in == nil {
- return nil
- }
- out := new(RangeAllocation)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *RangeAllocation) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationController) DeepCopyInto(out *ReplicationController) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationController.
-func (in *ReplicationController) DeepCopy() *ReplicationController {
- if in == nil {
- return nil
- }
- out := new(ReplicationController)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ReplicationController) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationControllerCondition) DeepCopyInto(out *ReplicationControllerCondition) {
- *out = *in
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerCondition.
-func (in *ReplicationControllerCondition) DeepCopy() *ReplicationControllerCondition {
- if in == nil {
- return nil
- }
- out := new(ReplicationControllerCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationControllerList) DeepCopyInto(out *ReplicationControllerList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ReplicationController, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerList.
-func (in *ReplicationControllerList) DeepCopy() *ReplicationControllerList {
- if in == nil {
- return nil
- }
- out := new(ReplicationControllerList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ReplicationControllerList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationControllerSpec) DeepCopyInto(out *ReplicationControllerSpec) {
- *out = *in
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.Template != nil {
- in, out := &in.Template, &out.Template
- if *in == nil {
- *out = nil
- } else {
- *out = new(PodTemplateSpec)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerSpec.
-func (in *ReplicationControllerSpec) DeepCopy() *ReplicationControllerSpec {
- if in == nil {
- return nil
- }
- out := new(ReplicationControllerSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationControllerStatus) DeepCopyInto(out *ReplicationControllerStatus) {
- *out = *in
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]ReplicationControllerCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerStatus.
-func (in *ReplicationControllerStatus) DeepCopy() *ReplicationControllerStatus {
- if in == nil {
- return nil
- }
- out := new(ReplicationControllerStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceFieldSelector) DeepCopyInto(out *ResourceFieldSelector) {
- *out = *in
- out.Divisor = in.Divisor.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceFieldSelector.
-func (in *ResourceFieldSelector) DeepCopy() *ResourceFieldSelector {
- if in == nil {
- return nil
- }
- out := new(ResourceFieldSelector)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in ResourceList) DeepCopyInto(out *ResourceList) {
- {
- in := &in
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- return
- }
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceList.
-func (in ResourceList) DeepCopy() ResourceList {
- if in == nil {
- return nil
- }
- out := new(ResourceList)
- in.DeepCopyInto(out)
- return *out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceQuota) DeepCopyInto(out *ResourceQuota) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceQuota.
-func (in *ResourceQuota) DeepCopy() *ResourceQuota {
- if in == nil {
- return nil
- }
- out := new(ResourceQuota)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ResourceQuota) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceQuotaList) DeepCopyInto(out *ResourceQuotaList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ResourceQuota, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceQuotaList.
-func (in *ResourceQuotaList) DeepCopy() *ResourceQuotaList {
- if in == nil {
- return nil
- }
- out := new(ResourceQuotaList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ResourceQuotaList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceQuotaSpec) DeepCopyInto(out *ResourceQuotaSpec) {
- *out = *in
- if in.Hard != nil {
- in, out := &in.Hard, &out.Hard
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Scopes != nil {
- in, out := &in.Scopes, &out.Scopes
- *out = make([]ResourceQuotaScope, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceQuotaSpec.
-func (in *ResourceQuotaSpec) DeepCopy() *ResourceQuotaSpec {
- if in == nil {
- return nil
- }
- out := new(ResourceQuotaSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceQuotaStatus) DeepCopyInto(out *ResourceQuotaStatus) {
- *out = *in
- if in.Hard != nil {
- in, out := &in.Hard, &out.Hard
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Used != nil {
- in, out := &in.Used, &out.Used
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceQuotaStatus.
-func (in *ResourceQuotaStatus) DeepCopy() *ResourceQuotaStatus {
- if in == nil {
- return nil
- }
- out := new(ResourceQuotaStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ResourceRequirements) DeepCopyInto(out *ResourceRequirements) {
- *out = *in
- if in.Limits != nil {
- in, out := &in.Limits, &out.Limits
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- if in.Requests != nil {
- in, out := &in.Requests, &out.Requests
- *out = make(ResourceList, len(*in))
- for key, val := range *in {
- (*out)[key] = val.DeepCopy()
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ResourceRequirements.
-func (in *ResourceRequirements) DeepCopy() *ResourceRequirements {
- if in == nil {
- return nil
- }
- out := new(ResourceRequirements)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SELinuxOptions) DeepCopyInto(out *SELinuxOptions) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SELinuxOptions.
-func (in *SELinuxOptions) DeepCopy() *SELinuxOptions {
- if in == nil {
- return nil
- }
- out := new(SELinuxOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ScaleIOPersistentVolumeSource) DeepCopyInto(out *ScaleIOPersistentVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ScaleIOPersistentVolumeSource.
-func (in *ScaleIOPersistentVolumeSource) DeepCopy() *ScaleIOPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ScaleIOPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ScaleIOVolumeSource) DeepCopyInto(out *ScaleIOVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ScaleIOVolumeSource.
-func (in *ScaleIOVolumeSource) DeepCopy() *ScaleIOVolumeSource {
- if in == nil {
- return nil
- }
- out := new(ScaleIOVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Secret) DeepCopyInto(out *Secret) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Data != nil {
- in, out := &in.Data, &out.Data
- *out = make(map[string][]byte, len(*in))
- for key, val := range *in {
- if val == nil {
- (*out)[key] = nil
- } else {
- (*out)[key] = make([]byte, len(val))
- copy((*out)[key], val)
- }
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Secret.
-func (in *Secret) DeepCopy() *Secret {
- if in == nil {
- return nil
- }
- out := new(Secret)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Secret) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretEnvSource) DeepCopyInto(out *SecretEnvSource) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretEnvSource.
-func (in *SecretEnvSource) DeepCopy() *SecretEnvSource {
- if in == nil {
- return nil
- }
- out := new(SecretEnvSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretKeySelector) DeepCopyInto(out *SecretKeySelector) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretKeySelector.
-func (in *SecretKeySelector) DeepCopy() *SecretKeySelector {
- if in == nil {
- return nil
- }
- out := new(SecretKeySelector)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretList) DeepCopyInto(out *SecretList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Secret, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretList.
-func (in *SecretList) DeepCopy() *SecretList {
- if in == nil {
- return nil
- }
- out := new(SecretList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *SecretList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretProjection) DeepCopyInto(out *SecretProjection) {
- *out = *in
- out.LocalObjectReference = in.LocalObjectReference
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]KeyToPath, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretProjection.
-func (in *SecretProjection) DeepCopy() *SecretProjection {
- if in == nil {
- return nil
- }
- out := new(SecretProjection)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretReference) DeepCopyInto(out *SecretReference) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretReference.
-func (in *SecretReference) DeepCopy() *SecretReference {
- if in == nil {
- return nil
- }
- out := new(SecretReference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecretVolumeSource) DeepCopyInto(out *SecretVolumeSource) {
- *out = *in
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]KeyToPath, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.DefaultMode != nil {
- in, out := &in.DefaultMode, &out.DefaultMode
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.Optional != nil {
- in, out := &in.Optional, &out.Optional
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecretVolumeSource.
-func (in *SecretVolumeSource) DeepCopy() *SecretVolumeSource {
- if in == nil {
- return nil
- }
- out := new(SecretVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SecurityContext) DeepCopyInto(out *SecurityContext) {
- *out = *in
- if in.Capabilities != nil {
- in, out := &in.Capabilities, &out.Capabilities
- if *in == nil {
- *out = nil
- } else {
- *out = new(Capabilities)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Privileged != nil {
- in, out := &in.Privileged, &out.Privileged
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.SELinuxOptions != nil {
- in, out := &in.SELinuxOptions, &out.SELinuxOptions
- if *in == nil {
- *out = nil
- } else {
- *out = new(SELinuxOptions)
- **out = **in
- }
- }
- if in.RunAsUser != nil {
- in, out := &in.RunAsUser, &out.RunAsUser
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- if in.RunAsNonRoot != nil {
- in, out := &in.RunAsNonRoot, &out.RunAsNonRoot
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.ReadOnlyRootFilesystem != nil {
- in, out := &in.ReadOnlyRootFilesystem, &out.ReadOnlyRootFilesystem
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.AllowPrivilegeEscalation != nil {
- in, out := &in.AllowPrivilegeEscalation, &out.AllowPrivilegeEscalation
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SecurityContext.
-func (in *SecurityContext) DeepCopy() *SecurityContext {
- if in == nil {
- return nil
- }
- out := new(SecurityContext)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SerializedReference) DeepCopyInto(out *SerializedReference) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.Reference = in.Reference
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SerializedReference.
-func (in *SerializedReference) DeepCopy() *SerializedReference {
- if in == nil {
- return nil
- }
- out := new(SerializedReference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *SerializedReference) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Service) DeepCopyInto(out *Service) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Service.
-func (in *Service) DeepCopy() *Service {
- if in == nil {
- return nil
- }
- out := new(Service)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Service) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceAccount) DeepCopyInto(out *ServiceAccount) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- if in.Secrets != nil {
- in, out := &in.Secrets, &out.Secrets
- *out = make([]ObjectReference, len(*in))
- copy(*out, *in)
- }
- if in.ImagePullSecrets != nil {
- in, out := &in.ImagePullSecrets, &out.ImagePullSecrets
- *out = make([]LocalObjectReference, len(*in))
- copy(*out, *in)
- }
- if in.AutomountServiceAccountToken != nil {
- in, out := &in.AutomountServiceAccountToken, &out.AutomountServiceAccountToken
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceAccount.
-func (in *ServiceAccount) DeepCopy() *ServiceAccount {
- if in == nil {
- return nil
- }
- out := new(ServiceAccount)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ServiceAccount) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceAccountList) DeepCopyInto(out *ServiceAccountList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ServiceAccount, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceAccountList.
-func (in *ServiceAccountList) DeepCopy() *ServiceAccountList {
- if in == nil {
- return nil
- }
- out := new(ServiceAccountList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ServiceAccountList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceList) DeepCopyInto(out *ServiceList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Service, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceList.
-func (in *ServiceList) DeepCopy() *ServiceList {
- if in == nil {
- return nil
- }
- out := new(ServiceList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ServiceList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServicePort) DeepCopyInto(out *ServicePort) {
- *out = *in
- out.TargetPort = in.TargetPort
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServicePort.
-func (in *ServicePort) DeepCopy() *ServicePort {
- if in == nil {
- return nil
- }
- out := new(ServicePort)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceProxyOptions) DeepCopyInto(out *ServiceProxyOptions) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceProxyOptions.
-func (in *ServiceProxyOptions) DeepCopy() *ServiceProxyOptions {
- if in == nil {
- return nil
- }
- out := new(ServiceProxyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ServiceProxyOptions) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceSpec) DeepCopyInto(out *ServiceSpec) {
- *out = *in
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]ServicePort, len(*in))
- copy(*out, *in)
- }
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- if in.ExternalIPs != nil {
- in, out := &in.ExternalIPs, &out.ExternalIPs
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- if in.SessionAffinityConfig != nil {
- in, out := &in.SessionAffinityConfig, &out.SessionAffinityConfig
- if *in == nil {
- *out = nil
- } else {
- *out = new(SessionAffinityConfig)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.LoadBalancerSourceRanges != nil {
- in, out := &in.LoadBalancerSourceRanges, &out.LoadBalancerSourceRanges
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceSpec.
-func (in *ServiceSpec) DeepCopy() *ServiceSpec {
- if in == nil {
- return nil
- }
- out := new(ServiceSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceStatus) DeepCopyInto(out *ServiceStatus) {
- *out = *in
- in.LoadBalancer.DeepCopyInto(&out.LoadBalancer)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceStatus.
-func (in *ServiceStatus) DeepCopy() *ServiceStatus {
- if in == nil {
- return nil
- }
- out := new(ServiceStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SessionAffinityConfig) DeepCopyInto(out *SessionAffinityConfig) {
- *out = *in
- if in.ClientIP != nil {
- in, out := &in.ClientIP, &out.ClientIP
- if *in == nil {
- *out = nil
- } else {
- *out = new(ClientIPConfig)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SessionAffinityConfig.
-func (in *SessionAffinityConfig) DeepCopy() *SessionAffinityConfig {
- if in == nil {
- return nil
- }
- out := new(SessionAffinityConfig)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *StorageOSPersistentVolumeSource) DeepCopyInto(out *StorageOSPersistentVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(ObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StorageOSPersistentVolumeSource.
-func (in *StorageOSPersistentVolumeSource) DeepCopy() *StorageOSPersistentVolumeSource {
- if in == nil {
- return nil
- }
- out := new(StorageOSPersistentVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *StorageOSVolumeSource) DeepCopyInto(out *StorageOSVolumeSource) {
- *out = *in
- if in.SecretRef != nil {
- in, out := &in.SecretRef, &out.SecretRef
- if *in == nil {
- *out = nil
- } else {
- *out = new(LocalObjectReference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StorageOSVolumeSource.
-func (in *StorageOSVolumeSource) DeepCopy() *StorageOSVolumeSource {
- if in == nil {
- return nil
- }
- out := new(StorageOSVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Sysctl) DeepCopyInto(out *Sysctl) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Sysctl.
-func (in *Sysctl) DeepCopy() *Sysctl {
- if in == nil {
- return nil
- }
- out := new(Sysctl)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *TCPSocketAction) DeepCopyInto(out *TCPSocketAction) {
- *out = *in
- out.Port = in.Port
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new TCPSocketAction.
-func (in *TCPSocketAction) DeepCopy() *TCPSocketAction {
- if in == nil {
- return nil
- }
- out := new(TCPSocketAction)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Taint) DeepCopyInto(out *Taint) {
- *out = *in
- if in.TimeAdded != nil {
- in, out := &in.TimeAdded, &out.TimeAdded
- if *in == nil {
- *out = nil
- } else {
- *out = (*in).DeepCopy()
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Taint.
-func (in *Taint) DeepCopy() *Taint {
- if in == nil {
- return nil
- }
- out := new(Taint)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Toleration) DeepCopyInto(out *Toleration) {
- *out = *in
- if in.TolerationSeconds != nil {
- in, out := &in.TolerationSeconds, &out.TolerationSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int64)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Toleration.
-func (in *Toleration) DeepCopy() *Toleration {
- if in == nil {
- return nil
- }
- out := new(Toleration)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Volume) DeepCopyInto(out *Volume) {
- *out = *in
- in.VolumeSource.DeepCopyInto(&out.VolumeSource)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Volume.
-func (in *Volume) DeepCopy() *Volume {
- if in == nil {
- return nil
- }
- out := new(Volume)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VolumeDevice) DeepCopyInto(out *VolumeDevice) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VolumeDevice.
-func (in *VolumeDevice) DeepCopy() *VolumeDevice {
- if in == nil {
- return nil
- }
- out := new(VolumeDevice)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VolumeMount) DeepCopyInto(out *VolumeMount) {
- *out = *in
- if in.MountPropagation != nil {
- in, out := &in.MountPropagation, &out.MountPropagation
- if *in == nil {
- *out = nil
- } else {
- *out = new(MountPropagationMode)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VolumeMount.
-func (in *VolumeMount) DeepCopy() *VolumeMount {
- if in == nil {
- return nil
- }
- out := new(VolumeMount)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VolumeNodeAffinity) DeepCopyInto(out *VolumeNodeAffinity) {
- *out = *in
- if in.Required != nil {
- in, out := &in.Required, &out.Required
- if *in == nil {
- *out = nil
- } else {
- *out = new(NodeSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VolumeNodeAffinity.
-func (in *VolumeNodeAffinity) DeepCopy() *VolumeNodeAffinity {
- if in == nil {
- return nil
- }
- out := new(VolumeNodeAffinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VolumeProjection) DeepCopyInto(out *VolumeProjection) {
- *out = *in
- if in.Secret != nil {
- in, out := &in.Secret, &out.Secret
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretProjection)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.DownwardAPI != nil {
- in, out := &in.DownwardAPI, &out.DownwardAPI
- if *in == nil {
- *out = nil
- } else {
- *out = new(DownwardAPIProjection)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.ConfigMap != nil {
- in, out := &in.ConfigMap, &out.ConfigMap
- if *in == nil {
- *out = nil
- } else {
- *out = new(ConfigMapProjection)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VolumeProjection.
-func (in *VolumeProjection) DeepCopy() *VolumeProjection {
- if in == nil {
- return nil
- }
- out := new(VolumeProjection)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VolumeSource) DeepCopyInto(out *VolumeSource) {
- *out = *in
- if in.HostPath != nil {
- in, out := &in.HostPath, &out.HostPath
- if *in == nil {
- *out = nil
- } else {
- *out = new(HostPathVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.EmptyDir != nil {
- in, out := &in.EmptyDir, &out.EmptyDir
- if *in == nil {
- *out = nil
- } else {
- *out = new(EmptyDirVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.GCEPersistentDisk != nil {
- in, out := &in.GCEPersistentDisk, &out.GCEPersistentDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(GCEPersistentDiskVolumeSource)
- **out = **in
- }
- }
- if in.AWSElasticBlockStore != nil {
- in, out := &in.AWSElasticBlockStore, &out.AWSElasticBlockStore
- if *in == nil {
- *out = nil
- } else {
- *out = new(AWSElasticBlockStoreVolumeSource)
- **out = **in
- }
- }
- if in.GitRepo != nil {
- in, out := &in.GitRepo, &out.GitRepo
- if *in == nil {
- *out = nil
- } else {
- *out = new(GitRepoVolumeSource)
- **out = **in
- }
- }
- if in.Secret != nil {
- in, out := &in.Secret, &out.Secret
- if *in == nil {
- *out = nil
- } else {
- *out = new(SecretVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.NFS != nil {
- in, out := &in.NFS, &out.NFS
- if *in == nil {
- *out = nil
- } else {
- *out = new(NFSVolumeSource)
- **out = **in
- }
- }
- if in.ISCSI != nil {
- in, out := &in.ISCSI, &out.ISCSI
- if *in == nil {
- *out = nil
- } else {
- *out = new(ISCSIVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Glusterfs != nil {
- in, out := &in.Glusterfs, &out.Glusterfs
- if *in == nil {
- *out = nil
- } else {
- *out = new(GlusterfsVolumeSource)
- **out = **in
- }
- }
- if in.PersistentVolumeClaim != nil {
- in, out := &in.PersistentVolumeClaim, &out.PersistentVolumeClaim
- if *in == nil {
- *out = nil
- } else {
- *out = new(PersistentVolumeClaimVolumeSource)
- **out = **in
- }
- }
- if in.RBD != nil {
- in, out := &in.RBD, &out.RBD
- if *in == nil {
- *out = nil
- } else {
- *out = new(RBDVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Quobyte != nil {
- in, out := &in.Quobyte, &out.Quobyte
- if *in == nil {
- *out = nil
- } else {
- *out = new(QuobyteVolumeSource)
- **out = **in
- }
- }
- if in.FlexVolume != nil {
- in, out := &in.FlexVolume, &out.FlexVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(FlexVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Cinder != nil {
- in, out := &in.Cinder, &out.Cinder
- if *in == nil {
- *out = nil
- } else {
- *out = new(CinderVolumeSource)
- **out = **in
- }
- }
- if in.CephFS != nil {
- in, out := &in.CephFS, &out.CephFS
- if *in == nil {
- *out = nil
- } else {
- *out = new(CephFSVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.Flocker != nil {
- in, out := &in.Flocker, &out.Flocker
- if *in == nil {
- *out = nil
- } else {
- *out = new(FlockerVolumeSource)
- **out = **in
- }
- }
- if in.DownwardAPI != nil {
- in, out := &in.DownwardAPI, &out.DownwardAPI
- if *in == nil {
- *out = nil
- } else {
- *out = new(DownwardAPIVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.FC != nil {
- in, out := &in.FC, &out.FC
- if *in == nil {
- *out = nil
- } else {
- *out = new(FCVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.AzureFile != nil {
- in, out := &in.AzureFile, &out.AzureFile
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureFileVolumeSource)
- **out = **in
- }
- }
- if in.ConfigMap != nil {
- in, out := &in.ConfigMap, &out.ConfigMap
- if *in == nil {
- *out = nil
- } else {
- *out = new(ConfigMapVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.VsphereVolume != nil {
- in, out := &in.VsphereVolume, &out.VsphereVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(VsphereVirtualDiskVolumeSource)
- **out = **in
- }
- }
- if in.AzureDisk != nil {
- in, out := &in.AzureDisk, &out.AzureDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(AzureDiskVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PhotonPersistentDisk != nil {
- in, out := &in.PhotonPersistentDisk, &out.PhotonPersistentDisk
- if *in == nil {
- *out = nil
- } else {
- *out = new(PhotonPersistentDiskVolumeSource)
- **out = **in
- }
- }
- if in.Projected != nil {
- in, out := &in.Projected, &out.Projected
- if *in == nil {
- *out = nil
- } else {
- *out = new(ProjectedVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.PortworxVolume != nil {
- in, out := &in.PortworxVolume, &out.PortworxVolume
- if *in == nil {
- *out = nil
- } else {
- *out = new(PortworxVolumeSource)
- **out = **in
- }
- }
- if in.ScaleIO != nil {
- in, out := &in.ScaleIO, &out.ScaleIO
- if *in == nil {
- *out = nil
- } else {
- *out = new(ScaleIOVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.StorageOS != nil {
- in, out := &in.StorageOS, &out.StorageOS
- if *in == nil {
- *out = nil
- } else {
- *out = new(StorageOSVolumeSource)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VolumeSource.
-func (in *VolumeSource) DeepCopy() *VolumeSource {
- if in == nil {
- return nil
- }
- out := new(VolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *VsphereVirtualDiskVolumeSource) DeepCopyInto(out *VsphereVirtualDiskVolumeSource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new VsphereVirtualDiskVolumeSource.
-func (in *VsphereVirtualDiskVolumeSource) DeepCopy() *VsphereVirtualDiskVolumeSource {
- if in == nil {
- return nil
- }
- out := new(VsphereVirtualDiskVolumeSource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *WeightedPodAffinityTerm) DeepCopyInto(out *WeightedPodAffinityTerm) {
- *out = *in
- in.PodAffinityTerm.DeepCopyInto(&out.PodAffinityTerm)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new WeightedPodAffinityTerm.
-func (in *WeightedPodAffinityTerm) DeepCopy() *WeightedPodAffinityTerm {
- if in == nil {
- return nil
- }
- out := new(WeightedPodAffinityTerm)
- in.DeepCopyInto(out)
- return out
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/extensions/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/extensions/doc.go
deleted file mode 100644
index d97cffdbc..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/extensions/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:deepcopy-gen=package
-
-package extensions // import "k8s.io/kubernetes/pkg/apis/extensions"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/extensions/helpers.go b/vendor/k8s.io/kubernetes/pkg/apis/extensions/helpers.go
deleted file mode 100644
index 27d3e23ad..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/extensions/helpers.go
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package extensions
-
-import (
- "strings"
-)
-
-// SysctlsFromPodSecurityPolicyAnnotation parses an annotation value of the key
-// SysctlsSecurityPolocyAnnotationKey into a slice of sysctls. An empty slice
-// is returned if annotation is the empty string.
-func SysctlsFromPodSecurityPolicyAnnotation(annotation string) ([]string, error) {
- if len(annotation) == 0 {
- return []string{}, nil
- }
-
- return strings.Split(annotation, ","), nil
-}
-
-// PodAnnotationsFromSysctls creates an annotation value for a slice of Sysctls.
-func PodAnnotationsFromSysctls(sysctls []string) string {
- return strings.Join(sysctls, ",")
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/extensions/register.go b/vendor/k8s.io/kubernetes/pkg/apis/extensions/register.go
deleted file mode 100644
index 48137fc69..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/extensions/register.go
+++ /dev/null
@@ -1,68 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package extensions
-
-import (
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
- "k8s.io/kubernetes/pkg/apis/autoscaling"
- "k8s.io/kubernetes/pkg/apis/networking"
-)
-
-// GroupName is the group name use in this package
-const GroupName = "extensions"
-
-// SchemeGroupVersion is group version used to register these objects
-var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: runtime.APIVersionInternal}
-
-// Kind takes an unqualified kind and returns a Group qualified GroupKind
-func Kind(kind string) schema.GroupKind {
- return SchemeGroupVersion.WithKind(kind).GroupKind()
-}
-
-// Resource takes an unqualified resource and returns a Group qualified GroupResource
-func Resource(resource string) schema.GroupResource {
- return SchemeGroupVersion.WithResource(resource).GroupResource()
-}
-
-var (
- SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes)
- AddToScheme = SchemeBuilder.AddToScheme
-)
-
-// Adds the list of known types to the given scheme.
-func addKnownTypes(scheme *runtime.Scheme) error {
- // TODO this gets cleaned up when the types are fixed
- scheme.AddKnownTypes(SchemeGroupVersion,
- &Deployment{},
- &DeploymentList{},
- &DeploymentRollback{},
- &ReplicationControllerDummy{},
- &DaemonSetList{},
- &DaemonSet{},
- &Ingress{},
- &IngressList{},
- &ReplicaSet{},
- &ReplicaSetList{},
- &PodSecurityPolicy{},
- &PodSecurityPolicyList{},
- &autoscaling.Scale{},
- &networking.NetworkPolicy{},
- &networking.NetworkPolicyList{},
- )
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/extensions/types.go b/vendor/k8s.io/kubernetes/pkg/apis/extensions/types.go
deleted file mode 100644
index e36972846..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/extensions/types.go
+++ /dev/null
@@ -1,1049 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-/*
-This file (together with pkg/apis/extensions/v1beta1/types.go) contain the experimental
-types in kubernetes. These API objects are experimental, meaning that the
-APIs may be broken at any time by the kubernetes team.
-
-DISCLAIMER: The implementation of the experimental API group itself is
-a temporary one meant as a stopgap solution until kubernetes has proper
-support for multiple API groups. The transition may require changes
-beyond registration differences. In other words, experimental API group
-support is experimental.
-*/
-
-package extensions
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/util/intstr"
- api "k8s.io/kubernetes/pkg/apis/core"
-)
-
-const (
- // SysctlsPodSecurityPolicyAnnotationKey represents the key of a whitelist of
- // allowed safe and unsafe sysctls in a pod spec. It's a comma-separated list of plain sysctl
- // names or sysctl patterns (which end in *). The string "*" matches all sysctls.
- SysctlsPodSecurityPolicyAnnotationKey string = "security.alpha.kubernetes.io/sysctls"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Dummy definition
-type ReplicationControllerDummy struct {
- metav1.TypeMeta
-}
-
-// Alpha-level support for Custom Metrics in HPA (as annotations).
-type CustomMetricTarget struct {
- // Custom Metric name.
- Name string
- // Custom Metric value (average).
- TargetValue resource.Quantity
-}
-
-type CustomMetricTargetList struct {
- Items []CustomMetricTarget
-}
-
-type CustomMetricCurrentStatus struct {
- // Custom Metric name.
- Name string
- // Custom Metric value (average).
- CurrentValue resource.Quantity
-}
-
-type CustomMetricCurrentStatusList struct {
- Items []CustomMetricCurrentStatus
-}
-
-// +genclient
-// +genclient:method=GetScale,verb=get,subresource=scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +genclient:method=UpdateScale,verb=update,subresource=scale,input=k8s.io/kubernetes/pkg/apis/autoscaling.Scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type Deployment struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Specification of the desired behavior of the Deployment.
- // +optional
- Spec DeploymentSpec
-
- // Most recently observed status of the Deployment.
- // +optional
- Status DeploymentStatus
-}
-
-type DeploymentSpec struct {
- // Number of desired pods. This is a pointer to distinguish between explicit
- // zero and not specified. Defaults to 1.
- // +optional
- Replicas int32
-
- // Label selector for pods. Existing ReplicaSets whose pods are
- // selected by this will be the ones affected by this deployment.
- // +optional
- Selector *metav1.LabelSelector
-
- // Template describes the pods that will be created.
- Template api.PodTemplateSpec
-
- // The deployment strategy to use to replace existing pods with new ones.
- // +optional
- Strategy DeploymentStrategy
-
- // Minimum number of seconds for which a newly created pod should be ready
- // without any of its container crashing, for it to be considered available.
- // Defaults to 0 (pod will be considered available as soon as it is ready)
- // +optional
- MinReadySeconds int32
-
- // The number of old ReplicaSets to retain to allow rollback.
- // This is a pointer to distinguish between explicit zero and not specified.
- // +optional
- RevisionHistoryLimit *int32
-
- // Indicates that the deployment is paused and will not be processed by the
- // deployment controller.
- // +optional
- Paused bool
-
- // DEPRECATED.
- // The config this deployment is rolling back to. Will be cleared after rollback is done.
- // +optional
- RollbackTo *RollbackConfig
-
- // The maximum time in seconds for a deployment to make progress before it
- // is considered to be failed. The deployment controller will continue to
- // process failed deployments and a condition with a ProgressDeadlineExceeded
- // reason will be surfaced in the deployment status. Note that progress will
- // not be estimated during the time a deployment is paused. This is not set
- // by default.
- // +optional
- ProgressDeadlineSeconds *int32
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// DEPRECATED.
-// DeploymentRollback stores the information required to rollback a deployment.
-type DeploymentRollback struct {
- metav1.TypeMeta
- // Required: This must match the Name of a deployment.
- Name string
- // The annotations to be updated to a deployment
- // +optional
- UpdatedAnnotations map[string]string
- // The config of this deployment rollback.
- RollbackTo RollbackConfig
-}
-
-// DEPRECATED.
-type RollbackConfig struct {
- // The revision to rollback to. If set to 0, rollback to the last revision.
- // +optional
- Revision int64
-}
-
-const (
- // DefaultDeploymentUniqueLabelKey is the default key of the selector that is added
- // to existing RCs (and label key that is added to its pods) to prevent the existing RCs
- // to select new pods (and old pods being select by new RC).
- DefaultDeploymentUniqueLabelKey string = "pod-template-hash"
-)
-
-type DeploymentStrategy struct {
- // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate.
- // +optional
- Type DeploymentStrategyType
-
- // Rolling update config params. Present only if DeploymentStrategyType =
- // RollingUpdate.
- //---
- // TODO: Update this to follow our convention for oneOf, whatever we decide it
- // to be.
- // +optional
- RollingUpdate *RollingUpdateDeployment
-}
-
-type DeploymentStrategyType string
-
-const (
- // Kill all existing pods before creating new ones.
- RecreateDeploymentStrategyType DeploymentStrategyType = "Recreate"
-
- // Replace the old RCs by new one using rolling update i.e gradually scale down the old RCs and scale up the new one.
- RollingUpdateDeploymentStrategyType DeploymentStrategyType = "RollingUpdate"
-)
-
-// Spec to control the desired behavior of rolling update.
-type RollingUpdateDeployment struct {
- // The maximum number of pods that can be unavailable during the update.
- // Value can be an absolute number (ex: 5) or a percentage of total pods at the start of update (ex: 10%).
- // Absolute number is calculated from percentage by rounding down.
- // This can not be 0 if MaxSurge is 0.
- // By default, a fixed value of 1 is used.
- // Example: when this is set to 30%, the old RC can be scaled down by 30%
- // immediately when the rolling update starts. Once new pods are ready, old RC
- // can be scaled down further, followed by scaling up the new RC, ensuring
- // that at least 70% of original number of pods are available at all times
- // during the update.
- // +optional
- MaxUnavailable intstr.IntOrString
-
- // The maximum number of pods that can be scheduled above the original number of
- // pods.
- // Value can be an absolute number (ex: 5) or a percentage of total pods at
- // the start of the update (ex: 10%). This can not be 0 if MaxUnavailable is 0.
- // Absolute number is calculated from percentage by rounding up.
- // By default, a value of 1 is used.
- // Example: when this is set to 30%, the new RC can be scaled up by 30%
- // immediately when the rolling update starts. Once old pods have been killed,
- // new RC can be scaled up further, ensuring that total number of pods running
- // at any time during the update is atmost 130% of original pods.
- // +optional
- MaxSurge intstr.IntOrString
-}
-
-type DeploymentStatus struct {
- // The generation observed by the deployment controller.
- // +optional
- ObservedGeneration int64
-
- // Total number of non-terminated pods targeted by this deployment (their labels match the selector).
- // +optional
- Replicas int32
-
- // Total number of non-terminated pods targeted by this deployment that have the desired template spec.
- // +optional
- UpdatedReplicas int32
-
- // Total number of ready pods targeted by this deployment.
- // +optional
- ReadyReplicas int32
-
- // Total number of available pods (ready for at least minReadySeconds) targeted by this deployment.
- // +optional
- AvailableReplicas int32
-
- // Total number of unavailable pods targeted by this deployment. This is the total number of
- // pods that are still required for the deployment to have 100% available capacity. They may
- // either be pods that are running but not yet available or pods that still have not been created.
- // +optional
- UnavailableReplicas int32
-
- // Represents the latest available observations of a deployment's current state.
- Conditions []DeploymentCondition
-
- // Count of hash collisions for the Deployment. The Deployment controller uses this
- // field as a collision avoidance mechanism when it needs to create the name for the
- // newest ReplicaSet.
- // +optional
- CollisionCount *int32
-}
-
-type DeploymentConditionType string
-
-// These are valid conditions of a deployment.
-const (
- // Available means the deployment is available, ie. at least the minimum available
- // replicas required are up and running for at least minReadySeconds.
- DeploymentAvailable DeploymentConditionType = "Available"
- // Progressing means the deployment is progressing. Progress for a deployment is
- // considered when a new replica set is created or adopted, and when new pods scale
- // up or old pods scale down. Progress is not estimated for paused deployments or
- // when progressDeadlineSeconds is not specified.
- DeploymentProgressing DeploymentConditionType = "Progressing"
- // ReplicaFailure is added in a deployment when one of its pods fails to be created
- // or deleted.
- DeploymentReplicaFailure DeploymentConditionType = "ReplicaFailure"
-)
-
-// DeploymentCondition describes the state of a deployment at a certain point.
-type DeploymentCondition struct {
- // Type of deployment condition.
- Type DeploymentConditionType
- // Status of the condition, one of True, False, Unknown.
- Status api.ConditionStatus
- // The last time this condition was updated.
- LastUpdateTime metav1.Time
- // Last time the condition transitioned from one status to another.
- LastTransitionTime metav1.Time
- // The reason for the condition's last transition.
- Reason string
- // A human readable message indicating details about the transition.
- Message string
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-type DeploymentList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- // Items is the list of deployments.
- Items []Deployment
-}
-
-type DaemonSetUpdateStrategy struct {
- // Type of daemon set update. Can be "RollingUpdate" or "OnDelete".
- // Default is OnDelete.
- // +optional
- Type DaemonSetUpdateStrategyType
-
- // Rolling update config params. Present only if type = "RollingUpdate".
- //---
- // TODO: Update this to follow our convention for oneOf, whatever we decide it
- // to be. Same as Deployment `strategy.rollingUpdate`.
- // See https://github.com/kubernetes/kubernetes/issues/35345
- // +optional
- RollingUpdate *RollingUpdateDaemonSet
-}
-
-type DaemonSetUpdateStrategyType string
-
-const (
- // Replace the old daemons by new ones using rolling update i.e replace them on each node one after the other.
- RollingUpdateDaemonSetStrategyType DaemonSetUpdateStrategyType = "RollingUpdate"
-
- // Replace the old daemons only when it's killed
- OnDeleteDaemonSetStrategyType DaemonSetUpdateStrategyType = "OnDelete"
-)
-
-// Spec to control the desired behavior of daemon set rolling update.
-type RollingUpdateDaemonSet struct {
- // The maximum number of DaemonSet pods that can be unavailable during the
- // update. Value can be an absolute number (ex: 5) or a percentage of total
- // number of DaemonSet pods at the start of the update (ex: 10%). Absolute
- // number is calculated from percentage by rounding up.
- // This cannot be 0.
- // Default value is 1.
- // Example: when this is set to 30%, at most 30% of the total number of nodes
- // that should be running the daemon pod (i.e. status.desiredNumberScheduled)
- // can have their pods stopped for an update at any given
- // time. The update starts by stopping at most 30% of those DaemonSet pods
- // and then brings up new DaemonSet pods in their place. Once the new pods
- // are available, it then proceeds onto other DaemonSet pods, thus ensuring
- // that at least 70% of original number of DaemonSet pods are available at
- // all times during the update.
- // +optional
- MaxUnavailable intstr.IntOrString
-}
-
-// DaemonSetSpec is the specification of a daemon set.
-type DaemonSetSpec struct {
- // A label query over pods that are managed by the daemon set.
- // Must match in order to be controlled.
- // If empty, defaulted to labels on Pod template.
- // More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors
- // +optional
- Selector *metav1.LabelSelector
-
- // An object that describes the pod that will be created.
- // The DaemonSet will create exactly one copy of this pod on every node
- // that matches the template's node selector (or on every node if no node
- // selector is specified).
- // More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template
- Template api.PodTemplateSpec
-
- // An update strategy to replace existing DaemonSet pods with new pods.
- // +optional
- UpdateStrategy DaemonSetUpdateStrategy
-
- // The minimum number of seconds for which a newly created DaemonSet pod should
- // be ready without any of its container crashing, for it to be considered
- // available. Defaults to 0 (pod will be considered available as soon as it
- // is ready).
- // +optional
- MinReadySeconds int32
-
- // DEPRECATED.
- // A sequence number representing a specific generation of the template.
- // Populated by the system. It can be set only during the creation.
- // +optional
- TemplateGeneration int64
-
- // The number of old history to retain to allow rollback.
- // This is a pointer to distinguish between explicit zero and not specified.
- // Defaults to 10.
- // +optional
- RevisionHistoryLimit *int32
-}
-
-// DaemonSetStatus represents the current status of a daemon set.
-type DaemonSetStatus struct {
- // The number of nodes that are running at least 1
- // daemon pod and are supposed to run the daemon pod.
- CurrentNumberScheduled int32
-
- // The number of nodes that are running the daemon pod, but are
- // not supposed to run the daemon pod.
- NumberMisscheduled int32
-
- // The total number of nodes that should be running the daemon
- // pod (including nodes correctly running the daemon pod).
- DesiredNumberScheduled int32
-
- // The number of nodes that should be running the daemon pod and have one
- // or more of the daemon pod running and ready.
- NumberReady int32
-
- // The most recent generation observed by the daemon set controller.
- // +optional
- ObservedGeneration int64
-
- // The total number of nodes that are running updated daemon pod
- // +optional
- UpdatedNumberScheduled int32
-
- // The number of nodes that should be running the
- // daemon pod and have one or more of the daemon pod running and
- // available (ready for at least spec.minReadySeconds)
- // +optional
- NumberAvailable int32
-
- // The number of nodes that should be running the
- // daemon pod and have none of the daemon pod running and available
- // (ready for at least spec.minReadySeconds)
- // +optional
- NumberUnavailable int32
-
- // Count of hash collisions for the DaemonSet. The DaemonSet controller
- // uses this field as a collision avoidance mechanism when it needs to
- // create the name for the newest ControllerRevision.
- // +optional
- CollisionCount *int32
-
- // Represents the latest available observations of a DaemonSet's current state.
- Conditions []DaemonSetCondition
-}
-
-type DaemonSetConditionType string
-
-// TODO: Add valid condition types of a DaemonSet.
-
-// DaemonSetCondition describes the state of a DaemonSet at a certain point.
-type DaemonSetCondition struct {
- // Type of DaemonSet condition.
- Type DaemonSetConditionType
- // Status of the condition, one of True, False, Unknown.
- Status api.ConditionStatus
- // Last time the condition transitioned from one status to another.
- LastTransitionTime metav1.Time
- // The reason for the condition's last transition.
- Reason string
- // A human readable message indicating details about the transition.
- Message string
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// DaemonSet represents the configuration of a daemon set.
-type DaemonSet struct {
- metav1.TypeMeta
- // Standard object's metadata.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- metav1.ObjectMeta
-
- // The desired behavior of this daemon set.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status
- // +optional
- Spec DaemonSetSpec
-
- // The current status of this daemon set. This data may be
- // out of date by some window of time.
- // Populated by the system.
- // Read-only.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status
- // +optional
- Status DaemonSetStatus
-}
-
-const (
- // DEPRECATED: DefaultDaemonSetUniqueLabelKey is used instead.
- // DaemonSetTemplateGenerationKey is the key of the labels that is added
- // to daemon set pods to distinguish between old and new pod templates
- // during DaemonSet template update.
- DaemonSetTemplateGenerationKey string = "pod-template-generation"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// DaemonSetList is a collection of daemon sets.
-type DaemonSetList struct {
- metav1.TypeMeta
- // Standard list metadata.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- metav1.ListMeta
-
- // A list of daemon sets.
- Items []DaemonSet
-}
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Ingress is a collection of rules that allow inbound connections to reach the
-// endpoints defined by a backend. An Ingress can be configured to give services
-// externally-reachable urls, load balance traffic, terminate SSL, offer name
-// based virtual hosting etc.
-type Ingress struct {
- metav1.TypeMeta
- // Standard object's metadata.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- metav1.ObjectMeta
-
- // Spec is the desired state of the Ingress.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status
- // +optional
- Spec IngressSpec
-
- // Status is the current state of the Ingress.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status
- // +optional
- Status IngressStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// IngressList is a collection of Ingress.
-type IngressList struct {
- metav1.TypeMeta
- // Standard object's metadata.
- // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata
- // +optional
- metav1.ListMeta
-
- // Items is the list of Ingress.
- Items []Ingress
-}
-
-// IngressSpec describes the Ingress the user wishes to exist.
-type IngressSpec struct {
- // A default backend capable of servicing requests that don't match any
- // rule. At least one of 'backend' or 'rules' must be specified. This field
- // is optional to allow the loadbalancer controller or defaulting logic to
- // specify a global default.
- // +optional
- Backend *IngressBackend
-
- // TLS configuration. Currently the Ingress only supports a single TLS
- // port, 443. If multiple members of this list specify different hosts, they
- // will be multiplexed on the same port according to the hostname specified
- // through the SNI TLS extension, if the ingress controller fulfilling the
- // ingress supports SNI.
- // +optional
- TLS []IngressTLS
-
- // A list of host rules used to configure the Ingress. If unspecified, or
- // no rule matches, all traffic is sent to the default backend.
- // +optional
- Rules []IngressRule
- // TODO: Add the ability to specify load-balancer IP through claims
-}
-
-// IngressTLS describes the transport layer security associated with an Ingress.
-type IngressTLS struct {
- // Hosts are a list of hosts included in the TLS certificate. The values in
- // this list must match the name/s used in the tlsSecret. Defaults to the
- // wildcard host setting for the loadbalancer controller fulfilling this
- // Ingress, if left unspecified.
- // +optional
- Hosts []string
- // SecretName is the name of the secret used to terminate SSL traffic on 443.
- // Field is left optional to allow SSL routing based on SNI hostname alone.
- // If the SNI host in a listener conflicts with the "Host" header field used
- // by an IngressRule, the SNI host is used for termination and value of the
- // Host header is used for routing.
- // +optional
- SecretName string
- // TODO: Consider specifying different modes of termination, protocols etc.
-}
-
-// IngressStatus describe the current state of the Ingress.
-type IngressStatus struct {
- // LoadBalancer contains the current status of the load-balancer.
- // +optional
- LoadBalancer api.LoadBalancerStatus
-}
-
-// IngressRule represents the rules mapping the paths under a specified host to
-// the related backend services. Incoming requests are first evaluated for a host
-// match, then routed to the backend associated with the matching IngressRuleValue.
-type IngressRule struct {
- // Host is the fully qualified domain name of a network host, as defined
- // by RFC 3986. Note the following deviations from the "host" part of the
- // URI as defined in the RFC:
- // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the
- // IP in the Spec of the parent Ingress.
- // 2. The `:` delimiter is not respected because ports are not allowed.
- // Currently the port of an Ingress is implicitly :80 for http and
- // :443 for https.
- // Both these may change in the future.
- // Incoming requests are matched against the host before the IngressRuleValue.
- // If the host is unspecified, the Ingress routes all traffic based on the
- // specified IngressRuleValue.
- // +optional
- Host string
- // IngressRuleValue represents a rule to route requests for this IngressRule.
- // If unspecified, the rule defaults to a http catch-all. Whether that sends
- // just traffic matching the host to the default backend or all traffic to the
- // default backend, is left to the controller fulfilling the Ingress. Http is
- // currently the only supported IngressRuleValue.
- // +optional
- IngressRuleValue
-}
-
-// IngressRuleValue represents a rule to apply against incoming requests. If the
-// rule is satisfied, the request is routed to the specified backend. Currently
-// mixing different types of rules in a single Ingress is disallowed, so exactly
-// one of the following must be set.
-type IngressRuleValue struct {
- //TODO:
- // 1. Consider renaming this resource and the associated rules so they
- // aren't tied to Ingress. They can be used to route intra-cluster traffic.
- // 2. Consider adding fields for ingress-type specific global options
- // usable by a loadbalancer, like http keep-alive.
-
- // +optional
- HTTP *HTTPIngressRuleValue
-}
-
-// HTTPIngressRuleValue is a list of http selectors pointing to backends.
-// In the example: http://<host>/<path>?<searchpart> -> backend where
-// where parts of the url correspond to RFC 3986, this resource will be used
-// to match against everything after the last '/' and before the first '?'
-// or '#'.
-type HTTPIngressRuleValue struct {
- // A collection of paths that map requests to backends.
- Paths []HTTPIngressPath
- // TODO: Consider adding fields for ingress-type specific global
- // options usable by a loadbalancer, like http keep-alive.
-}
-
-// HTTPIngressPath associates a path regex with a backend. Incoming urls matching
-// the path are forwarded to the backend.
-type HTTPIngressPath struct {
- // Path is an extended POSIX regex as defined by IEEE Std 1003.1,
- // (i.e this follows the egrep/unix syntax, not the perl syntax)
- // matched against the path of an incoming request. Currently it can
- // contain characters disallowed from the conventional "path"
- // part of a URL as defined by RFC 3986. Paths must begin with
- // a '/'. If unspecified, the path defaults to a catch all sending
- // traffic to the backend.
- // +optional
- Path string
-
- // Backend defines the referenced service endpoint to which the traffic
- // will be forwarded to.
- Backend IngressBackend
-}
-
-// IngressBackend describes all endpoints for a given service and port.
-type IngressBackend struct {
- // Specifies the name of the referenced service.
- ServiceName string
-
- // Specifies the port of the referenced service.
- ServicePort intstr.IntOrString
-}
-
-// +genclient
-// +genclient:method=GetScale,verb=get,subresource=scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +genclient:method=UpdateScale,verb=update,subresource=scale,input=k8s.io/kubernetes/pkg/apis/autoscaling.Scale,result=k8s.io/kubernetes/pkg/apis/autoscaling.Scale
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ReplicaSet ensures that a specified number of pod replicas are running at any given time.
-type ReplicaSet struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the desired behavior of this ReplicaSet.
- // +optional
- Spec ReplicaSetSpec
-
- // Status is the current status of this ReplicaSet. This data may be
- // out of date by some window of time.
- // +optional
- Status ReplicaSetStatus
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// ReplicaSetList is a collection of ReplicaSets.
-type ReplicaSetList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []ReplicaSet
-}
-
-// ReplicaSetSpec is the specification of a ReplicaSet.
-// As the internal representation of a ReplicaSet, it must have
-// a Template set.
-type ReplicaSetSpec struct {
- // Replicas is the number of desired replicas.
- Replicas int32
-
- // Minimum number of seconds for which a newly created pod should be ready
- // without any of its container crashing, for it to be considered available.
- // Defaults to 0 (pod will be considered available as soon as it is ready)
- // +optional
- MinReadySeconds int32
-
- // Selector is a label query over pods that should match the replica count.
- // Must match in order to be controlled.
- // If empty, defaulted to labels on pod template.
- // More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors
- // +optional
- Selector *metav1.LabelSelector
-
- // Template is the object that describes the pod that will be created if
- // insufficient replicas are detected.
- // +optional
- Template api.PodTemplateSpec
-}
-
-// ReplicaSetStatus represents the current status of a ReplicaSet.
-type ReplicaSetStatus struct {
- // Replicas is the number of actual replicas.
- Replicas int32
-
- // The number of pods that have labels matching the labels of the pod template of the replicaset.
- // +optional
- FullyLabeledReplicas int32
-
- // The number of ready replicas for this replica set.
- // +optional
- ReadyReplicas int32
-
- // The number of available replicas (ready for at least minReadySeconds) for this replica set.
- // +optional
- AvailableReplicas int32
-
- // ObservedGeneration is the most recent generation observed by the controller.
- // +optional
- ObservedGeneration int64
-
- // Represents the latest available observations of a replica set's current state.
- // +optional
- Conditions []ReplicaSetCondition
-}
-
-type ReplicaSetConditionType string
-
-// These are valid conditions of a replica set.
-const (
- // ReplicaSetReplicaFailure is added in a replica set when one of its pods fails to be created
- // due to insufficient quota, limit ranges, pod security policy, node selectors, etc. or deleted
- // due to kubelet being down or finalizers are failing.
- ReplicaSetReplicaFailure ReplicaSetConditionType = "ReplicaFailure"
-)
-
-// ReplicaSetCondition describes the state of a replica set at a certain point.
-type ReplicaSetCondition struct {
- // Type of replica set condition.
- Type ReplicaSetConditionType
- // Status of the condition, one of True, False, Unknown.
- Status api.ConditionStatus
- // The last time the condition transitioned from one status to another.
- // +optional
- LastTransitionTime metav1.Time
- // The reason for the condition's last transition.
- // +optional
- Reason string
- // A human readable message indicating details about the transition.
- // +optional
- Message string
-}
-
-// +genclient
-// +genclient:nonNamespaced
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodSecurityPolicy governs the ability to make requests that affect the SecurityContext
-// that will be applied to a pod and container.
-type PodSecurityPolicy struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Spec defines the policy enforced.
- // +optional
- Spec PodSecurityPolicySpec
-}
-
-// PodSecurityPolicySpec defines the policy enforced.
-type PodSecurityPolicySpec struct {
- // Privileged determines if a pod can request to be run as privileged.
- // +optional
- Privileged bool
- // DefaultAddCapabilities is the default set of capabilities that will be added to the container
- // unless the pod spec specifically drops the capability. You may not list a capability in both
- // DefaultAddCapabilities and RequiredDropCapabilities. Capabilities added here are implicitly
- // allowed, and need not be included in the AllowedCapabilities list.
- // +optional
- DefaultAddCapabilities []api.Capability
- // RequiredDropCapabilities are the capabilities that will be dropped from the container. These
- // are required to be dropped and cannot be added.
- // +optional
- RequiredDropCapabilities []api.Capability
- // AllowedCapabilities is a list of capabilities that can be requested to add to the container.
- // Capabilities in this field may be added at the pod author's discretion.
- // You must not list a capability in both AllowedCapabilities and RequiredDropCapabilities.
- // To allow all capabilities you may use '*'.
- // +optional
- AllowedCapabilities []api.Capability
- // Volumes is a white list of allowed volume plugins. Empty indicates that all plugins
- // may be used.
- // +optional
- Volumes []FSType
- // HostNetwork determines if the policy allows the use of HostNetwork in the pod spec.
- // +optional
- HostNetwork bool
- // HostPorts determines which host port ranges are allowed to be exposed.
- // +optional
- HostPorts []HostPortRange
- // HostPID determines if the policy allows the use of HostPID in the pod spec.
- // +optional
- HostPID bool
- // HostIPC determines if the policy allows the use of HostIPC in the pod spec.
- // +optional
- HostIPC bool
- // SELinux is the strategy that will dictate the allowable labels that may be set.
- SELinux SELinuxStrategyOptions
- // RunAsUser is the strategy that will dictate the allowable RunAsUser values that may be set.
- RunAsUser RunAsUserStrategyOptions
- // SupplementalGroups is the strategy that will dictate what supplemental groups are used by the SecurityContext.
- SupplementalGroups SupplementalGroupsStrategyOptions
- // FSGroup is the strategy that will dictate what fs group is used by the SecurityContext.
- FSGroup FSGroupStrategyOptions
- // ReadOnlyRootFilesystem when set to true will force containers to run with a read only root file
- // system. If the container specifically requests to run with a non-read only root file system
- // the PSP should deny the pod.
- // If set to false the container may run with a read only root file system if it wishes but it
- // will not be forced to.
- // +optional
- ReadOnlyRootFilesystem bool
- // DefaultAllowPrivilegeEscalation controls the default setting for whether a
- // process can gain more privileges than its parent process.
- // +optional
- DefaultAllowPrivilegeEscalation *bool
- // AllowPrivilegeEscalation determines if a pod can request to allow
- // privilege escalation. If unspecified, defaults to true.
- // +optional
- AllowPrivilegeEscalation bool
- // AllowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used.
- // +optional
- AllowedHostPaths []AllowedHostPath
- // AllowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all
- // Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes
- // is allowed in the "Volumes" field.
- // +optional
- AllowedFlexVolumes []AllowedFlexVolume
-}
-
-// AllowedHostPath defines the host volume conditions that will be enabled by a policy
-// for pods to use. It requires the path prefix to be defined.
-type AllowedHostPath struct {
- // PathPrefix is the path prefix that the host volume must match.
- // PathPrefix does not support `*`.
- // Trailing slashes are trimmed when validating the path prefix with a host path.
- //
- // Examples:
- // `/foo` would allow `/foo`, `/foo/` and `/foo/bar`
- // `/foo` would not allow `/food` or `/etc/foo`
- PathPrefix string
-}
-
-// HostPortRange defines a range of host ports that will be enabled by a policy
-// for pods to use. It requires both the start and end to be defined.
-type HostPortRange struct {
- // Min is the start of the range, inclusive.
- Min int32
- // Max is the end of the range, inclusive.
- Max int32
-}
-
-// AllowAllCapabilities can be used as a value for the PodSecurityPolicy.AllowAllCapabilities
-// field and means that any capabilities are allowed to be requested.
-var AllowAllCapabilities api.Capability = "*"
-
-// FSType gives strong typing to different file systems that are used by volumes.
-type FSType string
-
-var (
- AzureFile FSType = "azureFile"
- Flocker FSType = "flocker"
- FlexVolume FSType = "flexVolume"
- HostPath FSType = "hostPath"
- EmptyDir FSType = "emptyDir"
- GCEPersistentDisk FSType = "gcePersistentDisk"
- AWSElasticBlockStore FSType = "awsElasticBlockStore"
- GitRepo FSType = "gitRepo"
- Secret FSType = "secret"
- NFS FSType = "nfs"
- ISCSI FSType = "iscsi"
- Glusterfs FSType = "glusterfs"
- PersistentVolumeClaim FSType = "persistentVolumeClaim"
- RBD FSType = "rbd"
- Cinder FSType = "cinder"
- CephFS FSType = "cephFS"
- DownwardAPI FSType = "downwardAPI"
- FC FSType = "fc"
- ConfigMap FSType = "configMap"
- VsphereVolume FSType = "vsphereVolume"
- Quobyte FSType = "quobyte"
- AzureDisk FSType = "azureDisk"
- PhotonPersistentDisk FSType = "photonPersistentDisk"
- StorageOS FSType = "storageos"
- Projected FSType = "projected"
- PortworxVolume FSType = "portworxVolume"
- ScaleIO FSType = "scaleIO"
- CSI FSType = "csi"
- All FSType = "*"
-)
-
-// AllowedFlexVolume represents a single Flexvolume that is allowed to be used.
-type AllowedFlexVolume struct {
- // Driver is the name of the Flexvolume driver.
- Driver string
-}
-
-// SELinuxStrategyOptions defines the strategy type and any options used to create the strategy.
-type SELinuxStrategyOptions struct {
- // Rule is the strategy that will dictate the allowable labels that may be set.
- Rule SELinuxStrategy
- // seLinuxOptions required to run as; required for MustRunAs
- // More info: https://kubernetes.io/docs/concepts/policy/pod-security-policy/#selinux
- // +optional
- SELinuxOptions *api.SELinuxOptions
-}
-
-// SELinuxStrategy denotes strategy types for generating SELinux options for a
-// Security.
-type SELinuxStrategy string
-
-const (
- // container must have SELinux labels of X applied.
- SELinuxStrategyMustRunAs SELinuxStrategy = "MustRunAs"
- // container may make requests for any SELinux context labels.
- SELinuxStrategyRunAsAny SELinuxStrategy = "RunAsAny"
-)
-
-// RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy.
-type RunAsUserStrategyOptions struct {
- // Rule is the strategy that will dictate the allowable RunAsUser values that may be set.
- Rule RunAsUserStrategy
- // Ranges are the allowed ranges of uids that may be used.
- // +optional
- Ranges []UserIDRange
-}
-
-// UserIDRange provides a min/max of an allowed range of UserIDs.
-type UserIDRange struct {
- // Min is the start of the range, inclusive.
- Min int64
- // Max is the end of the range, inclusive.
- Max int64
-}
-
-// GroupIDRange provides a min/max of an allowed range of GroupIDs.
-type GroupIDRange struct {
- // Min is the start of the range, inclusive.
- Min int64
- // Max is the end of the range, inclusive.
- Max int64
-}
-
-// RunAsUserStrategy denotes strategy types for generating RunAsUser values for a
-// SecurityContext.
-type RunAsUserStrategy string
-
-const (
- // container must run as a particular uid.
- RunAsUserStrategyMustRunAs RunAsUserStrategy = "MustRunAs"
- // container must run as a non-root uid
- RunAsUserStrategyMustRunAsNonRoot RunAsUserStrategy = "MustRunAsNonRoot"
- // container may make requests for any uid.
- RunAsUserStrategyRunAsAny RunAsUserStrategy = "RunAsAny"
-)
-
-// FSGroupStrategyOptions defines the strategy type and options used to create the strategy.
-type FSGroupStrategyOptions struct {
- // Rule is the strategy that will dictate what FSGroup is used in the SecurityContext.
- // +optional
- Rule FSGroupStrategyType
- // Ranges are the allowed ranges of fs groups. If you would like to force a single
- // fs group then supply a single range with the same start and end.
- // +optional
- Ranges []GroupIDRange
-}
-
-// FSGroupStrategyType denotes strategy types for generating FSGroup values for a
-// SecurityContext
-type FSGroupStrategyType string
-
-const (
- // container must have FSGroup of X applied.
- FSGroupStrategyMustRunAs FSGroupStrategyType = "MustRunAs"
- // container may make requests for any FSGroup labels.
- FSGroupStrategyRunAsAny FSGroupStrategyType = "RunAsAny"
-)
-
-// SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy.
-type SupplementalGroupsStrategyOptions struct {
- // Rule is the strategy that will dictate what supplemental groups is used in the SecurityContext.
- // +optional
- Rule SupplementalGroupsStrategyType
- // Ranges are the allowed ranges of supplemental groups. If you would like to force a single
- // supplemental group then supply a single range with the same start and end.
- // +optional
- Ranges []GroupIDRange
-}
-
-// SupplementalGroupsStrategyType denotes strategy types for determining valid supplemental
-// groups for a SecurityContext.
-type SupplementalGroupsStrategyType string
-
-const (
- // container must run as a particular gid.
- SupplementalGroupsStrategyMustRunAs SupplementalGroupsStrategyType = "MustRunAs"
- // container may make requests for any gid.
- SupplementalGroupsStrategyRunAsAny SupplementalGroupsStrategyType = "RunAsAny"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// PodSecurityPolicyList is a list of PodSecurityPolicy objects.
-type PodSecurityPolicyList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []PodSecurityPolicy
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/extensions/zz_generated.deepcopy.go b/vendor/k8s.io/kubernetes/pkg/apis/extensions/zz_generated.deepcopy.go
deleted file mode 100644
index b7a91b05d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/extensions/zz_generated.deepcopy.go
+++ /dev/null
@@ -1,1231 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by deepcopy-gen. DO NOT EDIT.
-
-package extensions
-
-import (
- v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
- core "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AllowedFlexVolume) DeepCopyInto(out *AllowedFlexVolume) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AllowedFlexVolume.
-func (in *AllowedFlexVolume) DeepCopy() *AllowedFlexVolume {
- if in == nil {
- return nil
- }
- out := new(AllowedFlexVolume)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *AllowedHostPath) DeepCopyInto(out *AllowedHostPath) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AllowedHostPath.
-func (in *AllowedHostPath) DeepCopy() *AllowedHostPath {
- if in == nil {
- return nil
- }
- out := new(AllowedHostPath)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CustomMetricCurrentStatus) DeepCopyInto(out *CustomMetricCurrentStatus) {
- *out = *in
- out.CurrentValue = in.CurrentValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CustomMetricCurrentStatus.
-func (in *CustomMetricCurrentStatus) DeepCopy() *CustomMetricCurrentStatus {
- if in == nil {
- return nil
- }
- out := new(CustomMetricCurrentStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CustomMetricCurrentStatusList) DeepCopyInto(out *CustomMetricCurrentStatusList) {
- *out = *in
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]CustomMetricCurrentStatus, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CustomMetricCurrentStatusList.
-func (in *CustomMetricCurrentStatusList) DeepCopy() *CustomMetricCurrentStatusList {
- if in == nil {
- return nil
- }
- out := new(CustomMetricCurrentStatusList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CustomMetricTarget) DeepCopyInto(out *CustomMetricTarget) {
- *out = *in
- out.TargetValue = in.TargetValue.DeepCopy()
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CustomMetricTarget.
-func (in *CustomMetricTarget) DeepCopy() *CustomMetricTarget {
- if in == nil {
- return nil
- }
- out := new(CustomMetricTarget)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *CustomMetricTargetList) DeepCopyInto(out *CustomMetricTargetList) {
- *out = *in
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]CustomMetricTarget, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CustomMetricTargetList.
-func (in *CustomMetricTargetList) DeepCopy() *CustomMetricTargetList {
- if in == nil {
- return nil
- }
- out := new(CustomMetricTargetList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSet) DeepCopyInto(out *DaemonSet) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSet.
-func (in *DaemonSet) DeepCopy() *DaemonSet {
- if in == nil {
- return nil
- }
- out := new(DaemonSet)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *DaemonSet) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSetCondition) DeepCopyInto(out *DaemonSetCondition) {
- *out = *in
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSetCondition.
-func (in *DaemonSetCondition) DeepCopy() *DaemonSetCondition {
- if in == nil {
- return nil
- }
- out := new(DaemonSetCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSetList) DeepCopyInto(out *DaemonSetList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]DaemonSet, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSetList.
-func (in *DaemonSetList) DeepCopy() *DaemonSetList {
- if in == nil {
- return nil
- }
- out := new(DaemonSetList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *DaemonSetList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSetSpec) DeepCopyInto(out *DaemonSetSpec) {
- *out = *in
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- in.Template.DeepCopyInto(&out.Template)
- in.UpdateStrategy.DeepCopyInto(&out.UpdateStrategy)
- if in.RevisionHistoryLimit != nil {
- in, out := &in.RevisionHistoryLimit, &out.RevisionHistoryLimit
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSetSpec.
-func (in *DaemonSetSpec) DeepCopy() *DaemonSetSpec {
- if in == nil {
- return nil
- }
- out := new(DaemonSetSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSetStatus) DeepCopyInto(out *DaemonSetStatus) {
- *out = *in
- if in.CollisionCount != nil {
- in, out := &in.CollisionCount, &out.CollisionCount
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]DaemonSetCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSetStatus.
-func (in *DaemonSetStatus) DeepCopy() *DaemonSetStatus {
- if in == nil {
- return nil
- }
- out := new(DaemonSetStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DaemonSetUpdateStrategy) DeepCopyInto(out *DaemonSetUpdateStrategy) {
- *out = *in
- if in.RollingUpdate != nil {
- in, out := &in.RollingUpdate, &out.RollingUpdate
- if *in == nil {
- *out = nil
- } else {
- *out = new(RollingUpdateDaemonSet)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DaemonSetUpdateStrategy.
-func (in *DaemonSetUpdateStrategy) DeepCopy() *DaemonSetUpdateStrategy {
- if in == nil {
- return nil
- }
- out := new(DaemonSetUpdateStrategy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Deployment) DeepCopyInto(out *Deployment) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Deployment.
-func (in *Deployment) DeepCopy() *Deployment {
- if in == nil {
- return nil
- }
- out := new(Deployment)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Deployment) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentCondition) DeepCopyInto(out *DeploymentCondition) {
- *out = *in
- in.LastUpdateTime.DeepCopyInto(&out.LastUpdateTime)
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentCondition.
-func (in *DeploymentCondition) DeepCopy() *DeploymentCondition {
- if in == nil {
- return nil
- }
- out := new(DeploymentCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentList) DeepCopyInto(out *DeploymentList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Deployment, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentList.
-func (in *DeploymentList) DeepCopy() *DeploymentList {
- if in == nil {
- return nil
- }
- out := new(DeploymentList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *DeploymentList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentRollback) DeepCopyInto(out *DeploymentRollback) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.UpdatedAnnotations != nil {
- in, out := &in.UpdatedAnnotations, &out.UpdatedAnnotations
- *out = make(map[string]string, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- out.RollbackTo = in.RollbackTo
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentRollback.
-func (in *DeploymentRollback) DeepCopy() *DeploymentRollback {
- if in == nil {
- return nil
- }
- out := new(DeploymentRollback)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *DeploymentRollback) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentSpec) DeepCopyInto(out *DeploymentSpec) {
- *out = *in
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- in.Template.DeepCopyInto(&out.Template)
- in.Strategy.DeepCopyInto(&out.Strategy)
- if in.RevisionHistoryLimit != nil {
- in, out := &in.RevisionHistoryLimit, &out.RevisionHistoryLimit
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- if in.RollbackTo != nil {
- in, out := &in.RollbackTo, &out.RollbackTo
- if *in == nil {
- *out = nil
- } else {
- *out = new(RollbackConfig)
- **out = **in
- }
- }
- if in.ProgressDeadlineSeconds != nil {
- in, out := &in.ProgressDeadlineSeconds, &out.ProgressDeadlineSeconds
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentSpec.
-func (in *DeploymentSpec) DeepCopy() *DeploymentSpec {
- if in == nil {
- return nil
- }
- out := new(DeploymentSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentStatus) DeepCopyInto(out *DeploymentStatus) {
- *out = *in
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]DeploymentCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.CollisionCount != nil {
- in, out := &in.CollisionCount, &out.CollisionCount
- if *in == nil {
- *out = nil
- } else {
- *out = new(int32)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentStatus.
-func (in *DeploymentStatus) DeepCopy() *DeploymentStatus {
- if in == nil {
- return nil
- }
- out := new(DeploymentStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *DeploymentStrategy) DeepCopyInto(out *DeploymentStrategy) {
- *out = *in
- if in.RollingUpdate != nil {
- in, out := &in.RollingUpdate, &out.RollingUpdate
- if *in == nil {
- *out = nil
- } else {
- *out = new(RollingUpdateDeployment)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new DeploymentStrategy.
-func (in *DeploymentStrategy) DeepCopy() *DeploymentStrategy {
- if in == nil {
- return nil
- }
- out := new(DeploymentStrategy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *FSGroupStrategyOptions) DeepCopyInto(out *FSGroupStrategyOptions) {
- *out = *in
- if in.Ranges != nil {
- in, out := &in.Ranges, &out.Ranges
- *out = make([]GroupIDRange, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FSGroupStrategyOptions.
-func (in *FSGroupStrategyOptions) DeepCopy() *FSGroupStrategyOptions {
- if in == nil {
- return nil
- }
- out := new(FSGroupStrategyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *GroupIDRange) DeepCopyInto(out *GroupIDRange) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new GroupIDRange.
-func (in *GroupIDRange) DeepCopy() *GroupIDRange {
- if in == nil {
- return nil
- }
- out := new(GroupIDRange)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HTTPIngressPath) DeepCopyInto(out *HTTPIngressPath) {
- *out = *in
- out.Backend = in.Backend
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPIngressPath.
-func (in *HTTPIngressPath) DeepCopy() *HTTPIngressPath {
- if in == nil {
- return nil
- }
- out := new(HTTPIngressPath)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HTTPIngressRuleValue) DeepCopyInto(out *HTTPIngressRuleValue) {
- *out = *in
- if in.Paths != nil {
- in, out := &in.Paths, &out.Paths
- *out = make([]HTTPIngressPath, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPIngressRuleValue.
-func (in *HTTPIngressRuleValue) DeepCopy() *HTTPIngressRuleValue {
- if in == nil {
- return nil
- }
- out := new(HTTPIngressRuleValue)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HostPortRange) DeepCopyInto(out *HostPortRange) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HostPortRange.
-func (in *HostPortRange) DeepCopy() *HostPortRange {
- if in == nil {
- return nil
- }
- out := new(HostPortRange)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Ingress) DeepCopyInto(out *Ingress) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Ingress.
-func (in *Ingress) DeepCopy() *Ingress {
- if in == nil {
- return nil
- }
- out := new(Ingress)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Ingress) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressBackend) DeepCopyInto(out *IngressBackend) {
- *out = *in
- out.ServicePort = in.ServicePort
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressBackend.
-func (in *IngressBackend) DeepCopy() *IngressBackend {
- if in == nil {
- return nil
- }
- out := new(IngressBackend)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressList) DeepCopyInto(out *IngressList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]Ingress, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressList.
-func (in *IngressList) DeepCopy() *IngressList {
- if in == nil {
- return nil
- }
- out := new(IngressList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *IngressList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressRule) DeepCopyInto(out *IngressRule) {
- *out = *in
- in.IngressRuleValue.DeepCopyInto(&out.IngressRuleValue)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressRule.
-func (in *IngressRule) DeepCopy() *IngressRule {
- if in == nil {
- return nil
- }
- out := new(IngressRule)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressRuleValue) DeepCopyInto(out *IngressRuleValue) {
- *out = *in
- if in.HTTP != nil {
- in, out := &in.HTTP, &out.HTTP
- if *in == nil {
- *out = nil
- } else {
- *out = new(HTTPIngressRuleValue)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressRuleValue.
-func (in *IngressRuleValue) DeepCopy() *IngressRuleValue {
- if in == nil {
- return nil
- }
- out := new(IngressRuleValue)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressSpec) DeepCopyInto(out *IngressSpec) {
- *out = *in
- if in.Backend != nil {
- in, out := &in.Backend, &out.Backend
- if *in == nil {
- *out = nil
- } else {
- *out = new(IngressBackend)
- **out = **in
- }
- }
- if in.TLS != nil {
- in, out := &in.TLS, &out.TLS
- *out = make([]IngressTLS, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Rules != nil {
- in, out := &in.Rules, &out.Rules
- *out = make([]IngressRule, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressSpec.
-func (in *IngressSpec) DeepCopy() *IngressSpec {
- if in == nil {
- return nil
- }
- out := new(IngressSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressStatus) DeepCopyInto(out *IngressStatus) {
- *out = *in
- in.LoadBalancer.DeepCopyInto(&out.LoadBalancer)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressStatus.
-func (in *IngressStatus) DeepCopy() *IngressStatus {
- if in == nil {
- return nil
- }
- out := new(IngressStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IngressTLS) DeepCopyInto(out *IngressTLS) {
- *out = *in
- if in.Hosts != nil {
- in, out := &in.Hosts, &out.Hosts
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressTLS.
-func (in *IngressTLS) DeepCopy() *IngressTLS {
- if in == nil {
- return nil
- }
- out := new(IngressTLS)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSecurityPolicy) DeepCopyInto(out *PodSecurityPolicy) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSecurityPolicy.
-func (in *PodSecurityPolicy) DeepCopy() *PodSecurityPolicy {
- if in == nil {
- return nil
- }
- out := new(PodSecurityPolicy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodSecurityPolicy) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSecurityPolicyList) DeepCopyInto(out *PodSecurityPolicyList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]PodSecurityPolicy, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSecurityPolicyList.
-func (in *PodSecurityPolicyList) DeepCopy() *PodSecurityPolicyList {
- if in == nil {
- return nil
- }
- out := new(PodSecurityPolicyList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *PodSecurityPolicyList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PodSecurityPolicySpec) DeepCopyInto(out *PodSecurityPolicySpec) {
- *out = *in
- if in.DefaultAddCapabilities != nil {
- in, out := &in.DefaultAddCapabilities, &out.DefaultAddCapabilities
- *out = make([]core.Capability, len(*in))
- copy(*out, *in)
- }
- if in.RequiredDropCapabilities != nil {
- in, out := &in.RequiredDropCapabilities, &out.RequiredDropCapabilities
- *out = make([]core.Capability, len(*in))
- copy(*out, *in)
- }
- if in.AllowedCapabilities != nil {
- in, out := &in.AllowedCapabilities, &out.AllowedCapabilities
- *out = make([]core.Capability, len(*in))
- copy(*out, *in)
- }
- if in.Volumes != nil {
- in, out := &in.Volumes, &out.Volumes
- *out = make([]FSType, len(*in))
- copy(*out, *in)
- }
- if in.HostPorts != nil {
- in, out := &in.HostPorts, &out.HostPorts
- *out = make([]HostPortRange, len(*in))
- copy(*out, *in)
- }
- in.SELinux.DeepCopyInto(&out.SELinux)
- in.RunAsUser.DeepCopyInto(&out.RunAsUser)
- in.SupplementalGroups.DeepCopyInto(&out.SupplementalGroups)
- in.FSGroup.DeepCopyInto(&out.FSGroup)
- if in.DefaultAllowPrivilegeEscalation != nil {
- in, out := &in.DefaultAllowPrivilegeEscalation, &out.DefaultAllowPrivilegeEscalation
- if *in == nil {
- *out = nil
- } else {
- *out = new(bool)
- **out = **in
- }
- }
- if in.AllowedHostPaths != nil {
- in, out := &in.AllowedHostPaths, &out.AllowedHostPaths
- *out = make([]AllowedHostPath, len(*in))
- copy(*out, *in)
- }
- if in.AllowedFlexVolumes != nil {
- in, out := &in.AllowedFlexVolumes, &out.AllowedFlexVolumes
- *out = make([]AllowedFlexVolume, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSecurityPolicySpec.
-func (in *PodSecurityPolicySpec) DeepCopy() *PodSecurityPolicySpec {
- if in == nil {
- return nil
- }
- out := new(PodSecurityPolicySpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicaSet) DeepCopyInto(out *ReplicaSet) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- in.Status.DeepCopyInto(&out.Status)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicaSet.
-func (in *ReplicaSet) DeepCopy() *ReplicaSet {
- if in == nil {
- return nil
- }
- out := new(ReplicaSet)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ReplicaSet) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicaSetCondition) DeepCopyInto(out *ReplicaSetCondition) {
- *out = *in
- in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicaSetCondition.
-func (in *ReplicaSetCondition) DeepCopy() *ReplicaSetCondition {
- if in == nil {
- return nil
- }
- out := new(ReplicaSetCondition)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicaSetList) DeepCopyInto(out *ReplicaSetList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]ReplicaSet, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicaSetList.
-func (in *ReplicaSetList) DeepCopy() *ReplicaSetList {
- if in == nil {
- return nil
- }
- out := new(ReplicaSetList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ReplicaSetList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicaSetSpec) DeepCopyInto(out *ReplicaSetSpec) {
- *out = *in
- if in.Selector != nil {
- in, out := &in.Selector, &out.Selector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- in.Template.DeepCopyInto(&out.Template)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicaSetSpec.
-func (in *ReplicaSetSpec) DeepCopy() *ReplicaSetSpec {
- if in == nil {
- return nil
- }
- out := new(ReplicaSetSpec)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicaSetStatus) DeepCopyInto(out *ReplicaSetStatus) {
- *out = *in
- if in.Conditions != nil {
- in, out := &in.Conditions, &out.Conditions
- *out = make([]ReplicaSetCondition, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicaSetStatus.
-func (in *ReplicaSetStatus) DeepCopy() *ReplicaSetStatus {
- if in == nil {
- return nil
- }
- out := new(ReplicaSetStatus)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ReplicationControllerDummy) DeepCopyInto(out *ReplicationControllerDummy) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerDummy.
-func (in *ReplicationControllerDummy) DeepCopy() *ReplicationControllerDummy {
- if in == nil {
- return nil
- }
- out := new(ReplicationControllerDummy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *ReplicationControllerDummy) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RollbackConfig) DeepCopyInto(out *RollbackConfig) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RollbackConfig.
-func (in *RollbackConfig) DeepCopy() *RollbackConfig {
- if in == nil {
- return nil
- }
- out := new(RollbackConfig)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RollingUpdateDaemonSet) DeepCopyInto(out *RollingUpdateDaemonSet) {
- *out = *in
- out.MaxUnavailable = in.MaxUnavailable
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RollingUpdateDaemonSet.
-func (in *RollingUpdateDaemonSet) DeepCopy() *RollingUpdateDaemonSet {
- if in == nil {
- return nil
- }
- out := new(RollingUpdateDaemonSet)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RollingUpdateDeployment) DeepCopyInto(out *RollingUpdateDeployment) {
- *out = *in
- out.MaxUnavailable = in.MaxUnavailable
- out.MaxSurge = in.MaxSurge
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RollingUpdateDeployment.
-func (in *RollingUpdateDeployment) DeepCopy() *RollingUpdateDeployment {
- if in == nil {
- return nil
- }
- out := new(RollingUpdateDeployment)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *RunAsUserStrategyOptions) DeepCopyInto(out *RunAsUserStrategyOptions) {
- *out = *in
- if in.Ranges != nil {
- in, out := &in.Ranges, &out.Ranges
- *out = make([]UserIDRange, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new RunAsUserStrategyOptions.
-func (in *RunAsUserStrategyOptions) DeepCopy() *RunAsUserStrategyOptions {
- if in == nil {
- return nil
- }
- out := new(RunAsUserStrategyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SELinuxStrategyOptions) DeepCopyInto(out *SELinuxStrategyOptions) {
- *out = *in
- if in.SELinuxOptions != nil {
- in, out := &in.SELinuxOptions, &out.SELinuxOptions
- if *in == nil {
- *out = nil
- } else {
- *out = new(core.SELinuxOptions)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SELinuxStrategyOptions.
-func (in *SELinuxStrategyOptions) DeepCopy() *SELinuxStrategyOptions {
- if in == nil {
- return nil
- }
- out := new(SELinuxStrategyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *SupplementalGroupsStrategyOptions) DeepCopyInto(out *SupplementalGroupsStrategyOptions) {
- *out = *in
- if in.Ranges != nil {
- in, out := &in.Ranges, &out.Ranges
- *out = make([]GroupIDRange, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SupplementalGroupsStrategyOptions.
-func (in *SupplementalGroupsStrategyOptions) DeepCopy() *SupplementalGroupsStrategyOptions {
- if in == nil {
- return nil
- }
- out := new(SupplementalGroupsStrategyOptions)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *UserIDRange) DeepCopyInto(out *UserIDRange) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new UserIDRange.
-func (in *UserIDRange) DeepCopy() *UserIDRange {
- if in == nil {
- return nil
- }
- out := new(UserIDRange)
- in.DeepCopyInto(out)
- return out
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/networking/doc.go b/vendor/k8s.io/kubernetes/pkg/apis/networking/doc.go
deleted file mode 100644
index 8b013e348..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/networking/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:deepcopy-gen=package
-// +groupName=networking.k8s.io
-package networking // import "k8s.io/kubernetes/pkg/apis/networking"
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/networking/register.go b/vendor/k8s.io/kubernetes/pkg/apis/networking/register.go
deleted file mode 100644
index 4d58a210d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/networking/register.go
+++ /dev/null
@@ -1,51 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package networking
-
-import (
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-// GroupName is the group name use in this package
-const GroupName = "networking.k8s.io"
-
-// SchemeGroupVersion is group version used to register these objects
-var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: runtime.APIVersionInternal}
-
-// Kind takes an unqualified kind and returns a Group qualified GroupKind
-func Kind(kind string) schema.GroupKind {
- return SchemeGroupVersion.WithKind(kind).GroupKind()
-}
-
-// Resource takes an unqualified resource and returns a Group qualified GroupResource
-func Resource(resource string) schema.GroupResource {
- return SchemeGroupVersion.WithResource(resource).GroupResource()
-}
-
-var (
- SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes)
- AddToScheme = SchemeBuilder.AddToScheme
-)
-
-func addKnownTypes(scheme *runtime.Scheme) error {
- scheme.AddKnownTypes(SchemeGroupVersion,
- &NetworkPolicy{},
- &NetworkPolicyList{},
- )
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/networking/types.go b/vendor/k8s.io/kubernetes/pkg/apis/networking/types.go
deleted file mode 100644
index ae37fcd5e..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/networking/types.go
+++ /dev/null
@@ -1,191 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package networking
-
-import (
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/util/intstr"
- api "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// +genclient
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NetworkPolicy describes what network traffic is allowed for a set of Pods
-type NetworkPolicy struct {
- metav1.TypeMeta
- // +optional
- metav1.ObjectMeta
-
- // Specification of the desired behavior for this NetworkPolicy.
- // +optional
- Spec NetworkPolicySpec
-}
-
-// Policy Type string describes the NetworkPolicy type
-// This type is beta-level in 1.8
-type PolicyType string
-
-const (
- // PolicyTypeIngress is a NetworkPolicy that affects ingress traffic on selected pods
- PolicyTypeIngress PolicyType = "Ingress"
- // PolicyTypeEgress is a NetworkPolicy that affects egress traffic on selected pods
- PolicyTypeEgress PolicyType = "Egress"
-)
-
-// NetworkPolicySpec provides the specification of a NetworkPolicy
-type NetworkPolicySpec struct {
- // Selects the pods to which this NetworkPolicy object applies. The array of
- // ingress rules is applied to any pods selected by this field. Multiple network
- // policies can select the same set of pods. In this case, the ingress rules for
- // each are combined additively. This field is NOT optional and follows standard
- // label selector semantics. An empty podSelector matches all pods in this
- // namespace.
- PodSelector metav1.LabelSelector
-
- // List of ingress rules to be applied to the selected pods. Traffic is allowed to
- // a pod if there are no NetworkPolicies selecting the pod
- // (and cluster policy otherwise allows the traffic), OR if the traffic source is
- // the pod's local node, OR if the traffic matches at least one ingress rule
- // across all of the NetworkPolicy objects whose podSelector matches the pod. If
- // this field is empty then this NetworkPolicy does not allow any traffic (and serves
- // solely to ensure that the pods it selects are isolated by default)
- // +optional
- Ingress []NetworkPolicyIngressRule
-
- // List of egress rules to be applied to the selected pods. Outgoing traffic is
- // allowed if there are no NetworkPolicies selecting the pod (and cluster policy
- // otherwise allows the traffic), OR if the traffic matches at least one egress rule
- // across all of the NetworkPolicy objects whose podSelector matches the pod. If
- // this field is empty then this NetworkPolicy limits all outgoing traffic (and serves
- // solely to ensure that the pods it selects are isolated by default).
- // This field is beta-level in 1.8
- // +optional
- Egress []NetworkPolicyEgressRule
-
- // List of rule types that the NetworkPolicy relates to.
- // Valid options are Ingress, Egress, or Ingress,Egress.
- // If this field is not specified, it will default based on the existence of Ingress or Egress rules;
- // policies that contain an Egress section are assumed to affect Egress, and all policies
- // (whether or not they contain an Ingress section) are assumed to affect Ingress.
- // If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ].
- // Likewise, if you want to write a policy that specifies that no egress is allowed,
- // you must specify a policyTypes value that include "Egress" (since such a policy would not include
- // an Egress section and would otherwise default to just [ "Ingress" ]).
- // This field is beta-level in 1.8
- // +optional
- PolicyTypes []PolicyType
-}
-
-// NetworkPolicyIngressRule describes a particular set of traffic that is allowed to the pods
-// matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and from.
-type NetworkPolicyIngressRule struct {
- // List of ports which should be made accessible on the pods selected for this
- // rule. Each item in this list is combined using a logical OR. If this field is
- // empty or missing, this rule matches all ports (traffic not restricted by port).
- // If this field is present and contains at least one item, then this rule allows
- // traffic only if the traffic matches at least one port in the list.
- // +optional
- Ports []NetworkPolicyPort
-
- // List of sources which should be able to access the pods selected for this rule.
- // Items in this list are combined using a logical OR operation. If this field is
- // empty or missing, this rule matches all sources (traffic not restricted by
- // source). If this field is present and contains at least on item, this rule
- // allows traffic only if the traffic matches at least one item in the from list.
- // +optional
- From []NetworkPolicyPeer
-}
-
-// NetworkPolicyEgressRule describes a particular set of traffic that is allowed out of pods
-// matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and to.
-// This type is beta-level in 1.8
-type NetworkPolicyEgressRule struct {
- // List of destination ports for outgoing traffic.
- // Each item in this list is combined using a logical OR. If this field is
- // empty or missing, this rule matches all ports (traffic not restricted by port).
- // If this field is present and contains at least one item, then this rule allows
- // traffic only if the traffic matches at least one port in the list.
- // +optional
- Ports []NetworkPolicyPort
-
- // List of destinations for outgoing traffic of pods selected for this rule.
- // Items in this list are combined using a logical OR operation. If this field is
- // empty or missing, this rule matches all destinations (traffic not restricted by
- // destination). If this field is present and contains at least one item, this rule
- // allows traffic only if the traffic matches at least one item in the to list.
- // +optional
- To []NetworkPolicyPeer
-}
-
-// NetworkPolicyPort describes a port to allow traffic on
-type NetworkPolicyPort struct {
- // The protocol (TCP or UDP) which traffic must match. If not specified, this
- // field defaults to TCP.
- // +optional
- Protocol *api.Protocol
-
- // The port on the given protocol. This can either be a numerical or named port on
- // a pod. If this field is not provided, this matches all port names and numbers.
- // +optional
- Port *intstr.IntOrString
-}
-
-// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods
-// matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should
-// not be included within this rule.
-type IPBlock struct {
- // CIDR is a string representing the IP Block
- // Valid examples are "192.168.1.1/24"
- CIDR string
- // Except is a slice of CIDRs that should not be included within an IP Block
- // Valid examples are "192.168.1.1/24"
- // Except values will be rejected if they are outside the CIDR range
- // +optional
- Except []string
-}
-
-// NetworkPolicyPeer describes a peer to allow traffic from. Exactly one of its fields
-// must be specified.
-type NetworkPolicyPeer struct {
- // This is a label selector which selects Pods in this namespace. This field
- // follows standard label selector semantics. If present but empty, this selector
- // selects all pods in this namespace.
- // +optional
- PodSelector *metav1.LabelSelector
-
- // Selects Namespaces using cluster scoped-labels. This matches all pods in all
- // namespaces selected by this label selector. This field follows standard label
- // selector semantics. If present but empty, this selector selects all namespaces.
- // +optional
- NamespaceSelector *metav1.LabelSelector
-
- // IPBlock defines policy on a particular IPBlock
- // +optional
- IPBlock *IPBlock
-}
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// NetworkPolicyList is a list of NetworkPolicy objects.
-type NetworkPolicyList struct {
- metav1.TypeMeta
- // +optional
- metav1.ListMeta
-
- Items []NetworkPolicy
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/apis/networking/zz_generated.deepcopy.go b/vendor/k8s.io/kubernetes/pkg/apis/networking/zz_generated.deepcopy.go
deleted file mode 100644
index 47121a572..000000000
--- a/vendor/k8s.io/kubernetes/pkg/apis/networking/zz_generated.deepcopy.go
+++ /dev/null
@@ -1,282 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by deepcopy-gen. DO NOT EDIT.
-
-package networking
-
-import (
- v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
- intstr "k8s.io/apimachinery/pkg/util/intstr"
- core "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *IPBlock) DeepCopyInto(out *IPBlock) {
- *out = *in
- if in.Except != nil {
- in, out := &in.Except, &out.Except
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IPBlock.
-func (in *IPBlock) DeepCopy() *IPBlock {
- if in == nil {
- return nil
- }
- out := new(IPBlock)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicy) DeepCopyInto(out *NetworkPolicy) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
- in.Spec.DeepCopyInto(&out.Spec)
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicy.
-func (in *NetworkPolicy) DeepCopy() *NetworkPolicy {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NetworkPolicy) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicyEgressRule) DeepCopyInto(out *NetworkPolicyEgressRule) {
- *out = *in
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]NetworkPolicyPort, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.To != nil {
- in, out := &in.To, &out.To
- *out = make([]NetworkPolicyPeer, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicyEgressRule.
-func (in *NetworkPolicyEgressRule) DeepCopy() *NetworkPolicyEgressRule {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicyEgressRule)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicyIngressRule) DeepCopyInto(out *NetworkPolicyIngressRule) {
- *out = *in
- if in.Ports != nil {
- in, out := &in.Ports, &out.Ports
- *out = make([]NetworkPolicyPort, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.From != nil {
- in, out := &in.From, &out.From
- *out = make([]NetworkPolicyPeer, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicyIngressRule.
-func (in *NetworkPolicyIngressRule) DeepCopy() *NetworkPolicyIngressRule {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicyIngressRule)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicyList) DeepCopyInto(out *NetworkPolicyList) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- out.ListMeta = in.ListMeta
- if in.Items != nil {
- in, out := &in.Items, &out.Items
- *out = make([]NetworkPolicy, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicyList.
-func (in *NetworkPolicyList) DeepCopy() *NetworkPolicyList {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicyList)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *NetworkPolicyList) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicyPeer) DeepCopyInto(out *NetworkPolicyPeer) {
- *out = *in
- if in.PodSelector != nil {
- in, out := &in.PodSelector, &out.PodSelector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.NamespaceSelector != nil {
- in, out := &in.NamespaceSelector, &out.NamespaceSelector
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.LabelSelector)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.IPBlock != nil {
- in, out := &in.IPBlock, &out.IPBlock
- if *in == nil {
- *out = nil
- } else {
- *out = new(IPBlock)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicyPeer.
-func (in *NetworkPolicyPeer) DeepCopy() *NetworkPolicyPeer {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicyPeer)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicyPort) DeepCopyInto(out *NetworkPolicyPort) {
- *out = *in
- if in.Protocol != nil {
- in, out := &in.Protocol, &out.Protocol
- if *in == nil {
- *out = nil
- } else {
- *out = new(core.Protocol)
- **out = **in
- }
- }
- if in.Port != nil {
- in, out := &in.Port, &out.Port
- if *in == nil {
- *out = nil
- } else {
- *out = new(intstr.IntOrString)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicyPort.
-func (in *NetworkPolicyPort) DeepCopy() *NetworkPolicyPort {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicyPort)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *NetworkPolicySpec) DeepCopyInto(out *NetworkPolicySpec) {
- *out = *in
- in.PodSelector.DeepCopyInto(&out.PodSelector)
- if in.Ingress != nil {
- in, out := &in.Ingress, &out.Ingress
- *out = make([]NetworkPolicyIngressRule, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Egress != nil {
- in, out := &in.Egress, &out.Egress
- *out = make([]NetworkPolicyEgressRule, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.PolicyTypes != nil {
- in, out := &in.PolicyTypes, &out.PolicyTypes
- *out = make([]PolicyType, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new NetworkPolicySpec.
-func (in *NetworkPolicySpec) DeepCopy() *NetworkPolicySpec {
- if in == nil {
- return nil
- }
- out := new(NetworkPolicySpec)
- in.DeepCopyInto(out)
- return out
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/capabilities/capabilities.go b/vendor/k8s.io/kubernetes/pkg/capabilities/capabilities.go
deleted file mode 100644
index 0da7b9c8b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/capabilities/capabilities.go
+++ /dev/null
@@ -1,95 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package capabilities
-
-import (
- "sync"
-)
-
-// Capabilities defines the set of capabilities available within the system.
-// For now these are global. Eventually they may be per-user
-type Capabilities struct {
- AllowPrivileged bool
-
- // Pod sources from which to allow privileged capabilities like host networking, sharing the host
- // IPC namespace, and sharing the host PID namespace.
- PrivilegedSources PrivilegedSources
-
- // PerConnectionBandwidthLimitBytesPerSec limits the throughput of each connection (currently only used for proxy, exec, attach)
- PerConnectionBandwidthLimitBytesPerSec int64
-}
-
-// PrivilegedSources defines the pod sources allowed to make privileged requests for certain types
-// of capabilities like host networking, sharing the host IPC namespace, and sharing the host PID namespace.
-type PrivilegedSources struct {
- // List of pod sources for which using host network is allowed.
- HostNetworkSources []string
-
- // List of pod sources for which using host pid namespace is allowed.
- HostPIDSources []string
-
- // List of pod sources for which using host ipc is allowed.
- HostIPCSources []string
-}
-
-var capInstance struct {
- once sync.Once
- lock sync.Mutex
- capabilities *Capabilities
-}
-
-// Initialize the capability set. This can only be done once per binary, subsequent calls are ignored.
-func Initialize(c Capabilities) {
- // Only do this once
- capInstance.once.Do(func() {
- capInstance.capabilities = &c
- })
-}
-
-// Setup the capability set. It wraps Initialize for improving usability.
-func Setup(allowPrivileged bool, privilegedSources PrivilegedSources, perConnectionBytesPerSec int64) {
- Initialize(Capabilities{
- AllowPrivileged: allowPrivileged,
- PrivilegedSources: privilegedSources,
- PerConnectionBandwidthLimitBytesPerSec: perConnectionBytesPerSec,
- })
-}
-
-// SetForTests sets capabilities for tests. Convenience method for testing. This should only be called from tests.
-func SetForTests(c Capabilities) {
- capInstance.lock.Lock()
- defer capInstance.lock.Unlock()
- capInstance.capabilities = &c
-}
-
-// Returns a read-only copy of the system capabilities.
-func Get() Capabilities {
- capInstance.lock.Lock()
- defer capInstance.lock.Unlock()
- // This check prevents clobbering of capabilities that might've been set via SetForTests
- if capInstance.capabilities == nil {
- Initialize(Capabilities{
- AllowPrivileged: false,
- PrivilegedSources: PrivilegedSources{
- HostNetworkSources: []string{},
- HostPIDSources: []string{},
- HostIPCSources: []string{},
- },
- })
- }
- return *capInstance.capabilities
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/capabilities/doc.go b/vendor/k8s.io/kubernetes/pkg/capabilities/doc.go
deleted file mode 100644
index bbdc89d50..000000000
--- a/vendor/k8s.io/kubernetes/pkg/capabilities/doc.go
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// package capabilities manages system level capabilities
-package capabilities // import "k8s.io/kubernetes/pkg/capabilities"
diff --git a/vendor/k8s.io/kubernetes/pkg/cloudprovider/README.md b/vendor/k8s.io/kubernetes/pkg/cloudprovider/README.md
deleted file mode 100644
index b59eba409..000000000
--- a/vendor/k8s.io/kubernetes/pkg/cloudprovider/README.md
+++ /dev/null
@@ -1,16 +0,0 @@
-##### Deprecation Notice: This directory has entered maintenance mode and will not be accepting new providers. Cloud Providers in this directory will continue to be actively developed or maintained and supported at their current level of support as a longer-term solution evolves.
-
-## Overview:
-The mechanism for supporting cloud providers is currently in transition: the original method of implementing cloud provider-specific functionality within the main kubernetes tree (here) is no longer advised; however, the proposed solution is still in development.
-
-#### Guidance for potential cloud providers:
-* Support for cloud providers is currently in a state of flux. Background information on motivation and the proposal for improving is in the github [proposal](https://git.k8s.io/community/contributors/design-proposals/cloud-provider/cloud-provider-refactoring.md).
-* In support of this plan, a new cloud-controller-manager binary was added in 1.6. This was the first of several steps (see the proposal for more information).
-* Attempts to contribute new cloud providers or (to a lesser extent) persistent volumes to the core repo will likely meet with some pushback from reviewers/approvers.
-* It is understood that this is an unfortunate situation in which 'the old way is no longer supported but the new way is not ready yet', but the initial path is unsustainable, and contributors are encouraged to participate in the implementation of the proposed long-term solution, as there is risk that PRs for new cloud providers here will not be approved.
-* Though the fully productized support envisioned in the proposal is still 2 - 3 releases out, the foundational work is underway, and a motivated cloud provider could accomplish the work in a forward-looking way. Contributors are encouraged to assist with the implementation of the design outlined in the proposal.
-
-#### Some additional context on status / direction:
-* 1.6 added a new cloud-controller-manager binary that may be used for testing the new out-of-core cloudprovider flow.
-* Setting cloud-provider=external allows for creation of a separate controller-manager binary
-* 1.7 adds [extensible admission control](https://git.k8s.io/community/contributors/design-proposals/api-machinery/admission_control_extension.md), further enabling topology customization.
diff --git a/vendor/k8s.io/kubernetes/pkg/cloudprovider/cloud.go b/vendor/k8s.io/kubernetes/pkg/cloudprovider/cloud.go
deleted file mode 100644
index 9ca91ebf3..000000000
--- a/vendor/k8s.io/kubernetes/pkg/cloudprovider/cloud.go
+++ /dev/null
@@ -1,215 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package cloudprovider
-
-import (
- "context"
- "errors"
- "fmt"
- "strings"
-
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/types"
- "k8s.io/client-go/informers"
- "k8s.io/kubernetes/pkg/controller"
-)
-
-// Interface is an abstract, pluggable interface for cloud providers.
-type Interface interface {
- // Initialize provides the cloud with a kubernetes client builder and may spawn goroutines
- // to perform housekeeping activities within the cloud provider.
- Initialize(clientBuilder controller.ControllerClientBuilder)
- // LoadBalancer returns a balancer interface. Also returns true if the interface is supported, false otherwise.
- LoadBalancer() (LoadBalancer, bool)
- // Instances returns an instances interface. Also returns true if the interface is supported, false otherwise.
- Instances() (Instances, bool)
- // Zones returns a zones interface. Also returns true if the interface is supported, false otherwise.
- Zones() (Zones, bool)
- // Clusters returns a clusters interface. Also returns true if the interface is supported, false otherwise.
- Clusters() (Clusters, bool)
- // Routes returns a routes interface along with whether the interface is supported.
- Routes() (Routes, bool)
- // ProviderName returns the cloud provider ID.
- ProviderName() string
- // HasClusterID returns true if a ClusterID is required and set
- HasClusterID() bool
-}
-
-type InformerUser interface {
- // SetInformers sets the informer on the cloud object.
- SetInformers(informerFactory informers.SharedInformerFactory)
-}
-
-// Clusters is an abstract, pluggable interface for clusters of containers.
-type Clusters interface {
- // ListClusters lists the names of the available clusters.
- ListClusters(ctx context.Context) ([]string, error)
- // Master gets back the address (either DNS name or IP address) of the master node for the cluster.
- Master(ctx context.Context, clusterName string) (string, error)
-}
-
-// TODO(#6812): Use a shorter name that's less likely to be longer than cloud
-// providers' name length limits.
-func GetLoadBalancerName(service *v1.Service) string {
- //GCE requires that the name of a load balancer starts with a lower case letter.
- ret := "a" + string(service.UID)
- ret = strings.Replace(ret, "-", "", -1)
- //AWS requires that the name of a load balancer is shorter than 32 bytes.
- if len(ret) > 32 {
- ret = ret[:32]
- }
- return ret
-}
-
-// GetInstanceProviderID builds a ProviderID for a node in a cloud.
-func GetInstanceProviderID(ctx context.Context, cloud Interface, nodeName types.NodeName) (string, error) {
- instances, ok := cloud.Instances()
- if !ok {
- return "", fmt.Errorf("failed to get instances from cloud provider")
- }
- instanceID, err := instances.InstanceID(ctx, nodeName)
- if err != nil {
- return "", fmt.Errorf("failed to get instance ID from cloud provider: %v", err)
- }
- return cloud.ProviderName() + "://" + instanceID, nil
-}
-
-// LoadBalancer is an abstract, pluggable interface for load balancers.
-type LoadBalancer interface {
- // TODO: Break this up into different interfaces (LB, etc) when we have more than one type of service
- // GetLoadBalancer returns whether the specified load balancer exists, and
- // if so, what its status is.
- // Implementations must treat the *v1.Service parameter as read-only and not modify it.
- // Parameter 'clusterName' is the name of the cluster as presented to kube-controller-manager
- GetLoadBalancer(ctx context.Context, clusterName string, service *v1.Service) (status *v1.LoadBalancerStatus, exists bool, err error)
- // EnsureLoadBalancer creates a new load balancer 'name', or updates the existing one. Returns the status of the balancer
- // Implementations must treat the *v1.Service and *v1.Node
- // parameters as read-only and not modify them.
- // Parameter 'clusterName' is the name of the cluster as presented to kube-controller-manager
- EnsureLoadBalancer(ctx context.Context, clusterName string, service *v1.Service, nodes []*v1.Node) (*v1.LoadBalancerStatus, error)
- // UpdateLoadBalancer updates hosts under the specified load balancer.
- // Implementations must treat the *v1.Service and *v1.Node
- // parameters as read-only and not modify them.
- // Parameter 'clusterName' is the name of the cluster as presented to kube-controller-manager
- UpdateLoadBalancer(ctx context.Context, clusterName string, service *v1.Service, nodes []*v1.Node) error
- // EnsureLoadBalancerDeleted deletes the specified load balancer if it
- // exists, returning nil if the load balancer specified either didn't exist or
- // was successfully deleted.
- // This construction is useful because many cloud providers' load balancers
- // have multiple underlying components, meaning a Get could say that the LB
- // doesn't exist even if some part of it is still laying around.
- // Implementations must treat the *v1.Service parameter as read-only and not modify it.
- // Parameter 'clusterName' is the name of the cluster as presented to kube-controller-manager
- EnsureLoadBalancerDeleted(ctx context.Context, clusterName string, service *v1.Service) error
-}
-
-// Instances is an abstract, pluggable interface for sets of instances.
-type Instances interface {
- // NodeAddresses returns the addresses of the specified instance.
- // TODO(roberthbailey): This currently is only used in such a way that it
- // returns the address of the calling instance. We should do a rename to
- // make this clearer.
- NodeAddresses(ctx context.Context, name types.NodeName) ([]v1.NodeAddress, error)
- // NodeAddressesByProviderID returns the addresses of the specified instance.
- // The instance is specified using the providerID of the node. The
- // ProviderID is a unique identifier of the node. This will not be called
- // from the node whose nodeaddresses are being queried. i.e. local metadata
- // services cannot be used in this method to obtain nodeaddresses
- NodeAddressesByProviderID(ctx context.Context, providerID string) ([]v1.NodeAddress, error)
- // ExternalID returns the cloud provider ID of the node with the specified NodeName.
- // Note that if the instance does not exist or is no longer running, we must return ("", cloudprovider.InstanceNotFound)
- ExternalID(ctx context.Context, nodeName types.NodeName) (string, error)
- // InstanceID returns the cloud provider ID of the node with the specified NodeName.
- InstanceID(ctx context.Context, nodeName types.NodeName) (string, error)
- // InstanceType returns the type of the specified instance.
- InstanceType(ctx context.Context, name types.NodeName) (string, error)
- // InstanceTypeByProviderID returns the type of the specified instance.
- InstanceTypeByProviderID(ctx context.Context, providerID string) (string, error)
- // AddSSHKeyToAllInstances adds an SSH public key as a legal identity for all instances
- // expected format for the key is standard ssh-keygen format: <protocol> <blob>
- AddSSHKeyToAllInstances(ctx context.Context, user string, keyData []byte) error
- // CurrentNodeName returns the name of the node we are currently running on
- // On most clouds (e.g. GCE) this is the hostname, so we provide the hostname
- CurrentNodeName(ctx context.Context, hostname string) (types.NodeName, error)
- // InstanceExistsByProviderID returns true if the instance for the given provider id still is running.
- // If false is returned with no error, the instance will be immediately deleted by the cloud controller manager.
- InstanceExistsByProviderID(ctx context.Context, providerID string) (bool, error)
-}
-
-// Route is a representation of an advanced routing rule.
-type Route struct {
- // Name is the name of the routing rule in the cloud-provider.
- // It will be ignored in a Create (although nameHint may influence it)
- Name string
- // TargetNode is the NodeName of the target instance.
- TargetNode types.NodeName
- // DestinationCIDR is the CIDR format IP range that this routing rule
- // applies to.
- DestinationCIDR string
- // Blackhole is set to true if this is a blackhole route
- // The node controller will delete the route if it is in the managed range.
- Blackhole bool
-}
-
-// Routes is an abstract, pluggable interface for advanced routing rules.
-type Routes interface {
- // ListRoutes lists all managed routes that belong to the specified clusterName
- ListRoutes(ctx context.Context, clusterName string) ([]*Route, error)
- // CreateRoute creates the described managed route
- // route.Name will be ignored, although the cloud-provider may use nameHint
- // to create a more user-meaningful name.
- CreateRoute(ctx context.Context, clusterName string, nameHint string, route *Route) error
- // DeleteRoute deletes the specified managed route
- // Route should be as returned by ListRoutes
- DeleteRoute(ctx context.Context, clusterName string, route *Route) error
-}
-
-var (
- InstanceNotFound = errors.New("instance not found")
- DiskNotFound = errors.New("disk is not found")
- NotImplemented = errors.New("unimplemented")
-)
-
-// Zone represents the location of a particular machine.
-type Zone struct {
- FailureDomain string
- Region string
-}
-
-// Zones is an abstract, pluggable interface for zone enumeration.
-type Zones interface {
- // GetZone returns the Zone containing the current failure zone and locality region that the program is running in
- // In most cases, this method is called from the kubelet querying a local metadata service to acquire its zone.
- // For the case of external cloud providers, use GetZoneByProviderID or GetZoneByNodeName since GetZone
- // can no longer be called from the kubelets.
- GetZone(ctx context.Context) (Zone, error)
-
- // GetZoneByProviderID returns the Zone containing the current zone and locality region of the node specified by providerId
- // This method is particularly used in the context of external cloud providers where node initialization must be down
- // outside the kubelets.
- GetZoneByProviderID(ctx context.Context, providerID string) (Zone, error)
-
- // GetZoneByNodeName returns the Zone containing the current zone and locality region of the node specified by node name
- // This method is particularly used in the context of external cloud providers where node initialization must be down
- // outside the kubelets.
- GetZoneByNodeName(ctx context.Context, nodeName types.NodeName) (Zone, error)
-}
-
-// PVLabeler is an abstract, pluggable interface for fetching labels for volumes
-type PVLabeler interface {
- GetLabelsForVolume(ctx context.Context, pv *v1.PersistentVolume) (map[string]string, error)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/cloudprovider/doc.go b/vendor/k8s.io/kubernetes/pkg/cloudprovider/doc.go
deleted file mode 100644
index 41f8ae860..000000000
--- a/vendor/k8s.io/kubernetes/pkg/cloudprovider/doc.go
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package cloudprovider supplies interfaces and implementations for cloud service providers.
-package cloudprovider // import "k8s.io/kubernetes/pkg/cloudprovider"
diff --git a/vendor/k8s.io/kubernetes/pkg/cloudprovider/plugins.go b/vendor/k8s.io/kubernetes/pkg/cloudprovider/plugins.go
deleted file mode 100644
index 739c09613..000000000
--- a/vendor/k8s.io/kubernetes/pkg/cloudprovider/plugins.go
+++ /dev/null
@@ -1,122 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package cloudprovider
-
-import (
- "fmt"
- "io"
- "os"
- "sync"
-
- "github.com/golang/glog"
-)
-
-// Factory is a function that returns a cloudprovider.Interface.
-// The config parameter provides an io.Reader handler to the factory in
-// order to load specific configurations. If no configuration is provided
-// the parameter is nil.
-type Factory func(config io.Reader) (Interface, error)
-
-// All registered cloud providers.
-var (
- providersMutex sync.Mutex
- providers = make(map[string]Factory)
-)
-
-const externalCloudProvider = "external"
-
-// RegisterCloudProvider registers a cloudprovider.Factory by name. This
-// is expected to happen during app startup.
-func RegisterCloudProvider(name string, cloud Factory) {
- providersMutex.Lock()
- defer providersMutex.Unlock()
- if _, found := providers[name]; found {
- glog.Fatalf("Cloud provider %q was registered twice", name)
- }
- glog.V(1).Infof("Registered cloud provider %q", name)
- providers[name] = cloud
-}
-
-// IsCloudProvider returns true if name corresponds to an already registered
-// cloud provider.
-func IsCloudProvider(name string) bool {
- providersMutex.Lock()
- defer providersMutex.Unlock()
- _, found := providers[name]
- return found
-}
-
-// GetCloudProvider creates an instance of the named cloud provider, or nil if
-// the name is unknown. The error return is only used if the named provider
-// was known but failed to initialize. The config parameter specifies the
-// io.Reader handler of the configuration file for the cloud provider, or nil
-// for no configuration.
-func GetCloudProvider(name string, config io.Reader) (Interface, error) {
- providersMutex.Lock()
- defer providersMutex.Unlock()
- f, found := providers[name]
- if !found {
- return nil, nil
- }
- return f(config)
-}
-
-// Detects if the string is an external cloud provider
-func IsExternal(name string) bool {
- return name == externalCloudProvider
-}
-
-// InitCloudProvider creates an instance of the named cloud provider.
-func InitCloudProvider(name string, configFilePath string) (Interface, error) {
- var cloud Interface
- var err error
-
- if name == "" {
- glog.Info("No cloud provider specified.")
- return nil, nil
- }
-
- if IsExternal(name) {
- glog.Info("External cloud provider specified")
- return nil, nil
- }
-
- if configFilePath != "" {
- var config *os.File
- config, err = os.Open(configFilePath)
- if err != nil {
- glog.Fatalf("Couldn't open cloud provider configuration %s: %#v",
- configFilePath, err)
- }
-
- defer config.Close()
- cloud, err = GetCloudProvider(name, config)
- } else {
- // Pass explicit nil so plugins can actually check for nil. See
- // "Why is my nil error value not equal to nil?" in golang.org/doc/faq.
- cloud, err = GetCloudProvider(name, nil)
- }
-
- if err != nil {
- return nil, fmt.Errorf("could not init cloud provider %q: %v", name, err)
- }
- if cloud == nil {
- return nil, fmt.Errorf("unknown cloud provider %q", name)
- }
-
- return cloud, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/controller/client_builder.go b/vendor/k8s.io/kubernetes/pkg/controller/client_builder.go
deleted file mode 100644
index 041717623..000000000
--- a/vendor/k8s.io/kubernetes/pkg/controller/client_builder.go
+++ /dev/null
@@ -1,292 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package controller
-
-import (
- "fmt"
- "time"
-
- v1authenticationapi "k8s.io/api/authentication/v1"
- "k8s.io/api/core/v1"
- apierrors "k8s.io/apimachinery/pkg/api/errors"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/fields"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/watch"
- apiserverserviceaccount "k8s.io/apiserver/pkg/authentication/serviceaccount"
- clientset "k8s.io/client-go/kubernetes"
- v1authentication "k8s.io/client-go/kubernetes/typed/authentication/v1"
- v1core "k8s.io/client-go/kubernetes/typed/core/v1"
- restclient "k8s.io/client-go/rest"
- "k8s.io/client-go/tools/cache"
- "k8s.io/kubernetes/pkg/api/legacyscheme"
- api "k8s.io/kubernetes/pkg/apis/core"
- "k8s.io/kubernetes/pkg/serviceaccount"
-
- "github.com/golang/glog"
-)
-
-// ControllerClientBuilder allows you to get clients and configs for controllers
-type ControllerClientBuilder interface {
- Config(name string) (*restclient.Config, error)
- ConfigOrDie(name string) *restclient.Config
- Client(name string) (clientset.Interface, error)
- ClientOrDie(name string) clientset.Interface
- ClientGoClient(name string) (clientset.Interface, error)
- ClientGoClientOrDie(name string) clientset.Interface
-}
-
-// SimpleControllerClientBuilder returns a fixed client with different user agents
-type SimpleControllerClientBuilder struct {
- // ClientConfig is a skeleton config to clone and use as the basis for each controller client
- ClientConfig *restclient.Config
-}
-
-func (b SimpleControllerClientBuilder) Config(name string) (*restclient.Config, error) {
- clientConfig := *b.ClientConfig
- return restclient.AddUserAgent(&clientConfig, name), nil
-}
-
-func (b SimpleControllerClientBuilder) ConfigOrDie(name string) *restclient.Config {
- clientConfig, err := b.Config(name)
- if err != nil {
- glog.Fatal(err)
- }
- return clientConfig
-}
-
-func (b SimpleControllerClientBuilder) Client(name string) (clientset.Interface, error) {
- clientConfig, err := b.Config(name)
- if err != nil {
- return nil, err
- }
- return clientset.NewForConfig(clientConfig)
-}
-
-func (b SimpleControllerClientBuilder) ClientOrDie(name string) clientset.Interface {
- client, err := b.Client(name)
- if err != nil {
- glog.Fatal(err)
- }
- return client
-}
-
-func (b SimpleControllerClientBuilder) ClientGoClient(name string) (clientset.Interface, error) {
- clientConfig, err := b.Config(name)
- if err != nil {
- return nil, err
- }
- return clientset.NewForConfig(clientConfig)
-}
-
-func (b SimpleControllerClientBuilder) ClientGoClientOrDie(name string) clientset.Interface {
- client, err := b.ClientGoClient(name)
- if err != nil {
- glog.Fatal(err)
- }
- return client
-}
-
-// SAControllerClientBuilder is a ControllerClientBuilder that returns clients identifying as
-// service accounts
-type SAControllerClientBuilder struct {
- // ClientConfig is a skeleton config to clone and use as the basis for each controller client
- ClientConfig *restclient.Config
-
- // CoreClient is used to provision service accounts if needed and watch for their associated tokens
- // to construct a controller client
- CoreClient v1core.CoreV1Interface
-
- // AuthenticationClient is used to check API tokens to make sure they are valid before
- // building a controller client from them
- AuthenticationClient v1authentication.AuthenticationV1Interface
-
- // Namespace is the namespace used to host the service accounts that will back the
- // controllers. It must be highly privileged namespace which normal users cannot inspect.
- Namespace string
-}
-
-// config returns a complete clientConfig for constructing clients. This is separate in anticipation of composition
-// which means that not all clientsets are known here
-func (b SAControllerClientBuilder) Config(name string) (*restclient.Config, error) {
- sa, err := b.getOrCreateServiceAccount(name)
- if err != nil {
- return nil, err
- }
-
- var clientConfig *restclient.Config
-
- lw := &cache.ListWatch{
- ListFunc: func(options metav1.ListOptions) (runtime.Object, error) {
- options.FieldSelector = fields.SelectorFromSet(map[string]string{api.SecretTypeField: string(v1.SecretTypeServiceAccountToken)}).String()
- return b.CoreClient.Secrets(b.Namespace).List(options)
- },
- WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) {
- options.FieldSelector = fields.SelectorFromSet(map[string]string{api.SecretTypeField: string(v1.SecretTypeServiceAccountToken)}).String()
- return b.CoreClient.Secrets(b.Namespace).Watch(options)
- },
- }
- _, err = cache.ListWatchUntil(30*time.Second, lw,
- func(event watch.Event) (bool, error) {
- switch event.Type {
- case watch.Deleted:
- return false, nil
- case watch.Error:
- return false, fmt.Errorf("error watching")
-
- case watch.Added, watch.Modified:
- secret, ok := event.Object.(*v1.Secret)
- if !ok {
- return false, fmt.Errorf("unexpected object type: %T", event.Object)
- }
- if !serviceaccount.IsServiceAccountToken(secret, sa) {
- return false, nil
- }
- if len(secret.Data[v1.ServiceAccountTokenKey]) == 0 {
- return false, nil
- }
- validConfig, valid, err := b.getAuthenticatedConfig(sa, string(secret.Data[v1.ServiceAccountTokenKey]))
- if err != nil {
- glog.Warningf("error validating API token for %s/%s in secret %s: %v", sa.Name, sa.Namespace, secret.Name, err)
- // continue watching for good tokens
- return false, nil
- }
- if !valid {
- glog.Warningf("secret %s contained an invalid API token for %s/%s", secret.Name, sa.Name, sa.Namespace)
- // try to delete the secret containing the invalid token
- if err := b.CoreClient.Secrets(secret.Namespace).Delete(secret.Name, &metav1.DeleteOptions{}); err != nil && !apierrors.IsNotFound(err) {
- glog.Warningf("error deleting secret %s containing invalid API token for %s/%s: %v", secret.Name, sa.Name, sa.Namespace, err)
- }
- // continue watching for good tokens
- return false, nil
- }
- clientConfig = validConfig
- return true, nil
-
- default:
- return false, fmt.Errorf("unexpected event type: %v", event.Type)
- }
- })
- if err != nil {
- return nil, fmt.Errorf("unable to get token for service account: %v", err)
- }
-
- return clientConfig, nil
-}
-
-func (b SAControllerClientBuilder) getOrCreateServiceAccount(name string) (*v1.ServiceAccount, error) {
- sa, err := b.CoreClient.ServiceAccounts(b.Namespace).Get(name, metav1.GetOptions{})
- if err == nil {
- return sa, nil
- }
- if !apierrors.IsNotFound(err) {
- return nil, err
- }
-
- // Create the namespace if we can't verify it exists.
- // Tolerate errors, since we don't know whether this component has namespace creation permissions.
- if _, err := b.CoreClient.Namespaces().Get(b.Namespace, metav1.GetOptions{}); err != nil {
- b.CoreClient.Namespaces().Create(&v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: b.Namespace}})
- }
-
- // Create the service account
- sa, err = b.CoreClient.ServiceAccounts(b.Namespace).Create(&v1.ServiceAccount{ObjectMeta: metav1.ObjectMeta{Namespace: b.Namespace, Name: name}})
- if apierrors.IsAlreadyExists(err) {
- // If we're racing to init and someone else already created it, re-fetch
- return b.CoreClient.ServiceAccounts(b.Namespace).Get(name, metav1.GetOptions{})
- }
- return sa, err
-}
-
-func (b SAControllerClientBuilder) getAuthenticatedConfig(sa *v1.ServiceAccount, token string) (*restclient.Config, bool, error) {
- username := apiserverserviceaccount.MakeUsername(sa.Namespace, sa.Name)
-
- clientConfig := restclient.AnonymousClientConfig(b.ClientConfig)
- clientConfig.BearerToken = token
- restclient.AddUserAgent(clientConfig, username)
-
- // Try token review first
- tokenReview := &v1authenticationapi.TokenReview{Spec: v1authenticationapi.TokenReviewSpec{Token: token}}
- if tokenResult, err := b.AuthenticationClient.TokenReviews().Create(tokenReview); err == nil {
- if !tokenResult.Status.Authenticated {
- glog.Warningf("Token for %s/%s did not authenticate correctly", sa.Name, sa.Namespace)
- return nil, false, nil
- }
- if tokenResult.Status.User.Username != username {
- glog.Warningf("Token for %s/%s authenticated as unexpected username: %s", sa.Name, sa.Namespace, tokenResult.Status.User.Username)
- return nil, false, nil
- }
- glog.V(4).Infof("Verified credential for %s/%s", sa.Name, sa.Namespace)
- return clientConfig, true, nil
- }
-
- // If we couldn't run the token review, the API might be disabled or we might not have permission.
- // Try to make a request to /apis with the token. If we get a 401 we should consider the token invalid.
- clientConfigCopy := *clientConfig
- clientConfigCopy.NegotiatedSerializer = legacyscheme.Codecs
- client, err := restclient.UnversionedRESTClientFor(&clientConfigCopy)
- if err != nil {
- return nil, false, err
- }
- err = client.Get().AbsPath("/apis").Do().Error()
- if apierrors.IsUnauthorized(err) {
- glog.Warningf("Token for %s/%s did not authenticate correctly: %v", sa.Name, sa.Namespace, err)
- return nil, false, nil
- }
-
- return clientConfig, true, nil
-}
-
-func (b SAControllerClientBuilder) ConfigOrDie(name string) *restclient.Config {
- clientConfig, err := b.Config(name)
- if err != nil {
- glog.Fatal(err)
- }
- return clientConfig
-}
-
-func (b SAControllerClientBuilder) Client(name string) (clientset.Interface, error) {
- clientConfig, err := b.Config(name)
- if err != nil {
- return nil, err
- }
- return clientset.NewForConfig(clientConfig)
-}
-
-func (b SAControllerClientBuilder) ClientOrDie(name string) clientset.Interface {
- client, err := b.Client(name)
- if err != nil {
- glog.Fatal(err)
- }
- return client
-}
-
-func (b SAControllerClientBuilder) ClientGoClient(name string) (clientset.Interface, error) {
- clientConfig, err := b.Config(name)
- if err != nil {
- return nil, err
- }
- return clientset.NewForConfig(clientConfig)
-}
-
-func (b SAControllerClientBuilder) ClientGoClientOrDie(name string) clientset.Interface {
- client, err := b.ClientGoClient(name)
- if err != nil {
- glog.Fatal(err)
- }
- return client
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/controller/controller_ref_manager.go b/vendor/k8s.io/kubernetes/pkg/controller/controller_ref_manager.go
deleted file mode 100644
index 21d7aa302..000000000
--- a/vendor/k8s.io/kubernetes/pkg/controller/controller_ref_manager.go
+++ /dev/null
@@ -1,502 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package controller
-
-import (
- "fmt"
- "sync"
-
- "github.com/golang/glog"
- apps "k8s.io/api/apps/v1"
- "k8s.io/api/core/v1"
- extensions "k8s.io/api/extensions/v1beta1"
- "k8s.io/apimachinery/pkg/api/errors"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/runtime/schema"
- utilerrors "k8s.io/apimachinery/pkg/util/errors"
-)
-
-type BaseControllerRefManager struct {
- Controller metav1.Object
- Selector labels.Selector
-
- canAdoptErr error
- canAdoptOnce sync.Once
- CanAdoptFunc func() error
-}
-
-func (m *BaseControllerRefManager) CanAdopt() error {
- m.canAdoptOnce.Do(func() {
- if m.CanAdoptFunc != nil {
- m.canAdoptErr = m.CanAdoptFunc()
- }
- })
- return m.canAdoptErr
-}
-
-// ClaimObject tries to take ownership of an object for this controller.
-//
-// It will reconcile the following:
-// * Adopt orphans if the match function returns true.
-// * Release owned objects if the match function returns false.
-//
-// A non-nil error is returned if some form of reconciliation was attempted and
-// failed. Usually, controllers should try again later in case reconciliation
-// is still needed.
-//
-// If the error is nil, either the reconciliation succeeded, or no
-// reconciliation was necessary. The returned boolean indicates whether you now
-// own the object.
-//
-// No reconciliation will be attempted if the controller is being deleted.
-func (m *BaseControllerRefManager) ClaimObject(obj metav1.Object, match func(metav1.Object) bool, adopt, release func(metav1.Object) error) (bool, error) {
- controllerRef := metav1.GetControllerOf(obj)
- if controllerRef != nil {
- if controllerRef.UID != m.Controller.GetUID() {
- // Owned by someone else. Ignore.
- return false, nil
- }
- if match(obj) {
- // We already own it and the selector matches.
- // Return true (successfully claimed) before checking deletion timestamp.
- // We're still allowed to claim things we already own while being deleted
- // because doing so requires taking no actions.
- return true, nil
- }
- // Owned by us but selector doesn't match.
- // Try to release, unless we're being deleted.
- if m.Controller.GetDeletionTimestamp() != nil {
- return false, nil
- }
- if err := release(obj); err != nil {
- // If the pod no longer exists, ignore the error.
- if errors.IsNotFound(err) {
- return false, nil
- }
- // Either someone else released it, or there was a transient error.
- // The controller should requeue and try again if it's still stale.
- return false, err
- }
- // Successfully released.
- return false, nil
- }
-
- // It's an orphan.
- if m.Controller.GetDeletionTimestamp() != nil || !match(obj) {
- // Ignore if we're being deleted or selector doesn't match.
- return false, nil
- }
- if obj.GetDeletionTimestamp() != nil {
- // Ignore if the object is being deleted
- return false, nil
- }
- // Selector matches. Try to adopt.
- if err := adopt(obj); err != nil {
- // If the pod no longer exists, ignore the error.
- if errors.IsNotFound(err) {
- return false, nil
- }
- // Either someone else claimed it first, or there was a transient error.
- // The controller should requeue and try again if it's still orphaned.
- return false, err
- }
- // Successfully adopted.
- return true, nil
-}
-
-type PodControllerRefManager struct {
- BaseControllerRefManager
- controllerKind schema.GroupVersionKind
- podControl PodControlInterface
-}
-
-// NewPodControllerRefManager returns a PodControllerRefManager that exposes
-// methods to manage the controllerRef of pods.
-//
-// The CanAdopt() function can be used to perform a potentially expensive check
-// (such as a live GET from the API server) prior to the first adoption.
-// It will only be called (at most once) if an adoption is actually attempted.
-// If CanAdopt() returns a non-nil error, all adoptions will fail.
-//
-// NOTE: Once CanAdopt() is called, it will not be called again by the same
-// PodControllerRefManager instance. Create a new instance if it makes
-// sense to check CanAdopt() again (e.g. in a different sync pass).
-func NewPodControllerRefManager(
- podControl PodControlInterface,
- controller metav1.Object,
- selector labels.Selector,
- controllerKind schema.GroupVersionKind,
- canAdopt func() error,
-) *PodControllerRefManager {
- return &PodControllerRefManager{
- BaseControllerRefManager: BaseControllerRefManager{
- Controller: controller,
- Selector: selector,
- CanAdoptFunc: canAdopt,
- },
- controllerKind: controllerKind,
- podControl: podControl,
- }
-}
-
-// ClaimPods tries to take ownership of a list of Pods.
-//
-// It will reconcile the following:
-// * Adopt orphans if the selector matches.
-// * Release owned objects if the selector no longer matches.
-//
-// Optional: If one or more filters are specified, a Pod will only be claimed if
-// all filters return true.
-//
-// A non-nil error is returned if some form of reconciliation was attempted and
-// failed. Usually, controllers should try again later in case reconciliation
-// is still needed.
-//
-// If the error is nil, either the reconciliation succeeded, or no
-// reconciliation was necessary. The list of Pods that you now own is returned.
-func (m *PodControllerRefManager) ClaimPods(pods []*v1.Pod, filters ...func(*v1.Pod) bool) ([]*v1.Pod, error) {
- var claimed []*v1.Pod
- var errlist []error
-
- match := func(obj metav1.Object) bool {
- pod := obj.(*v1.Pod)
- // Check selector first so filters only run on potentially matching Pods.
- if !m.Selector.Matches(labels.Set(pod.Labels)) {
- return false
- }
- for _, filter := range filters {
- if !filter(pod) {
- return false
- }
- }
- return true
- }
- adopt := func(obj metav1.Object) error {
- return m.AdoptPod(obj.(*v1.Pod))
- }
- release := func(obj metav1.Object) error {
- return m.ReleasePod(obj.(*v1.Pod))
- }
-
- for _, pod := range pods {
- ok, err := m.ClaimObject(pod, match, adopt, release)
- if err != nil {
- errlist = append(errlist, err)
- continue
- }
- if ok {
- claimed = append(claimed, pod)
- }
- }
- return claimed, utilerrors.NewAggregate(errlist)
-}
-
-// AdoptPod sends a patch to take control of the pod. It returns the error if
-// the patching fails.
-func (m *PodControllerRefManager) AdoptPod(pod *v1.Pod) error {
- if err := m.CanAdopt(); err != nil {
- return fmt.Errorf("can't adopt Pod %v/%v (%v): %v", pod.Namespace, pod.Name, pod.UID, err)
- }
- // Note that ValidateOwnerReferences() will reject this patch if another
- // OwnerReference exists with controller=true.
- addControllerPatch := fmt.Sprintf(
- `{"metadata":{"ownerReferences":[{"apiVersion":"%s","kind":"%s","name":"%s","uid":"%s","controller":true,"blockOwnerDeletion":true}],"uid":"%s"}}`,
- m.controllerKind.GroupVersion(), m.controllerKind.Kind,
- m.Controller.GetName(), m.Controller.GetUID(), pod.UID)
- return m.podControl.PatchPod(pod.Namespace, pod.Name, []byte(addControllerPatch))
-}
-
-// ReleasePod sends a patch to free the pod from the control of the controller.
-// It returns the error if the patching fails. 404 and 422 errors are ignored.
-func (m *PodControllerRefManager) ReleasePod(pod *v1.Pod) error {
- glog.V(2).Infof("patching pod %s_%s to remove its controllerRef to %s/%s:%s",
- pod.Namespace, pod.Name, m.controllerKind.GroupVersion(), m.controllerKind.Kind, m.Controller.GetName())
- deleteOwnerRefPatch := fmt.Sprintf(`{"metadata":{"ownerReferences":[{"$patch":"delete","uid":"%s"}],"uid":"%s"}}`, m.Controller.GetUID(), pod.UID)
- err := m.podControl.PatchPod(pod.Namespace, pod.Name, []byte(deleteOwnerRefPatch))
- if err != nil {
- if errors.IsNotFound(err) {
- // If the pod no longer exists, ignore it.
- return nil
- }
- if errors.IsInvalid(err) {
- // Invalid error will be returned in two cases: 1. the pod
- // has no owner reference, 2. the uid of the pod doesn't
- // match, which means the pod is deleted and then recreated.
- // In both cases, the error can be ignored.
-
- // TODO: If the pod has owner references, but none of them
- // has the owner.UID, server will silently ignore the patch.
- // Investigate why.
- return nil
- }
- }
- return err
-}
-
-// ReplicaSetControllerRefManager is used to manage controllerRef of ReplicaSets.
-// Three methods are defined on this object 1: Classify 2: AdoptReplicaSet and
-// 3: ReleaseReplicaSet which are used to classify the ReplicaSets into appropriate
-// categories and accordingly adopt or release them. See comments on these functions
-// for more details.
-type ReplicaSetControllerRefManager struct {
- BaseControllerRefManager
- controllerKind schema.GroupVersionKind
- rsControl RSControlInterface
-}
-
-// NewReplicaSetControllerRefManager returns a ReplicaSetControllerRefManager that exposes
-// methods to manage the controllerRef of ReplicaSets.
-//
-// The CanAdopt() function can be used to perform a potentially expensive check
-// (such as a live GET from the API server) prior to the first adoption.
-// It will only be called (at most once) if an adoption is actually attempted.
-// If CanAdopt() returns a non-nil error, all adoptions will fail.
-//
-// NOTE: Once CanAdopt() is called, it will not be called again by the same
-// ReplicaSetControllerRefManager instance. Create a new instance if it
-// makes sense to check CanAdopt() again (e.g. in a different sync pass).
-func NewReplicaSetControllerRefManager(
- rsControl RSControlInterface,
- controller metav1.Object,
- selector labels.Selector,
- controllerKind schema.GroupVersionKind,
- canAdopt func() error,
-) *ReplicaSetControllerRefManager {
- return &ReplicaSetControllerRefManager{
- BaseControllerRefManager: BaseControllerRefManager{
- Controller: controller,
- Selector: selector,
- CanAdoptFunc: canAdopt,
- },
- controllerKind: controllerKind,
- rsControl: rsControl,
- }
-}
-
-// ClaimReplicaSets tries to take ownership of a list of ReplicaSets.
-//
-// It will reconcile the following:
-// * Adopt orphans if the selector matches.
-// * Release owned objects if the selector no longer matches.
-//
-// A non-nil error is returned if some form of reconciliation was attempted and
-// failed. Usually, controllers should try again later in case reconciliation
-// is still needed.
-//
-// If the error is nil, either the reconciliation succeeded, or no
-// reconciliation was necessary. The list of ReplicaSets that you now own is
-// returned.
-func (m *ReplicaSetControllerRefManager) ClaimReplicaSets(sets []*extensions.ReplicaSet) ([]*extensions.ReplicaSet, error) {
- var claimed []*extensions.ReplicaSet
- var errlist []error
-
- match := func(obj metav1.Object) bool {
- return m.Selector.Matches(labels.Set(obj.GetLabels()))
- }
- adopt := func(obj metav1.Object) error {
- return m.AdoptReplicaSet(obj.(*extensions.ReplicaSet))
- }
- release := func(obj metav1.Object) error {
- return m.ReleaseReplicaSet(obj.(*extensions.ReplicaSet))
- }
-
- for _, rs := range sets {
- ok, err := m.ClaimObject(rs, match, adopt, release)
- if err != nil {
- errlist = append(errlist, err)
- continue
- }
- if ok {
- claimed = append(claimed, rs)
- }
- }
- return claimed, utilerrors.NewAggregate(errlist)
-}
-
-// AdoptReplicaSet sends a patch to take control of the ReplicaSet. It returns
-// the error if the patching fails.
-func (m *ReplicaSetControllerRefManager) AdoptReplicaSet(rs *extensions.ReplicaSet) error {
- if err := m.CanAdopt(); err != nil {
- return fmt.Errorf("can't adopt ReplicaSet %v/%v (%v): %v", rs.Namespace, rs.Name, rs.UID, err)
- }
- // Note that ValidateOwnerReferences() will reject this patch if another
- // OwnerReference exists with controller=true.
- addControllerPatch := fmt.Sprintf(
- `{"metadata":{"ownerReferences":[{"apiVersion":"%s","kind":"%s","name":"%s","uid":"%s","controller":true,"blockOwnerDeletion":true}],"uid":"%s"}}`,
- m.controllerKind.GroupVersion(), m.controllerKind.Kind,
- m.Controller.GetName(), m.Controller.GetUID(), rs.UID)
- return m.rsControl.PatchReplicaSet(rs.Namespace, rs.Name, []byte(addControllerPatch))
-}
-
-// ReleaseReplicaSet sends a patch to free the ReplicaSet from the control of the Deployment controller.
-// It returns the error if the patching fails. 404 and 422 errors are ignored.
-func (m *ReplicaSetControllerRefManager) ReleaseReplicaSet(replicaSet *extensions.ReplicaSet) error {
- glog.V(2).Infof("patching ReplicaSet %s_%s to remove its controllerRef to %s/%s:%s",
- replicaSet.Namespace, replicaSet.Name, m.controllerKind.GroupVersion(), m.controllerKind.Kind, m.Controller.GetName())
- deleteOwnerRefPatch := fmt.Sprintf(`{"metadata":{"ownerReferences":[{"$patch":"delete","uid":"%s"}],"uid":"%s"}}`, m.Controller.GetUID(), replicaSet.UID)
- err := m.rsControl.PatchReplicaSet(replicaSet.Namespace, replicaSet.Name, []byte(deleteOwnerRefPatch))
- if err != nil {
- if errors.IsNotFound(err) {
- // If the ReplicaSet no longer exists, ignore it.
- return nil
- }
- if errors.IsInvalid(err) {
- // Invalid error will be returned in two cases: 1. the ReplicaSet
- // has no owner reference, 2. the uid of the ReplicaSet doesn't
- // match, which means the ReplicaSet is deleted and then recreated.
- // In both cases, the error can be ignored.
- return nil
- }
- }
- return err
-}
-
-// RecheckDeletionTimestamp returns a CanAdopt() function to recheck deletion.
-//
-// The CanAdopt() function calls getObject() to fetch the latest value,
-// and denies adoption attempts if that object has a non-nil DeletionTimestamp.
-func RecheckDeletionTimestamp(getObject func() (metav1.Object, error)) func() error {
- return func() error {
- obj, err := getObject()
- if err != nil {
- return fmt.Errorf("can't recheck DeletionTimestamp: %v", err)
- }
- if obj.GetDeletionTimestamp() != nil {
- return fmt.Errorf("%v/%v has just been deleted at %v", obj.GetNamespace(), obj.GetName(), obj.GetDeletionTimestamp())
- }
- return nil
- }
-}
-
-// ControllerRevisionControllerRefManager is used to manage controllerRef of ControllerRevisions.
-// Three methods are defined on this object 1: Classify 2: AdoptControllerRevision and
-// 3: ReleaseControllerRevision which are used to classify the ControllerRevisions into appropriate
-// categories and accordingly adopt or release them. See comments on these functions
-// for more details.
-type ControllerRevisionControllerRefManager struct {
- BaseControllerRefManager
- controllerKind schema.GroupVersionKind
- crControl ControllerRevisionControlInterface
-}
-
-// NewControllerRevisionControllerRefManager returns a ControllerRevisionControllerRefManager that exposes
-// methods to manage the controllerRef of ControllerRevisions.
-//
-// The canAdopt() function can be used to perform a potentially expensive check
-// (such as a live GET from the API server) prior to the first adoption.
-// It will only be called (at most once) if an adoption is actually attempted.
-// If canAdopt() returns a non-nil error, all adoptions will fail.
-//
-// NOTE: Once canAdopt() is called, it will not be called again by the same
-// ControllerRevisionControllerRefManager instance. Create a new instance if it
-// makes sense to check canAdopt() again (e.g. in a different sync pass).
-func NewControllerRevisionControllerRefManager(
- crControl ControllerRevisionControlInterface,
- controller metav1.Object,
- selector labels.Selector,
- controllerKind schema.GroupVersionKind,
- canAdopt func() error,
-) *ControllerRevisionControllerRefManager {
- return &ControllerRevisionControllerRefManager{
- BaseControllerRefManager: BaseControllerRefManager{
- Controller: controller,
- Selector: selector,
- CanAdoptFunc: canAdopt,
- },
- controllerKind: controllerKind,
- crControl: crControl,
- }
-}
-
-// ClaimControllerRevisions tries to take ownership of a list of ControllerRevisions.
-//
-// It will reconcile the following:
-// * Adopt orphans if the selector matches.
-// * Release owned objects if the selector no longer matches.
-//
-// A non-nil error is returned if some form of reconciliation was attempted and
-// failed. Usually, controllers should try again later in case reconciliation
-// is still needed.
-//
-// If the error is nil, either the reconciliation succeeded, or no
-// reconciliation was necessary. The list of ControllerRevisions that you now own is
-// returned.
-func (m *ControllerRevisionControllerRefManager) ClaimControllerRevisions(histories []*apps.ControllerRevision) ([]*apps.ControllerRevision, error) {
- var claimed []*apps.ControllerRevision
- var errlist []error
-
- match := func(obj metav1.Object) bool {
- return m.Selector.Matches(labels.Set(obj.GetLabels()))
- }
- adopt := func(obj metav1.Object) error {
- return m.AdoptControllerRevision(obj.(*apps.ControllerRevision))
- }
- release := func(obj metav1.Object) error {
- return m.ReleaseControllerRevision(obj.(*apps.ControllerRevision))
- }
-
- for _, h := range histories {
- ok, err := m.ClaimObject(h, match, adopt, release)
- if err != nil {
- errlist = append(errlist, err)
- continue
- }
- if ok {
- claimed = append(claimed, h)
- }
- }
- return claimed, utilerrors.NewAggregate(errlist)
-}
-
-// AdoptControllerRevision sends a patch to take control of the ControllerRevision. It returns the error if
-// the patching fails.
-func (m *ControllerRevisionControllerRefManager) AdoptControllerRevision(history *apps.ControllerRevision) error {
- if err := m.CanAdopt(); err != nil {
- return fmt.Errorf("can't adopt ControllerRevision %v/%v (%v): %v", history.Namespace, history.Name, history.UID, err)
- }
- // Note that ValidateOwnerReferences() will reject this patch if another
- // OwnerReference exists with controller=true.
- addControllerPatch := fmt.Sprintf(
- `{"metadata":{"ownerReferences":[{"apiVersion":"%s","kind":"%s","name":"%s","uid":"%s","controller":true,"blockOwnerDeletion":true}],"uid":"%s"}}`,
- m.controllerKind.GroupVersion(), m.controllerKind.Kind,
- m.Controller.GetName(), m.Controller.GetUID(), history.UID)
- return m.crControl.PatchControllerRevision(history.Namespace, history.Name, []byte(addControllerPatch))
-}
-
-// ReleaseControllerRevision sends a patch to free the ControllerRevision from the control of its controller.
-// It returns the error if the patching fails. 404 and 422 errors are ignored.
-func (m *ControllerRevisionControllerRefManager) ReleaseControllerRevision(history *apps.ControllerRevision) error {
- glog.V(2).Infof("patching ControllerRevision %s_%s to remove its controllerRef to %s/%s:%s",
- history.Namespace, history.Name, m.controllerKind.GroupVersion(), m.controllerKind.Kind, m.Controller.GetName())
- deleteOwnerRefPatch := fmt.Sprintf(`{"metadata":{"ownerReferences":[{"$patch":"delete","uid":"%s"}],"uid":"%s"}}`, m.Controller.GetUID(), history.UID)
- err := m.crControl.PatchControllerRevision(history.Namespace, history.Name, []byte(deleteOwnerRefPatch))
- if err != nil {
- if errors.IsNotFound(err) {
- // If the ControllerRevision no longer exists, ignore it.
- return nil
- }
- if errors.IsInvalid(err) {
- // Invalid error will be returned in two cases: 1. the ControllerRevision
- // has no owner reference, 2. the uid of the ControllerRevision doesn't
- // match, which means the ControllerRevision is deleted and then recreated.
- // In both cases, the error can be ignored.
- return nil
- }
- }
- return err
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/controller/controller_utils.go b/vendor/k8s.io/kubernetes/pkg/controller/controller_utils.go
deleted file mode 100644
index d7d755ebf..000000000
--- a/vendor/k8s.io/kubernetes/pkg/controller/controller_utils.go
+++ /dev/null
@@ -1,1043 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package controller
-
-import (
- "encoding/binary"
- "encoding/json"
- "fmt"
- "hash/fnv"
- "sync"
- "sync/atomic"
- "time"
-
- "k8s.io/api/core/v1"
- extensions "k8s.io/api/extensions/v1beta1"
- "k8s.io/apimachinery/pkg/api/meta"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/types"
- "k8s.io/apimachinery/pkg/util/clock"
- utilruntime "k8s.io/apimachinery/pkg/util/runtime"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/strategicpatch"
- "k8s.io/apimachinery/pkg/util/wait"
- clientset "k8s.io/client-go/kubernetes"
- "k8s.io/client-go/tools/cache"
- "k8s.io/client-go/tools/record"
- "k8s.io/client-go/util/integer"
- clientretry "k8s.io/client-go/util/retry"
- podutil "k8s.io/kubernetes/pkg/api/v1/pod"
- _ "k8s.io/kubernetes/pkg/apis/core/install"
- "k8s.io/kubernetes/pkg/apis/core/validation"
- hashutil "k8s.io/kubernetes/pkg/util/hash"
- taintutils "k8s.io/kubernetes/pkg/util/taints"
-
- "github.com/golang/glog"
-)
-
-const (
- // If a watch drops a delete event for a pod, it'll take this long
- // before a dormant controller waiting for those packets is woken up anyway. It is
- // specifically targeted at the case where some problem prevents an update
- // of expectations, without it the controller could stay asleep forever. This should
- // be set based on the expected latency of watch events.
- //
- // Currently a controller can service (create *and* observe the watch events for said
- // creation) about 10 pods a second, so it takes about 1 min to service
- // 500 pods. Just creation is limited to 20qps, and watching happens with ~10-30s
- // latency/pod at the scale of 3000 pods over 100 nodes.
- ExpectationsTimeout = 5 * time.Minute
- // When batching pod creates, SlowStartInitialBatchSize is the size of the
- // initial batch. The size of each successive batch is twice the size of
- // the previous batch. For example, for a value of 1, batch sizes would be
- // 1, 2, 4, 8, ... and for a value of 10, batch sizes would be
- // 10, 20, 40, 80, ... Setting the value higher means that quota denials
- // will result in more doomed API calls and associated event spam. Setting
- // the value lower will result in more API call round trip periods for
- // large batches.
- //
- // Given a number of pods to start "N":
- // The number of doomed calls per sync once quota is exceeded is given by:
- // min(N,SlowStartInitialBatchSize)
- // The number of batches is given by:
- // 1+floor(log_2(ceil(N/SlowStartInitialBatchSize)))
- SlowStartInitialBatchSize = 1
-)
-
-var UpdateTaintBackoff = wait.Backoff{
- Steps: 5,
- Duration: 100 * time.Millisecond,
- Jitter: 1.0,
-}
-
-var (
- KeyFunc = cache.DeletionHandlingMetaNamespaceKeyFunc
-)
-
-type ResyncPeriodFunc func() time.Duration
-
-// Returns 0 for resyncPeriod in case resyncing is not needed.
-func NoResyncPeriodFunc() time.Duration {
- return 0
-}
-
-// StaticResyncPeriodFunc returns the resync period specified
-func StaticResyncPeriodFunc(resyncPeriod time.Duration) ResyncPeriodFunc {
- return func() time.Duration {
- return resyncPeriod
- }
-}
-
-// Expectations are a way for controllers to tell the controller manager what they expect. eg:
-// ControllerExpectations: {
-// controller1: expects 2 adds in 2 minutes
-// controller2: expects 2 dels in 2 minutes
-// controller3: expects -1 adds in 2 minutes => controller3's expectations have already been met
-// }
-//
-// Implementation:
-// ControlleeExpectation = pair of atomic counters to track controllee's creation/deletion
-// ControllerExpectationsStore = TTLStore + a ControlleeExpectation per controller
-//
-// * Once set expectations can only be lowered
-// * A controller isn't synced till its expectations are either fulfilled, or expire
-// * Controllers that don't set expectations will get woken up for every matching controllee
-
-// ExpKeyFunc to parse out the key from a ControlleeExpectation
-var ExpKeyFunc = func(obj interface{}) (string, error) {
- if e, ok := obj.(*ControlleeExpectations); ok {
- return e.key, nil
- }
- return "", fmt.Errorf("Could not find key for obj %#v", obj)
-}
-
-// ControllerExpectationsInterface is an interface that allows users to set and wait on expectations.
-// Only abstracted out for testing.
-// Warning: if using KeyFunc it is not safe to use a single ControllerExpectationsInterface with different
-// types of controllers, because the keys might conflict across types.
-type ControllerExpectationsInterface interface {
- GetExpectations(controllerKey string) (*ControlleeExpectations, bool, error)
- SatisfiedExpectations(controllerKey string) bool
- DeleteExpectations(controllerKey string)
- SetExpectations(controllerKey string, add, del int) error
- ExpectCreations(controllerKey string, adds int) error
- ExpectDeletions(controllerKey string, dels int) error
- CreationObserved(controllerKey string)
- DeletionObserved(controllerKey string)
- RaiseExpectations(controllerKey string, add, del int)
- LowerExpectations(controllerKey string, add, del int)
-}
-
-// ControllerExpectations is a cache mapping controllers to what they expect to see before being woken up for a sync.
-type ControllerExpectations struct {
- cache.Store
-}
-
-// GetExpectations returns the ControlleeExpectations of the given controller.
-func (r *ControllerExpectations) GetExpectations(controllerKey string) (*ControlleeExpectations, bool, error) {
- if exp, exists, err := r.GetByKey(controllerKey); err == nil && exists {
- return exp.(*ControlleeExpectations), true, nil
- } else {
- return nil, false, err
- }
-}
-
-// DeleteExpectations deletes the expectations of the given controller from the TTLStore.
-func (r *ControllerExpectations) DeleteExpectations(controllerKey string) {
- if exp, exists, err := r.GetByKey(controllerKey); err == nil && exists {
- if err := r.Delete(exp); err != nil {
- glog.V(2).Infof("Error deleting expectations for controller %v: %v", controllerKey, err)
- }
- }
-}
-
-// SatisfiedExpectations returns true if the required adds/dels for the given controller have been observed.
-// Add/del counts are established by the controller at sync time, and updated as controllees are observed by the controller
-// manager.
-func (r *ControllerExpectations) SatisfiedExpectations(controllerKey string) bool {
- if exp, exists, err := r.GetExpectations(controllerKey); exists {
- if exp.Fulfilled() {
- glog.V(4).Infof("Controller expectations fulfilled %#v", exp)
- return true
- } else if exp.isExpired() {
- glog.V(4).Infof("Controller expectations expired %#v", exp)
- return true
- } else {
- glog.V(4).Infof("Controller still waiting on expectations %#v", exp)
- return false
- }
- } else if err != nil {
- glog.V(2).Infof("Error encountered while checking expectations %#v, forcing sync", err)
- } else {
- // When a new controller is created, it doesn't have expectations.
- // When it doesn't see expected watch events for > TTL, the expectations expire.
- // - In this case it wakes up, creates/deletes controllees, and sets expectations again.
- // When it has satisfied expectations and no controllees need to be created/destroyed > TTL, the expectations expire.
- // - In this case it continues without setting expectations till it needs to create/delete controllees.
- glog.V(4).Infof("Controller %v either never recorded expectations, or the ttl expired.", controllerKey)
- }
- // Trigger a sync if we either encountered and error (which shouldn't happen since we're
- // getting from local store) or this controller hasn't established expectations.
- return true
-}
-
-// TODO: Extend ExpirationCache to support explicit expiration.
-// TODO: Make this possible to disable in tests.
-// TODO: Support injection of clock.
-func (exp *ControlleeExpectations) isExpired() bool {
- return clock.RealClock{}.Since(exp.timestamp) > ExpectationsTimeout
-}
-
-// SetExpectations registers new expectations for the given controller. Forgets existing expectations.
-func (r *ControllerExpectations) SetExpectations(controllerKey string, add, del int) error {
- exp := &ControlleeExpectations{add: int64(add), del: int64(del), key: controllerKey, timestamp: clock.RealClock{}.Now()}
- glog.V(4).Infof("Setting expectations %#v", exp)
- return r.Add(exp)
-}
-
-func (r *ControllerExpectations) ExpectCreations(controllerKey string, adds int) error {
- return r.SetExpectations(controllerKey, adds, 0)
-}
-
-func (r *ControllerExpectations) ExpectDeletions(controllerKey string, dels int) error {
- return r.SetExpectations(controllerKey, 0, dels)
-}
-
-// Decrements the expectation counts of the given controller.
-func (r *ControllerExpectations) LowerExpectations(controllerKey string, add, del int) {
- if exp, exists, err := r.GetExpectations(controllerKey); err == nil && exists {
- exp.Add(int64(-add), int64(-del))
- // The expectations might've been modified since the update on the previous line.
- glog.V(4).Infof("Lowered expectations %#v", exp)
- }
-}
-
-// Increments the expectation counts of the given controller.
-func (r *ControllerExpectations) RaiseExpectations(controllerKey string, add, del int) {
- if exp, exists, err := r.GetExpectations(controllerKey); err == nil && exists {
- exp.Add(int64(add), int64(del))
- // The expectations might've been modified since the update on the previous line.
- glog.V(4).Infof("Raised expectations %#v", exp)
- }
-}
-
-// CreationObserved atomically decrements the `add` expectation count of the given controller.
-func (r *ControllerExpectations) CreationObserved(controllerKey string) {
- r.LowerExpectations(controllerKey, 1, 0)
-}
-
-// DeletionObserved atomically decrements the `del` expectation count of the given controller.
-func (r *ControllerExpectations) DeletionObserved(controllerKey string) {
- r.LowerExpectations(controllerKey, 0, 1)
-}
-
-// Expectations are either fulfilled, or expire naturally.
-type Expectations interface {
- Fulfilled() bool
-}
-
-// ControlleeExpectations track controllee creates/deletes.
-type ControlleeExpectations struct {
- // Important: Since these two int64 fields are using sync/atomic, they have to be at the top of the struct due to a bug on 32-bit platforms
- // See: https://golang.org/pkg/sync/atomic/ for more information
- add int64
- del int64
- key string
- timestamp time.Time
-}
-
-// Add increments the add and del counters.
-func (e *ControlleeExpectations) Add(add, del int64) {
- atomic.AddInt64(&e.add, add)
- atomic.AddInt64(&e.del, del)
-}
-
-// Fulfilled returns true if this expectation has been fulfilled.
-func (e *ControlleeExpectations) Fulfilled() bool {
- // TODO: think about why this line being atomic doesn't matter
- return atomic.LoadInt64(&e.add) <= 0 && atomic.LoadInt64(&e.del) <= 0
-}
-
-// GetExpectations returns the add and del expectations of the controllee.
-func (e *ControlleeExpectations) GetExpectations() (int64, int64) {
- return atomic.LoadInt64(&e.add), atomic.LoadInt64(&e.del)
-}
-
-// NewControllerExpectations returns a store for ControllerExpectations.
-func NewControllerExpectations() *ControllerExpectations {
- return &ControllerExpectations{cache.NewStore(ExpKeyFunc)}
-}
-
-// UIDSetKeyFunc to parse out the key from a UIDSet.
-var UIDSetKeyFunc = func(obj interface{}) (string, error) {
- if u, ok := obj.(*UIDSet); ok {
- return u.key, nil
- }
- return "", fmt.Errorf("Could not find key for obj %#v", obj)
-}
-
-// UIDSet holds a key and a set of UIDs. Used by the
-// UIDTrackingControllerExpectations to remember which UID it has seen/still
-// waiting for.
-type UIDSet struct {
- sets.String
- key string
-}
-
-// UIDTrackingControllerExpectations tracks the UID of the pods it deletes.
-// This cache is needed over plain old expectations to safely handle graceful
-// deletion. The desired behavior is to treat an update that sets the
-// DeletionTimestamp on an object as a delete. To do so consistently, one needs
-// to remember the expected deletes so they aren't double counted.
-// TODO: Track creates as well (#22599)
-type UIDTrackingControllerExpectations struct {
- ControllerExpectationsInterface
- // TODO: There is a much nicer way to do this that involves a single store,
- // a lock per entry, and a ControlleeExpectationsInterface type.
- uidStoreLock sync.Mutex
- // Store used for the UIDs associated with any expectation tracked via the
- // ControllerExpectationsInterface.
- uidStore cache.Store
-}
-
-// GetUIDs is a convenience method to avoid exposing the set of expected uids.
-// The returned set is not thread safe, all modifications must be made holding
-// the uidStoreLock.
-func (u *UIDTrackingControllerExpectations) GetUIDs(controllerKey string) sets.String {
- if uid, exists, err := u.uidStore.GetByKey(controllerKey); err == nil && exists {
- return uid.(*UIDSet).String
- }
- return nil
-}
-
-// ExpectDeletions records expectations for the given deleteKeys, against the given controller.
-func (u *UIDTrackingControllerExpectations) ExpectDeletions(rcKey string, deletedKeys []string) error {
- u.uidStoreLock.Lock()
- defer u.uidStoreLock.Unlock()
-
- if existing := u.GetUIDs(rcKey); existing != nil && existing.Len() != 0 {
- glog.Errorf("Clobbering existing delete keys: %+v", existing)
- }
- expectedUIDs := sets.NewString()
- for _, k := range deletedKeys {
- expectedUIDs.Insert(k)
- }
- glog.V(4).Infof("Controller %v waiting on deletions for: %+v", rcKey, deletedKeys)
- if err := u.uidStore.Add(&UIDSet{expectedUIDs, rcKey}); err != nil {
- return err
- }
- return u.ControllerExpectationsInterface.ExpectDeletions(rcKey, expectedUIDs.Len())
-}
-
-// DeletionObserved records the given deleteKey as a deletion, for the given rc.
-func (u *UIDTrackingControllerExpectations) DeletionObserved(rcKey, deleteKey string) {
- u.uidStoreLock.Lock()
- defer u.uidStoreLock.Unlock()
-
- uids := u.GetUIDs(rcKey)
- if uids != nil && uids.Has(deleteKey) {
- glog.V(4).Infof("Controller %v received delete for pod %v", rcKey, deleteKey)
- u.ControllerExpectationsInterface.DeletionObserved(rcKey)
- uids.Delete(deleteKey)
- }
-}
-
-// DeleteExpectations deletes the UID set and invokes DeleteExpectations on the
-// underlying ControllerExpectationsInterface.
-func (u *UIDTrackingControllerExpectations) DeleteExpectations(rcKey string) {
- u.uidStoreLock.Lock()
- defer u.uidStoreLock.Unlock()
-
- u.ControllerExpectationsInterface.DeleteExpectations(rcKey)
- if uidExp, exists, err := u.uidStore.GetByKey(rcKey); err == nil && exists {
- if err := u.uidStore.Delete(uidExp); err != nil {
- glog.V(2).Infof("Error deleting uid expectations for controller %v: %v", rcKey, err)
- }
- }
-}
-
-// NewUIDTrackingControllerExpectations returns a wrapper around
-// ControllerExpectations that is aware of deleteKeys.
-func NewUIDTrackingControllerExpectations(ce ControllerExpectationsInterface) *UIDTrackingControllerExpectations {
- return &UIDTrackingControllerExpectations{ControllerExpectationsInterface: ce, uidStore: cache.NewStore(UIDSetKeyFunc)}
-}
-
-// Reasons for pod events
-const (
- // FailedCreatePodReason is added in an event and in a replica set condition
- // when a pod for a replica set is failed to be created.
- FailedCreatePodReason = "FailedCreate"
- // SuccessfulCreatePodReason is added in an event when a pod for a replica set
- // is successfully created.
- SuccessfulCreatePodReason = "SuccessfulCreate"
- // FailedDeletePodReason is added in an event and in a replica set condition
- // when a pod for a replica set is failed to be deleted.
- FailedDeletePodReason = "FailedDelete"
- // SuccessfulDeletePodReason is added in an event when a pod for a replica set
- // is successfully deleted.
- SuccessfulDeletePodReason = "SuccessfulDelete"
-)
-
-// RSControlInterface is an interface that knows how to add or delete
-// ReplicaSets, as well as increment or decrement them. It is used
-// by the deployment controller to ease testing of actions that it takes.
-type RSControlInterface interface {
- PatchReplicaSet(namespace, name string, data []byte) error
-}
-
-// RealRSControl is the default implementation of RSControllerInterface.
-type RealRSControl struct {
- KubeClient clientset.Interface
- Recorder record.EventRecorder
-}
-
-var _ RSControlInterface = &RealRSControl{}
-
-func (r RealRSControl) PatchReplicaSet(namespace, name string, data []byte) error {
- _, err := r.KubeClient.ExtensionsV1beta1().ReplicaSets(namespace).Patch(name, types.StrategicMergePatchType, data)
- return err
-}
-
-// TODO: merge the controller revision interface in controller_history.go with this one
-// ControllerRevisionControlInterface is an interface that knows how to patch
-// ControllerRevisions, as well as increment or decrement them. It is used
-// by the daemonset controller to ease testing of actions that it takes.
-type ControllerRevisionControlInterface interface {
- PatchControllerRevision(namespace, name string, data []byte) error
-}
-
-// RealControllerRevisionControl is the default implementation of ControllerRevisionControlInterface.
-type RealControllerRevisionControl struct {
- KubeClient clientset.Interface
-}
-
-var _ ControllerRevisionControlInterface = &RealControllerRevisionControl{}
-
-func (r RealControllerRevisionControl) PatchControllerRevision(namespace, name string, data []byte) error {
- _, err := r.KubeClient.AppsV1beta1().ControllerRevisions(namespace).Patch(name, types.StrategicMergePatchType, data)
- return err
-}
-
-// PodControlInterface is an interface that knows how to add or delete pods
-// created as an interface to allow testing.
-type PodControlInterface interface {
- // CreatePods creates new pods according to the spec.
- CreatePods(namespace string, template *v1.PodTemplateSpec, object runtime.Object) error
- // CreatePodsOnNode creates a new pod according to the spec on the specified node,
- // and sets the ControllerRef.
- CreatePodsOnNode(nodeName, namespace string, template *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error
- // CreatePodsWithControllerRef creates new pods according to the spec, and sets object as the pod's controller.
- CreatePodsWithControllerRef(namespace string, template *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error
- // DeletePod deletes the pod identified by podID.
- DeletePod(namespace string, podID string, object runtime.Object) error
- // PatchPod patches the pod.
- PatchPod(namespace, name string, data []byte) error
-}
-
-// RealPodControl is the default implementation of PodControlInterface.
-type RealPodControl struct {
- KubeClient clientset.Interface
- Recorder record.EventRecorder
-}
-
-var _ PodControlInterface = &RealPodControl{}
-
-func getPodsLabelSet(template *v1.PodTemplateSpec) labels.Set {
- desiredLabels := make(labels.Set)
- for k, v := range template.Labels {
- desiredLabels[k] = v
- }
- return desiredLabels
-}
-
-func getPodsFinalizers(template *v1.PodTemplateSpec) []string {
- desiredFinalizers := make([]string, len(template.Finalizers))
- copy(desiredFinalizers, template.Finalizers)
- return desiredFinalizers
-}
-
-func getPodsAnnotationSet(template *v1.PodTemplateSpec) labels.Set {
- desiredAnnotations := make(labels.Set)
- for k, v := range template.Annotations {
- desiredAnnotations[k] = v
- }
- return desiredAnnotations
-}
-
-func getPodsPrefix(controllerName string) string {
- // use the dash (if the name isn't too long) to make the pod name a bit prettier
- prefix := fmt.Sprintf("%s-", controllerName)
- if len(validation.ValidatePodName(prefix, true)) != 0 {
- prefix = controllerName
- }
- return prefix
-}
-
-func validateControllerRef(controllerRef *metav1.OwnerReference) error {
- if controllerRef == nil {
- return fmt.Errorf("controllerRef is nil")
- }
- if len(controllerRef.APIVersion) == 0 {
- return fmt.Errorf("controllerRef has empty APIVersion")
- }
- if len(controllerRef.Kind) == 0 {
- return fmt.Errorf("controllerRef has empty Kind")
- }
- if controllerRef.Controller == nil || *controllerRef.Controller != true {
- return fmt.Errorf("controllerRef.Controller is not set to true")
- }
- if controllerRef.BlockOwnerDeletion == nil || *controllerRef.BlockOwnerDeletion != true {
- return fmt.Errorf("controllerRef.BlockOwnerDeletion is not set")
- }
- return nil
-}
-
-func (r RealPodControl) CreatePods(namespace string, template *v1.PodTemplateSpec, object runtime.Object) error {
- return r.createPods("", namespace, template, object, nil)
-}
-
-func (r RealPodControl) CreatePodsWithControllerRef(namespace string, template *v1.PodTemplateSpec, controllerObject runtime.Object, controllerRef *metav1.OwnerReference) error {
- if err := validateControllerRef(controllerRef); err != nil {
- return err
- }
- return r.createPods("", namespace, template, controllerObject, controllerRef)
-}
-
-func (r RealPodControl) CreatePodsOnNode(nodeName, namespace string, template *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error {
- if err := validateControllerRef(controllerRef); err != nil {
- return err
- }
- return r.createPods(nodeName, namespace, template, object, controllerRef)
-}
-
-func (r RealPodControl) PatchPod(namespace, name string, data []byte) error {
- _, err := r.KubeClient.CoreV1().Pods(namespace).Patch(name, types.StrategicMergePatchType, data)
- return err
-}
-
-func GetPodFromTemplate(template *v1.PodTemplateSpec, parentObject runtime.Object, controllerRef *metav1.OwnerReference) (*v1.Pod, error) {
- desiredLabels := getPodsLabelSet(template)
- desiredFinalizers := getPodsFinalizers(template)
- desiredAnnotations := getPodsAnnotationSet(template)
- accessor, err := meta.Accessor(parentObject)
- if err != nil {
- return nil, fmt.Errorf("parentObject does not have ObjectMeta, %v", err)
- }
- prefix := getPodsPrefix(accessor.GetName())
-
- pod := &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Labels: desiredLabels,
- Annotations: desiredAnnotations,
- GenerateName: prefix,
- Finalizers: desiredFinalizers,
- },
- }
- if controllerRef != nil {
- pod.OwnerReferences = append(pod.OwnerReferences, *controllerRef)
- }
- pod.Spec = *template.Spec.DeepCopy()
- return pod, nil
-}
-
-func (r RealPodControl) createPods(nodeName, namespace string, template *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error {
- pod, err := GetPodFromTemplate(template, object, controllerRef)
- if err != nil {
- return err
- }
- if len(nodeName) != 0 {
- pod.Spec.NodeName = nodeName
- }
- if labels.Set(pod.Labels).AsSelectorPreValidated().Empty() {
- return fmt.Errorf("unable to create pods, no labels")
- }
- if newPod, err := r.KubeClient.CoreV1().Pods(namespace).Create(pod); err != nil {
- r.Recorder.Eventf(object, v1.EventTypeWarning, FailedCreatePodReason, "Error creating: %v", err)
- return err
- } else {
- accessor, err := meta.Accessor(object)
- if err != nil {
- glog.Errorf("parentObject does not have ObjectMeta, %v", err)
- return nil
- }
- glog.V(4).Infof("Controller %v created pod %v", accessor.GetName(), newPod.Name)
- r.Recorder.Eventf(object, v1.EventTypeNormal, SuccessfulCreatePodReason, "Created pod: %v", newPod.Name)
- }
- return nil
-}
-
-func (r RealPodControl) DeletePod(namespace string, podID string, object runtime.Object) error {
- accessor, err := meta.Accessor(object)
- if err != nil {
- return fmt.Errorf("object does not have ObjectMeta, %v", err)
- }
- glog.V(2).Infof("Controller %v deleting pod %v/%v", accessor.GetName(), namespace, podID)
- if err := r.KubeClient.CoreV1().Pods(namespace).Delete(podID, nil); err != nil {
- r.Recorder.Eventf(object, v1.EventTypeWarning, FailedDeletePodReason, "Error deleting: %v", err)
- return fmt.Errorf("unable to delete pods: %v", err)
- } else {
- r.Recorder.Eventf(object, v1.EventTypeNormal, SuccessfulDeletePodReason, "Deleted pod: %v", podID)
- }
- return nil
-}
-
-type FakePodControl struct {
- sync.Mutex
- Templates []v1.PodTemplateSpec
- ControllerRefs []metav1.OwnerReference
- DeletePodName []string
- Patches [][]byte
- Err error
- CreateLimit int
- CreateCallCount int
-}
-
-var _ PodControlInterface = &FakePodControl{}
-
-func (f *FakePodControl) PatchPod(namespace, name string, data []byte) error {
- f.Lock()
- defer f.Unlock()
- f.Patches = append(f.Patches, data)
- if f.Err != nil {
- return f.Err
- }
- return nil
-}
-
-func (f *FakePodControl) CreatePods(namespace string, spec *v1.PodTemplateSpec, object runtime.Object) error {
- f.Lock()
- defer f.Unlock()
- f.CreateCallCount++
- if f.CreateLimit != 0 && f.CreateCallCount > f.CreateLimit {
- return fmt.Errorf("Not creating pod, limit %d already reached (create call %d)", f.CreateLimit, f.CreateCallCount)
- }
- f.Templates = append(f.Templates, *spec)
- if f.Err != nil {
- return f.Err
- }
- return nil
-}
-
-func (f *FakePodControl) CreatePodsWithControllerRef(namespace string, spec *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error {
- f.Lock()
- defer f.Unlock()
- f.CreateCallCount++
- if f.CreateLimit != 0 && f.CreateCallCount > f.CreateLimit {
- return fmt.Errorf("Not creating pod, limit %d already reached (create call %d)", f.CreateLimit, f.CreateCallCount)
- }
- f.Templates = append(f.Templates, *spec)
- f.ControllerRefs = append(f.ControllerRefs, *controllerRef)
- if f.Err != nil {
- return f.Err
- }
- return nil
-}
-
-func (f *FakePodControl) CreatePodsOnNode(nodeName, namespace string, template *v1.PodTemplateSpec, object runtime.Object, controllerRef *metav1.OwnerReference) error {
- f.Lock()
- defer f.Unlock()
- f.CreateCallCount++
- if f.CreateLimit != 0 && f.CreateCallCount > f.CreateLimit {
- return fmt.Errorf("Not creating pod, limit %d already reached (create call %d)", f.CreateLimit, f.CreateCallCount)
- }
- f.Templates = append(f.Templates, *template)
- f.ControllerRefs = append(f.ControllerRefs, *controllerRef)
- if f.Err != nil {
- return f.Err
- }
- return nil
-}
-
-func (f *FakePodControl) DeletePod(namespace string, podID string, object runtime.Object) error {
- f.Lock()
- defer f.Unlock()
- f.DeletePodName = append(f.DeletePodName, podID)
- if f.Err != nil {
- return f.Err
- }
- return nil
-}
-
-func (f *FakePodControl) Clear() {
- f.Lock()
- defer f.Unlock()
- f.DeletePodName = []string{}
- f.Templates = []v1.PodTemplateSpec{}
- f.ControllerRefs = []metav1.OwnerReference{}
- f.Patches = [][]byte{}
- f.CreateLimit = 0
- f.CreateCallCount = 0
-}
-
-// ByLogging allows custom sorting of pods so the best one can be picked for getting its logs.
-type ByLogging []*v1.Pod
-
-func (s ByLogging) Len() int { return len(s) }
-func (s ByLogging) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
-
-func (s ByLogging) Less(i, j int) bool {
- // 1. assigned < unassigned
- if s[i].Spec.NodeName != s[j].Spec.NodeName && (len(s[i].Spec.NodeName) == 0 || len(s[j].Spec.NodeName) == 0) {
- return len(s[i].Spec.NodeName) > 0
- }
- // 2. PodRunning < PodUnknown < PodPending
- m := map[v1.PodPhase]int{v1.PodRunning: 0, v1.PodUnknown: 1, v1.PodPending: 2}
- if m[s[i].Status.Phase] != m[s[j].Status.Phase] {
- return m[s[i].Status.Phase] < m[s[j].Status.Phase]
- }
- // 3. ready < not ready
- if podutil.IsPodReady(s[i]) != podutil.IsPodReady(s[j]) {
- return podutil.IsPodReady(s[i])
- }
- // TODO: take availability into account when we push minReadySeconds information from deployment into pods,
- // see https://github.com/kubernetes/kubernetes/issues/22065
- // 4. Been ready for more time < less time < empty time
- if podutil.IsPodReady(s[i]) && podutil.IsPodReady(s[j]) && !podReadyTime(s[i]).Equal(podReadyTime(s[j])) {
- return afterOrZero(podReadyTime(s[j]), podReadyTime(s[i]))
- }
- // 5. Pods with containers with higher restart counts < lower restart counts
- if maxContainerRestarts(s[i]) != maxContainerRestarts(s[j]) {
- return maxContainerRestarts(s[i]) > maxContainerRestarts(s[j])
- }
- // 6. older pods < newer pods < empty timestamp pods
- if !s[i].CreationTimestamp.Equal(&s[j].CreationTimestamp) {
- return afterOrZero(&s[j].CreationTimestamp, &s[i].CreationTimestamp)
- }
- return false
-}
-
-// ActivePods type allows custom sorting of pods so a controller can pick the best ones to delete.
-type ActivePods []*v1.Pod
-
-func (s ActivePods) Len() int { return len(s) }
-func (s ActivePods) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
-
-func (s ActivePods) Less(i, j int) bool {
- // 1. Unassigned < assigned
- // If only one of the pods is unassigned, the unassigned one is smaller
- if s[i].Spec.NodeName != s[j].Spec.NodeName && (len(s[i].Spec.NodeName) == 0 || len(s[j].Spec.NodeName) == 0) {
- return len(s[i].Spec.NodeName) == 0
- }
- // 2. PodPending < PodUnknown < PodRunning
- m := map[v1.PodPhase]int{v1.PodPending: 0, v1.PodUnknown: 1, v1.PodRunning: 2}
- if m[s[i].Status.Phase] != m[s[j].Status.Phase] {
- return m[s[i].Status.Phase] < m[s[j].Status.Phase]
- }
- // 3. Not ready < ready
- // If only one of the pods is not ready, the not ready one is smaller
- if podutil.IsPodReady(s[i]) != podutil.IsPodReady(s[j]) {
- return !podutil.IsPodReady(s[i])
- }
- // TODO: take availability into account when we push minReadySeconds information from deployment into pods,
- // see https://github.com/kubernetes/kubernetes/issues/22065
- // 4. Been ready for empty time < less time < more time
- // If both pods are ready, the latest ready one is smaller
- if podutil.IsPodReady(s[i]) && podutil.IsPodReady(s[j]) && !podReadyTime(s[i]).Equal(podReadyTime(s[j])) {
- return afterOrZero(podReadyTime(s[i]), podReadyTime(s[j]))
- }
- // 5. Pods with containers with higher restart counts < lower restart counts
- if maxContainerRestarts(s[i]) != maxContainerRestarts(s[j]) {
- return maxContainerRestarts(s[i]) > maxContainerRestarts(s[j])
- }
- // 6. Empty creation time pods < newer pods < older pods
- if !s[i].CreationTimestamp.Equal(&s[j].CreationTimestamp) {
- return afterOrZero(&s[i].CreationTimestamp, &s[j].CreationTimestamp)
- }
- return false
-}
-
-// afterOrZero checks if time t1 is after time t2; if one of them
-// is zero, the zero time is seen as after non-zero time.
-func afterOrZero(t1, t2 *metav1.Time) bool {
- if t1.Time.IsZero() || t2.Time.IsZero() {
- return t1.Time.IsZero()
- }
- return t1.After(t2.Time)
-}
-
-func podReadyTime(pod *v1.Pod) *metav1.Time {
- if podutil.IsPodReady(pod) {
- for _, c := range pod.Status.Conditions {
- // we only care about pod ready conditions
- if c.Type == v1.PodReady && c.Status == v1.ConditionTrue {
- return &c.LastTransitionTime
- }
- }
- }
- return &metav1.Time{}
-}
-
-func maxContainerRestarts(pod *v1.Pod) int {
- maxRestarts := 0
- for _, c := range pod.Status.ContainerStatuses {
- maxRestarts = integer.IntMax(maxRestarts, int(c.RestartCount))
- }
- return maxRestarts
-}
-
-// FilterActivePods returns pods that have not terminated.
-func FilterActivePods(pods []*v1.Pod) []*v1.Pod {
- var result []*v1.Pod
- for _, p := range pods {
- if IsPodActive(p) {
- result = append(result, p)
- } else {
- glog.V(4).Infof("Ignoring inactive pod %v/%v in state %v, deletion time %v",
- p.Namespace, p.Name, p.Status.Phase, p.DeletionTimestamp)
- }
- }
- return result
-}
-
-func IsPodActive(p *v1.Pod) bool {
- return v1.PodSucceeded != p.Status.Phase &&
- v1.PodFailed != p.Status.Phase &&
- p.DeletionTimestamp == nil
-}
-
-// FilterActiveReplicaSets returns replica sets that have (or at least ought to have) pods.
-func FilterActiveReplicaSets(replicaSets []*extensions.ReplicaSet) []*extensions.ReplicaSet {
- activeFilter := func(rs *extensions.ReplicaSet) bool {
- return rs != nil && *(rs.Spec.Replicas) > 0
- }
- return FilterReplicaSets(replicaSets, activeFilter)
-}
-
-type filterRS func(rs *extensions.ReplicaSet) bool
-
-// FilterReplicaSets returns replica sets that are filtered by filterFn (all returned ones should match filterFn).
-func FilterReplicaSets(RSes []*extensions.ReplicaSet, filterFn filterRS) []*extensions.ReplicaSet {
- var filtered []*extensions.ReplicaSet
- for i := range RSes {
- if filterFn(RSes[i]) {
- filtered = append(filtered, RSes[i])
- }
- }
- return filtered
-}
-
-// PodKey returns a key unique to the given pod within a cluster.
-// It's used so we consistently use the same key scheme in this module.
-// It does exactly what cache.MetaNamespaceKeyFunc would have done
-// except there's not possibility for error since we know the exact type.
-func PodKey(pod *v1.Pod) string {
- return fmt.Sprintf("%v/%v", pod.Namespace, pod.Name)
-}
-
-// ControllersByCreationTimestamp sorts a list of ReplicationControllers by creation timestamp, using their names as a tie breaker.
-type ControllersByCreationTimestamp []*v1.ReplicationController
-
-func (o ControllersByCreationTimestamp) Len() int { return len(o) }
-func (o ControllersByCreationTimestamp) Swap(i, j int) { o[i], o[j] = o[j], o[i] }
-func (o ControllersByCreationTimestamp) Less(i, j int) bool {
- if o[i].CreationTimestamp.Equal(&o[j].CreationTimestamp) {
- return o[i].Name < o[j].Name
- }
- return o[i].CreationTimestamp.Before(&o[j].CreationTimestamp)
-}
-
-// ReplicaSetsByCreationTimestamp sorts a list of ReplicaSet by creation timestamp, using their names as a tie breaker.
-type ReplicaSetsByCreationTimestamp []*extensions.ReplicaSet
-
-func (o ReplicaSetsByCreationTimestamp) Len() int { return len(o) }
-func (o ReplicaSetsByCreationTimestamp) Swap(i, j int) { o[i], o[j] = o[j], o[i] }
-func (o ReplicaSetsByCreationTimestamp) Less(i, j int) bool {
- if o[i].CreationTimestamp.Equal(&o[j].CreationTimestamp) {
- return o[i].Name < o[j].Name
- }
- return o[i].CreationTimestamp.Before(&o[j].CreationTimestamp)
-}
-
-// ReplicaSetsBySizeOlder sorts a list of ReplicaSet by size in descending order, using their creation timestamp or name as a tie breaker.
-// By using the creation timestamp, this sorts from old to new replica sets.
-type ReplicaSetsBySizeOlder []*extensions.ReplicaSet
-
-func (o ReplicaSetsBySizeOlder) Len() int { return len(o) }
-func (o ReplicaSetsBySizeOlder) Swap(i, j int) { o[i], o[j] = o[j], o[i] }
-func (o ReplicaSetsBySizeOlder) Less(i, j int) bool {
- if *(o[i].Spec.Replicas) == *(o[j].Spec.Replicas) {
- return ReplicaSetsByCreationTimestamp(o).Less(i, j)
- }
- return *(o[i].Spec.Replicas) > *(o[j].Spec.Replicas)
-}
-
-// ReplicaSetsBySizeNewer sorts a list of ReplicaSet by size in descending order, using their creation timestamp or name as a tie breaker.
-// By using the creation timestamp, this sorts from new to old replica sets.
-type ReplicaSetsBySizeNewer []*extensions.ReplicaSet
-
-func (o ReplicaSetsBySizeNewer) Len() int { return len(o) }
-func (o ReplicaSetsBySizeNewer) Swap(i, j int) { o[i], o[j] = o[j], o[i] }
-func (o ReplicaSetsBySizeNewer) Less(i, j int) bool {
- if *(o[i].Spec.Replicas) == *(o[j].Spec.Replicas) {
- return ReplicaSetsByCreationTimestamp(o).Less(j, i)
- }
- return *(o[i].Spec.Replicas) > *(o[j].Spec.Replicas)
-}
-
-// AddOrUpdateTaintOnNode add taints to the node. If taint was added into node, it'll issue API calls
-// to update nodes; otherwise, no API calls. Return error if any.
-func AddOrUpdateTaintOnNode(c clientset.Interface, nodeName string, taints ...*v1.Taint) error {
- if len(taints) == 0 {
- return nil
- }
- firstTry := true
- return clientretry.RetryOnConflict(UpdateTaintBackoff, func() error {
- var err error
- var oldNode *v1.Node
- // First we try getting node from the API server cache, as it's cheaper. If it fails
- // we get it from etcd to be sure to have fresh data.
- if firstTry {
- oldNode, err = c.CoreV1().Nodes().Get(nodeName, metav1.GetOptions{ResourceVersion: "0"})
- firstTry = false
- } else {
- oldNode, err = c.CoreV1().Nodes().Get(nodeName, metav1.GetOptions{})
- }
- if err != nil {
- return err
- }
-
- var newNode *v1.Node
- oldNodeCopy := oldNode
- updated := false
- for _, taint := range taints {
- curNewNode, ok, err := taintutils.AddOrUpdateTaint(oldNodeCopy, taint)
- if err != nil {
- return fmt.Errorf("Failed to update taint of node!")
- }
- updated = updated || ok
- newNode = curNewNode
- oldNodeCopy = curNewNode
- }
- if !updated {
- return nil
- }
- return PatchNodeTaints(c, nodeName, oldNode, newNode)
- })
-}
-
-// RemoveTaintOffNode is for cleaning up taints temporarily added to node,
-// won't fail if target taint doesn't exist or has been removed.
-// If passed a node it'll check if there's anything to be done, if taint is not present it won't issue
-// any API calls.
-func RemoveTaintOffNode(c clientset.Interface, nodeName string, node *v1.Node, taints ...*v1.Taint) error {
- if len(taints) == 0 {
- return nil
- }
- // Short circuit for limiting amount of API calls.
- if node != nil {
- match := false
- for _, taint := range taints {
- if taintutils.TaintExists(node.Spec.Taints, taint) {
- match = true
- break
- }
- }
- if !match {
- return nil
- }
- }
-
- firstTry := true
- return clientretry.RetryOnConflict(UpdateTaintBackoff, func() error {
- var err error
- var oldNode *v1.Node
- // First we try getting node from the API server cache, as it's cheaper. If it fails
- // we get it from etcd to be sure to have fresh data.
- if firstTry {
- oldNode, err = c.CoreV1().Nodes().Get(nodeName, metav1.GetOptions{ResourceVersion: "0"})
- firstTry = false
- } else {
- oldNode, err = c.CoreV1().Nodes().Get(nodeName, metav1.GetOptions{})
- }
- if err != nil {
- return err
- }
-
- var newNode *v1.Node
- oldNodeCopy := oldNode
- updated := false
- for _, taint := range taints {
- curNewNode, ok, err := taintutils.RemoveTaint(oldNodeCopy, taint)
- if err != nil {
- return fmt.Errorf("Failed to remove taint of node!")
- }
- updated = updated || ok
- newNode = curNewNode
- oldNodeCopy = curNewNode
- }
- if !updated {
- return nil
- }
- return PatchNodeTaints(c, nodeName, oldNode, newNode)
- })
-}
-
-// PatchNodeTaints patches node's taints.
-func PatchNodeTaints(c clientset.Interface, nodeName string, oldNode *v1.Node, newNode *v1.Node) error {
- oldData, err := json.Marshal(oldNode)
- if err != nil {
- return fmt.Errorf("failed to marshal old node %#v for node %q: %v", oldNode, nodeName, err)
- }
-
- newTaints := newNode.Spec.Taints
- newNodeClone := oldNode.DeepCopy()
- newNodeClone.Spec.Taints = newTaints
- newData, err := json.Marshal(newNodeClone)
- if err != nil {
- return fmt.Errorf("failed to marshal new node %#v for node %q: %v", newNodeClone, nodeName, err)
- }
-
- patchBytes, err := strategicpatch.CreateTwoWayMergePatch(oldData, newData, v1.Node{})
- if err != nil {
- return fmt.Errorf("failed to create patch for node %q: %v", nodeName, err)
- }
-
- _, err = c.CoreV1().Nodes().Patch(nodeName, types.StrategicMergePatchType, patchBytes)
- return err
-}
-
-// WaitForCacheSync is a wrapper around cache.WaitForCacheSync that generates log messages
-// indicating that the controller identified by controllerName is waiting for syncs, followed by
-// either a successful or failed sync.
-func WaitForCacheSync(controllerName string, stopCh <-chan struct{}, cacheSyncs ...cache.InformerSynced) bool {
- glog.Infof("Waiting for caches to sync for %s controller", controllerName)
-
- if !cache.WaitForCacheSync(stopCh, cacheSyncs...) {
- utilruntime.HandleError(fmt.Errorf("Unable to sync caches for %s controller", controllerName))
- return false
- }
-
- glog.Infof("Caches are synced for %s controller", controllerName)
- return true
-}
-
-// ComputeHash returns a hash value calculated from pod template and a collisionCount to avoid hash collision
-func ComputeHash(template *v1.PodTemplateSpec, collisionCount *int32) uint32 {
- podTemplateSpecHasher := fnv.New32a()
- hashutil.DeepHashObject(podTemplateSpecHasher, *template)
-
- // Add collisionCount in the hash if it exists.
- if collisionCount != nil {
- collisionCountBytes := make([]byte, 8)
- binary.LittleEndian.PutUint32(collisionCountBytes, uint32(*collisionCount))
- podTemplateSpecHasher.Write(collisionCountBytes)
- }
-
- return podTemplateSpecHasher.Sum32()
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/controller/doc.go b/vendor/k8s.io/kubernetes/pkg/controller/doc.go
deleted file mode 100644
index 3c5c943da..000000000
--- a/vendor/k8s.io/kubernetes/pkg/controller/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package controller contains code for controllers (like the replication
-// controller).
-package controller // import "k8s.io/kubernetes/pkg/controller"
diff --git a/vendor/k8s.io/kubernetes/pkg/controller/lookup_cache.go b/vendor/k8s.io/kubernetes/pkg/controller/lookup_cache.go
deleted file mode 100644
index 160aa6e08..000000000
--- a/vendor/k8s.io/kubernetes/pkg/controller/lookup_cache.go
+++ /dev/null
@@ -1,92 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package controller
-
-import (
- "hash/fnv"
- "sync"
-
- "github.com/golang/groupcache/lru"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- hashutil "k8s.io/kubernetes/pkg/util/hash"
-)
-
-type objectWithMeta interface {
- metav1.Object
-}
-
-// keyFunc returns the key of an object, which is used to look up in the cache for it's matching object.
-// Since we match objects by namespace and Labels/Selector, so if two objects have the same namespace and labels,
-// they will have the same key.
-func keyFunc(obj objectWithMeta) uint64 {
- hash := fnv.New32a()
- hashutil.DeepHashObject(hash, &equivalenceLabelObj{
- namespace: obj.GetNamespace(),
- labels: obj.GetLabels(),
- })
- return uint64(hash.Sum32())
-}
-
-type equivalenceLabelObj struct {
- namespace string
- labels map[string]string
-}
-
-// MatchingCache save label and selector matching relationship
-type MatchingCache struct {
- mutex sync.RWMutex
- cache *lru.Cache
-}
-
-// NewMatchingCache return a NewMatchingCache, which save label and selector matching relationship.
-func NewMatchingCache(maxCacheEntries int) *MatchingCache {
- return &MatchingCache{
- cache: lru.New(maxCacheEntries),
- }
-}
-
-// Add will add matching information to the cache.
-func (c *MatchingCache) Add(labelObj objectWithMeta, selectorObj objectWithMeta) {
- key := keyFunc(labelObj)
- c.mutex.Lock()
- defer c.mutex.Unlock()
- c.cache.Add(key, selectorObj)
-}
-
-// GetMatchingObject lookup the matching object for a given object.
-// Note: the cache information may be invalid since the controller may be deleted or updated,
-// we need check in the external request to ensure the cache data is not dirty.
-func (c *MatchingCache) GetMatchingObject(labelObj objectWithMeta) (controller interface{}, exists bool) {
- key := keyFunc(labelObj)
- // NOTE: we use Lock() instead of RLock() here because lru's Get() method also modifies state(
- // it need update the least recently usage information). So we can not call it concurrently.
- c.mutex.Lock()
- defer c.mutex.Unlock()
- return c.cache.Get(key)
-}
-
-// Update update the cached matching information.
-func (c *MatchingCache) Update(labelObj objectWithMeta, selectorObj objectWithMeta) {
- c.Add(labelObj, selectorObj)
-}
-
-// InvalidateAll invalidate the whole cache.
-func (c *MatchingCache) InvalidateAll() {
- c.mutex.Lock()
- defer c.mutex.Unlock()
- c.cache = lru.New(c.cache.MaxEntries)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/features/kube_features.go b/vendor/k8s.io/kubernetes/pkg/features/kube_features.go
deleted file mode 100644
index c2fb4896d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/features/kube_features.go
+++ /dev/null
@@ -1,326 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package features
-
-import (
- apiextensionsfeatures "k8s.io/apiextensions-apiserver/pkg/features"
- genericfeatures "k8s.io/apiserver/pkg/features"
- utilfeature "k8s.io/apiserver/pkg/util/feature"
-)
-
-const (
- // Every feature gate should add method here following this template:
- //
- // // owner: @username
- // // alpha: v1.X
- // MyFeature utilfeature.Feature = "MyFeature"
-
- // owner: @tallclair
- // beta: v1.4
- AppArmor utilfeature.Feature = "AppArmor"
-
- // owner: @mtaufen
- // alpha: v1.4
- DynamicKubeletConfig utilfeature.Feature = "DynamicKubeletConfig"
-
- // owner: @pweil-
- // alpha: v1.5
- //
- // Default userns=host for containers that are using other host namespaces, host mounts, the pod
- // contains a privileged container, or specific non-namespaced capabilities (MKNOD, SYS_MODULE,
- // SYS_TIME). This should only be enabled if user namespace remapping is enabled in the docker daemon.
- ExperimentalHostUserNamespaceDefaultingGate utilfeature.Feature = "ExperimentalHostUserNamespaceDefaulting"
-
- // owner: @vishh
- // alpha: v1.5
- //
- // Ensures guaranteed scheduling of pods marked with a special pod annotation `scheduler.alpha.kubernetes.io/critical-pod`
- // and also prevents them from being evicted from a node.
- // Note: This feature is not supported for `BestEffort` pods.
- ExperimentalCriticalPodAnnotation utilfeature.Feature = "ExperimentalCriticalPodAnnotation"
-
- // owner: @vishh
- // alpha: v1.6
- //
- // This is deprecated and will be removed in v1.11. Use DevicePlugins instead.
- //
- // Enables support for GPUs as a schedulable resource.
- // Only Nvidia GPUs are supported as of v1.6.
- // Works only with Docker Container Runtime.
- Accelerators utilfeature.Feature = "Accelerators"
-
- // owner: @jiayingz
- // beta: v1.10
- //
- // Enables support for Device Plugins
- DevicePlugins utilfeature.Feature = "DevicePlugins"
-
- // owner: @gmarek
- // alpha: v1.6
- //
- // Changes the logic behind evicting Pods from not ready Nodes
- // to take advantage of NoExecute Taints and Tolerations.
- TaintBasedEvictions utilfeature.Feature = "TaintBasedEvictions"
-
- // owner: @jcbsmpsn
- // alpha: v1.7
- //
- // Gets a server certificate for the kubelet from the Certificate Signing
- // Request API instead of generating one self signed and auto rotates the
- // certificate as expiration approaches.
- RotateKubeletServerCertificate utilfeature.Feature = "RotateKubeletServerCertificate"
-
- // owner: @jcbsmpsn
- // alpha: v1.7
- //
- // Automatically renews the client certificate used for communicating with
- // the API server as the certificate approaches expiration.
- RotateKubeletClientCertificate utilfeature.Feature = "RotateKubeletClientCertificate"
-
- // owner: @msau42
- // alpha: v1.7
- //
- // A new volume type that supports local disks on a node.
- PersistentLocalVolumes utilfeature.Feature = "PersistentLocalVolumes"
-
- // owner: @jinxu
- // alpha: v1.7
- //
- // New local storage types to support local storage capacity isolation
- LocalStorageCapacityIsolation utilfeature.Feature = "LocalStorageCapacityIsolation"
-
- // owner: @gnufied
- // alpha: v1.8
- // Ability to Expand persistent volumes
- ExpandPersistentVolumes utilfeature.Feature = "ExpandPersistentVolumes"
-
- // owner: @verb
- // alpha: v1.10
- //
- // Allows running a "debug container" in a pod namespaces to troubleshoot a running pod.
- DebugContainers utilfeature.Feature = "DebugContainers"
-
- // owner: @verb
- // alpha: v1.10
- //
- // Allows all containers in a pod to share a process namespace.
- PodShareProcessNamespace utilfeature.Feature = "PodShareProcessNamespace"
-
- // owner: @bsalamat
- // alpha: v1.8
- //
- // Add priority to pods. Priority affects scheduling and preemption of pods.
- PodPriority utilfeature.Feature = "PodPriority"
-
- // owner: @resouer
- // alpha: v1.8
- //
- // Enable equivalence class cache for scheduler.
- EnableEquivalenceClassCache utilfeature.Feature = "EnableEquivalenceClassCache"
-
- // owner: @k82cn
- // alpha: v1.8
- //
- // Taint nodes based on their condition status for 'NetworkUnavailable',
- // 'MemoryPressure', 'OutOfDisk' and 'DiskPressure'.
- TaintNodesByCondition utilfeature.Feature = "TaintNodesByCondition"
-
- // owner: @jsafrane
- // beta: v1.10
- //
- // Enable mount propagation of volumes.
- MountPropagation utilfeature.Feature = "MountPropagation"
-
- // owner: @ConnorDoyle
- // alpha: v1.8
- //
- // Alternative container-level CPU affinity policies.
- CPUManager utilfeature.Feature = "CPUManager"
-
- // owner: @derekwaynecarr
- // beta: v1.10
- //
- // Enable pods to consume pre-allocated huge pages of varying page sizes
- HugePages utilfeature.Feature = "HugePages"
-
- // owner @brendandburns
- // alpha: v1.9
- //
- // Enable nodes to exclude themselves from service load balancers
- ServiceNodeExclusion utilfeature.Feature = "ServiceNodeExclusion"
-
- // owner @brendandburns
- // deprecated: v1.10
- //
- // Enable the service proxy to contact external IP addresses. Note this feature is present
- // only for backward compatibility, it will be removed in the 1.11 release.
- ServiceProxyAllowExternalIPs utilfeature.Feature = "ServiceProxyAllowExternalIPs"
-
- // owner: @jsafrane
- // alpha: v1.9
- //
- // Enable running mount utilities in containers.
- MountContainers utilfeature.Feature = "MountContainers"
-
- // owner: @msau42
- // alpha: v1.9
- //
- // Extend the default scheduler to be aware of PV topology and handle PV binding
- // Before moving to beta, resolve Kubernetes issue #56180
- VolumeScheduling utilfeature.Feature = "VolumeScheduling"
-
- // owner: @vladimirvivien
- // alpha: v1.9
- //
- // Enable mount/attachment of Container Storage Interface (CSI) backed PVs
- CSIPersistentVolume utilfeature.Feature = "CSIPersistentVolume"
-
- // owner @MrHohn
- // beta: v1.10
- //
- // Support configurable pod DNS parameters.
- CustomPodDNS utilfeature.Feature = "CustomPodDNS"
-
- // owner: @screeley44
- // alpha: v1.9
- //
- // Enable Block volume support in containers.
- BlockVolume utilfeature.Feature = "BlockVolume"
-
- // owner: @pospispa
- // beta: v1.10
- //
- // Postpone deletion of a PV or a PVC when they are being used
- StorageObjectInUseProtection utilfeature.Feature = "StorageObjectInUseProtection"
-
- // owner: @aveshagarwal
- // alpha: v1.9
- //
- // Enable resource limits priority function
- ResourceLimitsPriorityFunction utilfeature.Feature = "ResourceLimitsPriorityFunction"
-
- // owner: @m1093782566
- // beta: v1.9
- //
- // Implement IPVS-based in-cluster service load balancing
- SupportIPVSProxyMode utilfeature.Feature = "SupportIPVSProxyMode"
-
- // owner: @dims
- // alpha: v1.10
- //
- // Implement support for limiting pids in pods
- SupportPodPidsLimit utilfeature.Feature = "SupportPodPidsLimit"
-
- // owner: @feiskyer
- // alpha: v1.10
- //
- // Enable Hyper-V containers on Windows
- HyperVContainer utilfeature.Feature = "HyperVContainer"
-
- // owner: @joelsmith
- // deprecated: v1.10
- //
- // Mount secret, configMap, downwardAPI and projected volumes ReadOnly. Note: this feature
- // gate is present only for backward compatibility, it will be removed in the 1.11 release.
- ReadOnlyAPIDataVolumes utilfeature.Feature = "ReadOnlyAPIDataVolumes"
-
- // owner: @k82cn
- // alpha: v1.10
- //
- // Schedule DaemonSet Pods by default scheduler instead of DaemonSet controller
- NoDaemonSetScheduler utilfeature.Feature = "NoDaemonSetScheduler"
-
- // owner: @mikedanese
- // alpha: v1.10
- //
- // Implement TokenRequest endpoint on service account resources.
- TokenRequest utilfeature.Feature = "TokenRequest"
-
- // owner: @Random-Liu
- // alpha: v1.10
- //
- // Enable container log rotation for cri container runtime
- CRIContainerLogRotation utilfeature.Feature = "CRIContainerLogRotation"
-
- // owner: @verult
- // beta: v1.10
- //
- // Enables the regional PD feature on GCE.
- GCERegionalPersistentDisk utilfeature.Feature = "GCERegionalPersistentDisk"
-)
-
-func init() {
- utilfeature.DefaultFeatureGate.Add(defaultKubernetesFeatureGates)
-}
-
-// defaultKubernetesFeatureGates consists of all known Kubernetes-specific feature keys.
-// To add a new feature, define a key for it above and add it here. The features will be
-// available throughout Kubernetes binaries.
-var defaultKubernetesFeatureGates = map[utilfeature.Feature]utilfeature.FeatureSpec{
- AppArmor: {Default: true, PreRelease: utilfeature.Beta},
- DynamicKubeletConfig: {Default: false, PreRelease: utilfeature.Alpha},
- ExperimentalHostUserNamespaceDefaultingGate: {Default: false, PreRelease: utilfeature.Beta},
- ExperimentalCriticalPodAnnotation: {Default: false, PreRelease: utilfeature.Alpha},
- Accelerators: {Default: false, PreRelease: utilfeature.Alpha},
- DevicePlugins: {Default: true, PreRelease: utilfeature.Beta},
- TaintBasedEvictions: {Default: false, PreRelease: utilfeature.Alpha},
- RotateKubeletServerCertificate: {Default: false, PreRelease: utilfeature.Alpha},
- RotateKubeletClientCertificate: {Default: true, PreRelease: utilfeature.Beta},
- PersistentLocalVolumes: {Default: true, PreRelease: utilfeature.Beta},
- LocalStorageCapacityIsolation: {Default: false, PreRelease: utilfeature.Alpha},
- HugePages: {Default: true, PreRelease: utilfeature.Beta},
- DebugContainers: {Default: false, PreRelease: utilfeature.Alpha},
- PodShareProcessNamespace: {Default: false, PreRelease: utilfeature.Alpha},
- PodPriority: {Default: false, PreRelease: utilfeature.Alpha},
- EnableEquivalenceClassCache: {Default: false, PreRelease: utilfeature.Alpha},
- TaintNodesByCondition: {Default: false, PreRelease: utilfeature.Alpha},
- MountPropagation: {Default: true, PreRelease: utilfeature.Beta},
- ExpandPersistentVolumes: {Default: false, PreRelease: utilfeature.Alpha},
- CPUManager: {Default: true, PreRelease: utilfeature.Beta},
- ServiceNodeExclusion: {Default: false, PreRelease: utilfeature.Alpha},
- MountContainers: {Default: false, PreRelease: utilfeature.Alpha},
- VolumeScheduling: {Default: true, PreRelease: utilfeature.Beta},
- CSIPersistentVolume: {Default: true, PreRelease: utilfeature.Beta},
- CustomPodDNS: {Default: true, PreRelease: utilfeature.Beta},
- BlockVolume: {Default: false, PreRelease: utilfeature.Alpha},
- StorageObjectInUseProtection: {Default: true, PreRelease: utilfeature.Beta},
- ResourceLimitsPriorityFunction: {Default: false, PreRelease: utilfeature.Alpha},
- SupportIPVSProxyMode: {Default: true, PreRelease: utilfeature.Beta},
- SupportPodPidsLimit: {Default: false, PreRelease: utilfeature.Alpha},
- HyperVContainer: {Default: false, PreRelease: utilfeature.Alpha},
- NoDaemonSetScheduler: {Default: false, PreRelease: utilfeature.Alpha},
- TokenRequest: {Default: false, PreRelease: utilfeature.Alpha},
- CRIContainerLogRotation: {Default: false, PreRelease: utilfeature.Alpha},
- GCERegionalPersistentDisk: {Default: true, PreRelease: utilfeature.Beta},
-
- // inherited features from generic apiserver, relisted here to get a conflict if it is changed
- // unintentionally on either side:
- genericfeatures.StreamingProxyRedirects: {Default: true, PreRelease: utilfeature.Beta},
- genericfeatures.AdvancedAuditing: {Default: true, PreRelease: utilfeature.Beta},
- genericfeatures.APIResponseCompression: {Default: false, PreRelease: utilfeature.Alpha},
- genericfeatures.Initializers: {Default: false, PreRelease: utilfeature.Alpha},
- genericfeatures.APIListChunking: {Default: true, PreRelease: utilfeature.Beta},
-
- // inherited features from apiextensions-apiserver, relisted here to get a conflict if it is changed
- // unintentionally on either side:
- apiextensionsfeatures.CustomResourceValidation: {Default: true, PreRelease: utilfeature.Beta},
- apiextensionsfeatures.CustomResourceSubresources: {Default: false, PreRelease: utilfeature.Alpha},
-
- // features that enable backwards compatibility but are scheduled to be removed
- ServiceProxyAllowExternalIPs: {Default: false, PreRelease: utilfeature.Deprecated},
- ReadOnlyAPIDataVolumes: {Default: true, PreRelease: utilfeature.Deprecated},
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/fieldpath/doc.go b/vendor/k8s.io/kubernetes/pkg/fieldpath/doc.go
deleted file mode 100644
index 400d001e7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/fieldpath/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package fieldpath supplies methods for extracting fields from objects
-// given a path to a field.
-package fieldpath // import "k8s.io/kubernetes/pkg/fieldpath"
diff --git a/vendor/k8s.io/kubernetes/pkg/fieldpath/fieldpath.go b/vendor/k8s.io/kubernetes/pkg/fieldpath/fieldpath.go
deleted file mode 100644
index 437541458..000000000
--- a/vendor/k8s.io/kubernetes/pkg/fieldpath/fieldpath.go
+++ /dev/null
@@ -1,103 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package fieldpath
-
-import (
- "fmt"
- "strings"
-
- "k8s.io/apimachinery/pkg/api/meta"
- "k8s.io/apimachinery/pkg/util/validation"
-)
-
-// FormatMap formats map[string]string to a string.
-func FormatMap(m map[string]string) (fmtStr string) {
- for key, value := range m {
- fmtStr += fmt.Sprintf("%v=%q\n", key, value)
- }
- fmtStr = strings.TrimSuffix(fmtStr, "\n")
-
- return
-}
-
-// ExtractFieldPathAsString extracts the field from the given object
-// and returns it as a string. The object must be a pointer to an
-// API type.
-func ExtractFieldPathAsString(obj interface{}, fieldPath string) (string, error) {
- accessor, err := meta.Accessor(obj)
- if err != nil {
- return "", nil
- }
-
- if path, subscript, ok := SplitMaybeSubscriptedPath(fieldPath); ok {
- switch path {
- case "metadata.annotations":
- if errs := validation.IsQualifiedName(strings.ToLower(subscript)); len(errs) != 0 {
- return "", fmt.Errorf("invalid key subscript in %s: %s", fieldPath, strings.Join(errs, ";"))
- }
- return accessor.GetAnnotations()[subscript], nil
- case "metadata.labels":
- if errs := validation.IsQualifiedName(subscript); len(errs) != 0 {
- return "", fmt.Errorf("invalid key subscript in %s: %s", fieldPath, strings.Join(errs, ";"))
- }
- return accessor.GetLabels()[subscript], nil
- default:
- return "", fmt.Errorf("fieldPath %q does not support subscript", fieldPath)
- }
- }
-
- switch fieldPath {
- case "metadata.annotations":
- return FormatMap(accessor.GetAnnotations()), nil
- case "metadata.labels":
- return FormatMap(accessor.GetLabels()), nil
- case "metadata.name":
- return accessor.GetName(), nil
- case "metadata.namespace":
- return accessor.GetNamespace(), nil
- case "metadata.uid":
- return string(accessor.GetUID()), nil
- }
-
- return "", fmt.Errorf("unsupported fieldPath: %v", fieldPath)
-}
-
-// SplitMaybeSubscriptedPath checks whether the specified fieldPath is
-// subscripted, and
-// - if yes, this function splits the fieldPath into path and subscript, and
-// returns (path, subscript, true).
-// - if no, this function returns (fieldPath, "", false).
-//
-// Example inputs and outputs:
-// - "metadata.annotations['myKey']" --> ("metadata.annotations", "myKey", true)
-// - "metadata.annotations['a[b]c']" --> ("metadata.annotations", "a[b]c", true)
-// - "metadata.labels['']" --> ("metadata.labels", "", true)
-// - "metadata.labels" --> ("metadata.labels", "", false)
-func SplitMaybeSubscriptedPath(fieldPath string) (string, string, bool) {
- if !strings.HasSuffix(fieldPath, "']") {
- return fieldPath, "", false
- }
- s := strings.TrimSuffix(fieldPath, "']")
- parts := strings.SplitN(s, "['", 2)
- if len(parts) < 2 {
- return fieldPath, "", false
- }
- if len(parts[0]) == 0 {
- return fieldPath, "", false
- }
- return parts[0], parts[1], true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.pb.go b/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.pb.go
deleted file mode 100644
index 653c72603..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.pb.go
+++ /dev/null
@@ -1,26627 +0,0 @@
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by protoc-gen-gogo.
-// source: api.proto
-// DO NOT EDIT!
-
-/*
- Package v1alpha2 is a generated protocol buffer package.
-
- It is generated from these files:
- api.proto
-
- It has these top-level messages:
- VersionRequest
- VersionResponse
- DNSConfig
- PortMapping
- Mount
- NamespaceOption
- Int64Value
- LinuxSandboxSecurityContext
- LinuxPodSandboxConfig
- PodSandboxMetadata
- PodSandboxConfig
- RunPodSandboxRequest
- RunPodSandboxResponse
- StopPodSandboxRequest
- StopPodSandboxResponse
- RemovePodSandboxRequest
- RemovePodSandboxResponse
- PodSandboxStatusRequest
- PodSandboxNetworkStatus
- Namespace
- LinuxPodSandboxStatus
- PodSandboxStatus
- PodSandboxStatusResponse
- PodSandboxStateValue
- PodSandboxFilter
- ListPodSandboxRequest
- PodSandbox
- ListPodSandboxResponse
- ImageSpec
- KeyValue
- LinuxContainerResources
- SELinuxOption
- Capability
- LinuxContainerSecurityContext
- LinuxContainerConfig
- WindowsContainerConfig
- WindowsContainerResources
- ContainerMetadata
- Device
- ContainerConfig
- CreateContainerRequest
- CreateContainerResponse
- StartContainerRequest
- StartContainerResponse
- StopContainerRequest
- StopContainerResponse
- RemoveContainerRequest
- RemoveContainerResponse
- ContainerStateValue
- ContainerFilter
- ListContainersRequest
- Container
- ListContainersResponse
- ContainerStatusRequest
- ContainerStatus
- ContainerStatusResponse
- UpdateContainerResourcesRequest
- UpdateContainerResourcesResponse
- ExecSyncRequest
- ExecSyncResponse
- ExecRequest
- ExecResponse
- AttachRequest
- AttachResponse
- PortForwardRequest
- PortForwardResponse
- ImageFilter
- ListImagesRequest
- Image
- ListImagesResponse
- ImageStatusRequest
- ImageStatusResponse
- AuthConfig
- PullImageRequest
- PullImageResponse
- RemoveImageRequest
- RemoveImageResponse
- NetworkConfig
- RuntimeConfig
- UpdateRuntimeConfigRequest
- UpdateRuntimeConfigResponse
- RuntimeCondition
- RuntimeStatus
- StatusRequest
- StatusResponse
- ImageFsInfoRequest
- UInt64Value
- FilesystemIdentifier
- FilesystemUsage
- ImageFsInfoResponse
- ContainerStatsRequest
- ContainerStatsResponse
- ListContainerStatsRequest
- ContainerStatsFilter
- ListContainerStatsResponse
- ContainerAttributes
- ContainerStats
- CpuUsage
- MemoryUsage
- ReopenContainerLogRequest
- ReopenContainerLogResponse
-*/
-package v1alpha2
-
-import proto "github.com/gogo/protobuf/proto"
-import fmt "fmt"
-import math "math"
-import _ "github.com/gogo/protobuf/gogoproto"
-
-import (
- context "golang.org/x/net/context"
- grpc "google.golang.org/grpc"
-)
-
-import strings "strings"
-import reflect "reflect"
-import github_com_gogo_protobuf_sortkeys "github.com/gogo/protobuf/sortkeys"
-
-import io "io"
-
-// Reference imports to suppress errors if they are not otherwise used.
-var _ = proto.Marshal
-var _ = fmt.Errorf
-var _ = math.Inf
-
-// This is a compile-time assertion to ensure that this generated file
-// is compatible with the proto package it is being compiled against.
-// A compilation error at this line likely means your copy of the
-// proto package needs to be updated.
-const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package
-
-type Protocol int32
-
-const (
- Protocol_TCP Protocol = 0
- Protocol_UDP Protocol = 1
-)
-
-var Protocol_name = map[int32]string{
- 0: "TCP",
- 1: "UDP",
-}
-var Protocol_value = map[string]int32{
- "TCP": 0,
- "UDP": 1,
-}
-
-func (x Protocol) String() string {
- return proto.EnumName(Protocol_name, int32(x))
-}
-func (Protocol) EnumDescriptor() ([]byte, []int) { return fileDescriptorApi, []int{0} }
-
-type MountPropagation int32
-
-const (
- // No mount propagation ("private" in Linux terminology).
- MountPropagation_PROPAGATION_PRIVATE MountPropagation = 0
- // Mounts get propagated from the host to the container ("rslave" in Linux).
- MountPropagation_PROPAGATION_HOST_TO_CONTAINER MountPropagation = 1
- // Mounts get propagated from the host to the container and from the
- // container to the host ("rshared" in Linux).
- MountPropagation_PROPAGATION_BIDIRECTIONAL MountPropagation = 2
-)
-
-var MountPropagation_name = map[int32]string{
- 0: "PROPAGATION_PRIVATE",
- 1: "PROPAGATION_HOST_TO_CONTAINER",
- 2: "PROPAGATION_BIDIRECTIONAL",
-}
-var MountPropagation_value = map[string]int32{
- "PROPAGATION_PRIVATE": 0,
- "PROPAGATION_HOST_TO_CONTAINER": 1,
- "PROPAGATION_BIDIRECTIONAL": 2,
-}
-
-func (x MountPropagation) String() string {
- return proto.EnumName(MountPropagation_name, int32(x))
-}
-func (MountPropagation) EnumDescriptor() ([]byte, []int) { return fileDescriptorApi, []int{1} }
-
-// A NamespaceMode describes the intended namespace configuration for each
-// of the namespaces (Network, PID, IPC) in NamespaceOption. Runtimes should
-// map these modes as appropriate for the technology underlying the runtime.
-type NamespaceMode int32
-
-const (
- // A POD namespace is common to all containers in a pod.
- // For example, a container with a PID namespace of POD expects to view
- // all of the processes in all of the containers in the pod.
- NamespaceMode_POD NamespaceMode = 0
- // A CONTAINER namespace is restricted to a single container.
- // For example, a container with a PID namespace of CONTAINER expects to
- // view only the processes in that container.
- NamespaceMode_CONTAINER NamespaceMode = 1
- // A NODE namespace is the namespace of the Kubernetes node.
- // For example, a container with a PID namespace of NODE expects to view
- // all of the processes on the host running the kubelet.
- NamespaceMode_NODE NamespaceMode = 2
-)
-
-var NamespaceMode_name = map[int32]string{
- 0: "POD",
- 1: "CONTAINER",
- 2: "NODE",
-}
-var NamespaceMode_value = map[string]int32{
- "POD": 0,
- "CONTAINER": 1,
- "NODE": 2,
-}
-
-func (x NamespaceMode) String() string {
- return proto.EnumName(NamespaceMode_name, int32(x))
-}
-func (NamespaceMode) EnumDescriptor() ([]byte, []int) { return fileDescriptorApi, []int{2} }
-
-type PodSandboxState int32
-
-const (
- PodSandboxState_SANDBOX_READY PodSandboxState = 0
- PodSandboxState_SANDBOX_NOTREADY PodSandboxState = 1
-)
-
-var PodSandboxState_name = map[int32]string{
- 0: "SANDBOX_READY",
- 1: "SANDBOX_NOTREADY",
-}
-var PodSandboxState_value = map[string]int32{
- "SANDBOX_READY": 0,
- "SANDBOX_NOTREADY": 1,
-}
-
-func (x PodSandboxState) String() string {
- return proto.EnumName(PodSandboxState_name, int32(x))
-}
-func (PodSandboxState) EnumDescriptor() ([]byte, []int) { return fileDescriptorApi, []int{3} }
-
-type ContainerState int32
-
-const (
- ContainerState_CONTAINER_CREATED ContainerState = 0
- ContainerState_CONTAINER_RUNNING ContainerState = 1
- ContainerState_CONTAINER_EXITED ContainerState = 2
- ContainerState_CONTAINER_UNKNOWN ContainerState = 3
-)
-
-var ContainerState_name = map[int32]string{
- 0: "CONTAINER_CREATED",
- 1: "CONTAINER_RUNNING",
- 2: "CONTAINER_EXITED",
- 3: "CONTAINER_UNKNOWN",
-}
-var ContainerState_value = map[string]int32{
- "CONTAINER_CREATED": 0,
- "CONTAINER_RUNNING": 1,
- "CONTAINER_EXITED": 2,
- "CONTAINER_UNKNOWN": 3,
-}
-
-func (x ContainerState) String() string {
- return proto.EnumName(ContainerState_name, int32(x))
-}
-func (ContainerState) EnumDescriptor() ([]byte, []int) { return fileDescriptorApi, []int{4} }
-
-type VersionRequest struct {
- // Version of the kubelet runtime API.
- Version string `protobuf:"bytes,1,opt,name=version,proto3" json:"version,omitempty"`
-}
-
-func (m *VersionRequest) Reset() { *m = VersionRequest{} }
-func (*VersionRequest) ProtoMessage() {}
-func (*VersionRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{0} }
-
-func (m *VersionRequest) GetVersion() string {
- if m != nil {
- return m.Version
- }
- return ""
-}
-
-type VersionResponse struct {
- // Version of the kubelet runtime API.
- Version string `protobuf:"bytes,1,opt,name=version,proto3" json:"version,omitempty"`
- // Name of the container runtime.
- RuntimeName string `protobuf:"bytes,2,opt,name=runtime_name,json=runtimeName,proto3" json:"runtime_name,omitempty"`
- // Version of the container runtime. The string must be
- // semver-compatible.
- RuntimeVersion string `protobuf:"bytes,3,opt,name=runtime_version,json=runtimeVersion,proto3" json:"runtime_version,omitempty"`
- // API version of the container runtime. The string must be
- // semver-compatible.
- RuntimeApiVersion string `protobuf:"bytes,4,opt,name=runtime_api_version,json=runtimeApiVersion,proto3" json:"runtime_api_version,omitempty"`
-}
-
-func (m *VersionResponse) Reset() { *m = VersionResponse{} }
-func (*VersionResponse) ProtoMessage() {}
-func (*VersionResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{1} }
-
-func (m *VersionResponse) GetVersion() string {
- if m != nil {
- return m.Version
- }
- return ""
-}
-
-func (m *VersionResponse) GetRuntimeName() string {
- if m != nil {
- return m.RuntimeName
- }
- return ""
-}
-
-func (m *VersionResponse) GetRuntimeVersion() string {
- if m != nil {
- return m.RuntimeVersion
- }
- return ""
-}
-
-func (m *VersionResponse) GetRuntimeApiVersion() string {
- if m != nil {
- return m.RuntimeApiVersion
- }
- return ""
-}
-
-// DNSConfig specifies the DNS servers and search domains of a sandbox.
-type DNSConfig struct {
- // List of DNS servers of the cluster.
- Servers []string `protobuf:"bytes,1,rep,name=servers" json:"servers,omitempty"`
- // List of DNS search domains of the cluster.
- Searches []string `protobuf:"bytes,2,rep,name=searches" json:"searches,omitempty"`
- // List of DNS options. See https://linux.die.net/man/5/resolv.conf
- // for all available options.
- Options []string `protobuf:"bytes,3,rep,name=options" json:"options,omitempty"`
-}
-
-func (m *DNSConfig) Reset() { *m = DNSConfig{} }
-func (*DNSConfig) ProtoMessage() {}
-func (*DNSConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{2} }
-
-func (m *DNSConfig) GetServers() []string {
- if m != nil {
- return m.Servers
- }
- return nil
-}
-
-func (m *DNSConfig) GetSearches() []string {
- if m != nil {
- return m.Searches
- }
- return nil
-}
-
-func (m *DNSConfig) GetOptions() []string {
- if m != nil {
- return m.Options
- }
- return nil
-}
-
-// PortMapping specifies the port mapping configurations of a sandbox.
-type PortMapping struct {
- // Protocol of the port mapping.
- Protocol Protocol `protobuf:"varint,1,opt,name=protocol,proto3,enum=runtime.v1alpha2.Protocol" json:"protocol,omitempty"`
- // Port number within the container. Default: 0 (not specified).
- ContainerPort int32 `protobuf:"varint,2,opt,name=container_port,json=containerPort,proto3" json:"container_port,omitempty"`
- // Port number on the host. Default: 0 (not specified).
- HostPort int32 `protobuf:"varint,3,opt,name=host_port,json=hostPort,proto3" json:"host_port,omitempty"`
- // Host IP.
- HostIp string `protobuf:"bytes,4,opt,name=host_ip,json=hostIp,proto3" json:"host_ip,omitempty"`
-}
-
-func (m *PortMapping) Reset() { *m = PortMapping{} }
-func (*PortMapping) ProtoMessage() {}
-func (*PortMapping) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{3} }
-
-func (m *PortMapping) GetProtocol() Protocol {
- if m != nil {
- return m.Protocol
- }
- return Protocol_TCP
-}
-
-func (m *PortMapping) GetContainerPort() int32 {
- if m != nil {
- return m.ContainerPort
- }
- return 0
-}
-
-func (m *PortMapping) GetHostPort() int32 {
- if m != nil {
- return m.HostPort
- }
- return 0
-}
-
-func (m *PortMapping) GetHostIp() string {
- if m != nil {
- return m.HostIp
- }
- return ""
-}
-
-// Mount specifies a host volume to mount into a container.
-type Mount struct {
- // Path of the mount within the container.
- ContainerPath string `protobuf:"bytes,1,opt,name=container_path,json=containerPath,proto3" json:"container_path,omitempty"`
- // Path of the mount on the host.
- HostPath string `protobuf:"bytes,2,opt,name=host_path,json=hostPath,proto3" json:"host_path,omitempty"`
- // If set, the mount is read-only.
- Readonly bool `protobuf:"varint,3,opt,name=readonly,proto3" json:"readonly,omitempty"`
- // If set, the mount needs SELinux relabeling.
- SelinuxRelabel bool `protobuf:"varint,4,opt,name=selinux_relabel,json=selinuxRelabel,proto3" json:"selinux_relabel,omitempty"`
- // Requested propagation mode.
- Propagation MountPropagation `protobuf:"varint,5,opt,name=propagation,proto3,enum=runtime.v1alpha2.MountPropagation" json:"propagation,omitempty"`
-}
-
-func (m *Mount) Reset() { *m = Mount{} }
-func (*Mount) ProtoMessage() {}
-func (*Mount) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{4} }
-
-func (m *Mount) GetContainerPath() string {
- if m != nil {
- return m.ContainerPath
- }
- return ""
-}
-
-func (m *Mount) GetHostPath() string {
- if m != nil {
- return m.HostPath
- }
- return ""
-}
-
-func (m *Mount) GetReadonly() bool {
- if m != nil {
- return m.Readonly
- }
- return false
-}
-
-func (m *Mount) GetSelinuxRelabel() bool {
- if m != nil {
- return m.SelinuxRelabel
- }
- return false
-}
-
-func (m *Mount) GetPropagation() MountPropagation {
- if m != nil {
- return m.Propagation
- }
- return MountPropagation_PROPAGATION_PRIVATE
-}
-
-// NamespaceOption provides options for Linux namespaces.
-type NamespaceOption struct {
- // Network namespace for this container/sandbox.
- // Note: There is currently no way to set CONTAINER scoped network in the Kubernetes API.
- // Namespaces currently set by the kubelet: POD, NODE
- Network NamespaceMode `protobuf:"varint,1,opt,name=network,proto3,enum=runtime.v1alpha2.NamespaceMode" json:"network,omitempty"`
- // PID namespace for this container/sandbox.
- // Note: The CRI default is POD, but the v1.PodSpec default is CONTAINER.
- // The kubelet's runtime manager will set this to CONTAINER explicitly for v1 pods.
- // Namespaces currently set by the kubelet: POD, CONTAINER, NODE
- Pid NamespaceMode `protobuf:"varint,2,opt,name=pid,proto3,enum=runtime.v1alpha2.NamespaceMode" json:"pid,omitempty"`
- // IPC namespace for this container/sandbox.
- // Note: There is currently no way to set CONTAINER scoped IPC in the Kubernetes API.
- // Namespaces currently set by the kubelet: POD, NODE
- Ipc NamespaceMode `protobuf:"varint,3,opt,name=ipc,proto3,enum=runtime.v1alpha2.NamespaceMode" json:"ipc,omitempty"`
-}
-
-func (m *NamespaceOption) Reset() { *m = NamespaceOption{} }
-func (*NamespaceOption) ProtoMessage() {}
-func (*NamespaceOption) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{5} }
-
-func (m *NamespaceOption) GetNetwork() NamespaceMode {
- if m != nil {
- return m.Network
- }
- return NamespaceMode_POD
-}
-
-func (m *NamespaceOption) GetPid() NamespaceMode {
- if m != nil {
- return m.Pid
- }
- return NamespaceMode_POD
-}
-
-func (m *NamespaceOption) GetIpc() NamespaceMode {
- if m != nil {
- return m.Ipc
- }
- return NamespaceMode_POD
-}
-
-// Int64Value is the wrapper of int64.
-type Int64Value struct {
- // The value.
- Value int64 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"`
-}
-
-func (m *Int64Value) Reset() { *m = Int64Value{} }
-func (*Int64Value) ProtoMessage() {}
-func (*Int64Value) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{6} }
-
-func (m *Int64Value) GetValue() int64 {
- if m != nil {
- return m.Value
- }
- return 0
-}
-
-// LinuxSandboxSecurityContext holds linux security configuration that will be
-// applied to a sandbox. Note that:
-// 1) It does not apply to containers in the pods.
-// 2) It may not be applicable to a PodSandbox which does not contain any running
-// process.
-type LinuxSandboxSecurityContext struct {
- // Configurations for the sandbox's namespaces.
- // This will be used only if the PodSandbox uses namespace for isolation.
- NamespaceOptions *NamespaceOption `protobuf:"bytes,1,opt,name=namespace_options,json=namespaceOptions" json:"namespace_options,omitempty"`
- // Optional SELinux context to be applied.
- SelinuxOptions *SELinuxOption `protobuf:"bytes,2,opt,name=selinux_options,json=selinuxOptions" json:"selinux_options,omitempty"`
- // UID to run sandbox processes as, when applicable.
- RunAsUser *Int64Value `protobuf:"bytes,3,opt,name=run_as_user,json=runAsUser" json:"run_as_user,omitempty"`
- // If set, the root filesystem of the sandbox is read-only.
- ReadonlyRootfs bool `protobuf:"varint,4,opt,name=readonly_rootfs,json=readonlyRootfs,proto3" json:"readonly_rootfs,omitempty"`
- // List of groups applied to the first process run in the sandbox, in
- // addition to the sandbox's primary GID.
- SupplementalGroups []int64 `protobuf:"varint,5,rep,packed,name=supplemental_groups,json=supplementalGroups" json:"supplemental_groups,omitempty"`
- // Indicates whether the sandbox will be asked to run a privileged
- // container. If a privileged container is to be executed within it, this
- // MUST be true.
- // This allows a sandbox to take additional security precautions if no
- // privileged containers are expected to be run.
- Privileged bool `protobuf:"varint,6,opt,name=privileged,proto3" json:"privileged,omitempty"`
- // Seccomp profile for the sandbox, candidate values are:
- // * docker/default: the default profile for the docker container runtime
- // * unconfined: unconfined profile, ie, no seccomp sandboxing
- // * localhost/<full-path-to-profile>: the profile installed on the node.
- // <full-path-to-profile> is the full path of the profile.
- // Default: "", which is identical with unconfined.
- SeccompProfilePath string `protobuf:"bytes,7,opt,name=seccomp_profile_path,json=seccompProfilePath,proto3" json:"seccomp_profile_path,omitempty"`
-}
-
-func (m *LinuxSandboxSecurityContext) Reset() { *m = LinuxSandboxSecurityContext{} }
-func (*LinuxSandboxSecurityContext) ProtoMessage() {}
-func (*LinuxSandboxSecurityContext) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{7} }
-
-func (m *LinuxSandboxSecurityContext) GetNamespaceOptions() *NamespaceOption {
- if m != nil {
- return m.NamespaceOptions
- }
- return nil
-}
-
-func (m *LinuxSandboxSecurityContext) GetSelinuxOptions() *SELinuxOption {
- if m != nil {
- return m.SelinuxOptions
- }
- return nil
-}
-
-func (m *LinuxSandboxSecurityContext) GetRunAsUser() *Int64Value {
- if m != nil {
- return m.RunAsUser
- }
- return nil
-}
-
-func (m *LinuxSandboxSecurityContext) GetReadonlyRootfs() bool {
- if m != nil {
- return m.ReadonlyRootfs
- }
- return false
-}
-
-func (m *LinuxSandboxSecurityContext) GetSupplementalGroups() []int64 {
- if m != nil {
- return m.SupplementalGroups
- }
- return nil
-}
-
-func (m *LinuxSandboxSecurityContext) GetPrivileged() bool {
- if m != nil {
- return m.Privileged
- }
- return false
-}
-
-func (m *LinuxSandboxSecurityContext) GetSeccompProfilePath() string {
- if m != nil {
- return m.SeccompProfilePath
- }
- return ""
-}
-
-// LinuxPodSandboxConfig holds platform-specific configurations for Linux
-// host platforms and Linux-based containers.
-type LinuxPodSandboxConfig struct {
- // Parent cgroup of the PodSandbox.
- // The cgroupfs style syntax will be used, but the container runtime can
- // convert it to systemd semantics if needed.
- CgroupParent string `protobuf:"bytes,1,opt,name=cgroup_parent,json=cgroupParent,proto3" json:"cgroup_parent,omitempty"`
- // LinuxSandboxSecurityContext holds sandbox security attributes.
- SecurityContext *LinuxSandboxSecurityContext `protobuf:"bytes,2,opt,name=security_context,json=securityContext" json:"security_context,omitempty"`
- // Sysctls holds linux sysctls config for the sandbox.
- Sysctls map[string]string `protobuf:"bytes,3,rep,name=sysctls" json:"sysctls,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *LinuxPodSandboxConfig) Reset() { *m = LinuxPodSandboxConfig{} }
-func (*LinuxPodSandboxConfig) ProtoMessage() {}
-func (*LinuxPodSandboxConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{8} }
-
-func (m *LinuxPodSandboxConfig) GetCgroupParent() string {
- if m != nil {
- return m.CgroupParent
- }
- return ""
-}
-
-func (m *LinuxPodSandboxConfig) GetSecurityContext() *LinuxSandboxSecurityContext {
- if m != nil {
- return m.SecurityContext
- }
- return nil
-}
-
-func (m *LinuxPodSandboxConfig) GetSysctls() map[string]string {
- if m != nil {
- return m.Sysctls
- }
- return nil
-}
-
-// PodSandboxMetadata holds all necessary information for building the sandbox name.
-// The container runtime is encouraged to expose the metadata associated with the
-// PodSandbox in its user interface for better user experience. For example,
-// the runtime can construct a unique PodSandboxName based on the metadata.
-type PodSandboxMetadata struct {
- // Pod name of the sandbox. Same as the pod name in the PodSpec.
- Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
- // Pod UID of the sandbox. Same as the pod UID in the PodSpec.
- Uid string `protobuf:"bytes,2,opt,name=uid,proto3" json:"uid,omitempty"`
- // Pod namespace of the sandbox. Same as the pod namespace in the PodSpec.
- Namespace string `protobuf:"bytes,3,opt,name=namespace,proto3" json:"namespace,omitempty"`
- // Attempt number of creating the sandbox. Default: 0.
- Attempt uint32 `protobuf:"varint,4,opt,name=attempt,proto3" json:"attempt,omitempty"`
-}
-
-func (m *PodSandboxMetadata) Reset() { *m = PodSandboxMetadata{} }
-func (*PodSandboxMetadata) ProtoMessage() {}
-func (*PodSandboxMetadata) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{9} }
-
-func (m *PodSandboxMetadata) GetName() string {
- if m != nil {
- return m.Name
- }
- return ""
-}
-
-func (m *PodSandboxMetadata) GetUid() string {
- if m != nil {
- return m.Uid
- }
- return ""
-}
-
-func (m *PodSandboxMetadata) GetNamespace() string {
- if m != nil {
- return m.Namespace
- }
- return ""
-}
-
-func (m *PodSandboxMetadata) GetAttempt() uint32 {
- if m != nil {
- return m.Attempt
- }
- return 0
-}
-
-// PodSandboxConfig holds all the required and optional fields for creating a
-// sandbox.
-type PodSandboxConfig struct {
- // Metadata of the sandbox. This information will uniquely identify the
- // sandbox, and the runtime should leverage this to ensure correct
- // operation. The runtime may also use this information to improve UX, such
- // as by constructing a readable name.
- Metadata *PodSandboxMetadata `protobuf:"bytes,1,opt,name=metadata" json:"metadata,omitempty"`
- // Hostname of the sandbox.
- Hostname string `protobuf:"bytes,2,opt,name=hostname,proto3" json:"hostname,omitempty"`
- // Path to the directory on the host in which container log files are
- // stored.
- // By default the log of a container going into the LogDirectory will be
- // hooked up to STDOUT and STDERR. However, the LogDirectory may contain
- // binary log files with structured logging data from the individual
- // containers. For example, the files might be newline separated JSON
- // structured logs, systemd-journald journal files, gRPC trace files, etc.
- // E.g.,
- // PodSandboxConfig.LogDirectory = `/var/log/pods/<podUID>/`
- // ContainerConfig.LogPath = `containerName_Instance#.log`
- //
- // WARNING: Log management and how kubelet should interface with the
- // container logs are under active discussion in
- // https://issues.k8s.io/24677. There *may* be future change of direction
- // for logging as the discussion carries on.
- LogDirectory string `protobuf:"bytes,3,opt,name=log_directory,json=logDirectory,proto3" json:"log_directory,omitempty"`
- // DNS config for the sandbox.
- DnsConfig *DNSConfig `protobuf:"bytes,4,opt,name=dns_config,json=dnsConfig" json:"dns_config,omitempty"`
- // Port mappings for the sandbox.
- PortMappings []*PortMapping `protobuf:"bytes,5,rep,name=port_mappings,json=portMappings" json:"port_mappings,omitempty"`
- // Key-value pairs that may be used to scope and select individual resources.
- Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map that may be set by the kubelet to store and
- // retrieve arbitrary metadata. This will include any annotations set on a
- // pod through the Kubernetes API.
- //
- // Annotations MUST NOT be altered by the runtime; the annotations stored
- // here MUST be returned in the PodSandboxStatus associated with the pod
- // this PodSandboxConfig creates.
- //
- // In general, in order to preserve a well-defined interface between the
- // kubelet and the container runtime, annotations SHOULD NOT influence
- // runtime behaviour.
- //
- // Annotations can also be useful for runtime authors to experiment with
- // new features that are opaque to the Kubernetes APIs (both user-facing
- // and the CRI). Whenever possible, however, runtime authors SHOULD
- // consider proposing new typed fields for any new features instead.
- Annotations map[string]string `protobuf:"bytes,7,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Optional configurations specific to Linux hosts.
- Linux *LinuxPodSandboxConfig `protobuf:"bytes,8,opt,name=linux" json:"linux,omitempty"`
-}
-
-func (m *PodSandboxConfig) Reset() { *m = PodSandboxConfig{} }
-func (*PodSandboxConfig) ProtoMessage() {}
-func (*PodSandboxConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{10} }
-
-func (m *PodSandboxConfig) GetMetadata() *PodSandboxMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *PodSandboxConfig) GetHostname() string {
- if m != nil {
- return m.Hostname
- }
- return ""
-}
-
-func (m *PodSandboxConfig) GetLogDirectory() string {
- if m != nil {
- return m.LogDirectory
- }
- return ""
-}
-
-func (m *PodSandboxConfig) GetDnsConfig() *DNSConfig {
- if m != nil {
- return m.DnsConfig
- }
- return nil
-}
-
-func (m *PodSandboxConfig) GetPortMappings() []*PortMapping {
- if m != nil {
- return m.PortMappings
- }
- return nil
-}
-
-func (m *PodSandboxConfig) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *PodSandboxConfig) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-func (m *PodSandboxConfig) GetLinux() *LinuxPodSandboxConfig {
- if m != nil {
- return m.Linux
- }
- return nil
-}
-
-type RunPodSandboxRequest struct {
- // Configuration for creating a PodSandbox.
- Config *PodSandboxConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"`
-}
-
-func (m *RunPodSandboxRequest) Reset() { *m = RunPodSandboxRequest{} }
-func (*RunPodSandboxRequest) ProtoMessage() {}
-func (*RunPodSandboxRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{11} }
-
-func (m *RunPodSandboxRequest) GetConfig() *PodSandboxConfig {
- if m != nil {
- return m.Config
- }
- return nil
-}
-
-type RunPodSandboxResponse struct {
- // ID of the PodSandbox to run.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
-}
-
-func (m *RunPodSandboxResponse) Reset() { *m = RunPodSandboxResponse{} }
-func (*RunPodSandboxResponse) ProtoMessage() {}
-func (*RunPodSandboxResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{12} }
-
-func (m *RunPodSandboxResponse) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-type StopPodSandboxRequest struct {
- // ID of the PodSandbox to stop.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
-}
-
-func (m *StopPodSandboxRequest) Reset() { *m = StopPodSandboxRequest{} }
-func (*StopPodSandboxRequest) ProtoMessage() {}
-func (*StopPodSandboxRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{13} }
-
-func (m *StopPodSandboxRequest) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-type StopPodSandboxResponse struct {
-}
-
-func (m *StopPodSandboxResponse) Reset() { *m = StopPodSandboxResponse{} }
-func (*StopPodSandboxResponse) ProtoMessage() {}
-func (*StopPodSandboxResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{14} }
-
-type RemovePodSandboxRequest struct {
- // ID of the PodSandbox to remove.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
-}
-
-func (m *RemovePodSandboxRequest) Reset() { *m = RemovePodSandboxRequest{} }
-func (*RemovePodSandboxRequest) ProtoMessage() {}
-func (*RemovePodSandboxRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{15} }
-
-func (m *RemovePodSandboxRequest) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-type RemovePodSandboxResponse struct {
-}
-
-func (m *RemovePodSandboxResponse) Reset() { *m = RemovePodSandboxResponse{} }
-func (*RemovePodSandboxResponse) ProtoMessage() {}
-func (*RemovePodSandboxResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{16} }
-
-type PodSandboxStatusRequest struct {
- // ID of the PodSandbox for which to retrieve status.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // Verbose indicates whether to return extra information about the pod sandbox.
- Verbose bool `protobuf:"varint,2,opt,name=verbose,proto3" json:"verbose,omitempty"`
-}
-
-func (m *PodSandboxStatusRequest) Reset() { *m = PodSandboxStatusRequest{} }
-func (*PodSandboxStatusRequest) ProtoMessage() {}
-func (*PodSandboxStatusRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{17} }
-
-func (m *PodSandboxStatusRequest) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *PodSandboxStatusRequest) GetVerbose() bool {
- if m != nil {
- return m.Verbose
- }
- return false
-}
-
-// PodSandboxNetworkStatus is the status of the network for a PodSandbox.
-type PodSandboxNetworkStatus struct {
- // IP address of the PodSandbox.
- Ip string `protobuf:"bytes,1,opt,name=ip,proto3" json:"ip,omitempty"`
-}
-
-func (m *PodSandboxNetworkStatus) Reset() { *m = PodSandboxNetworkStatus{} }
-func (*PodSandboxNetworkStatus) ProtoMessage() {}
-func (*PodSandboxNetworkStatus) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{18} }
-
-func (m *PodSandboxNetworkStatus) GetIp() string {
- if m != nil {
- return m.Ip
- }
- return ""
-}
-
-// Namespace contains paths to the namespaces.
-type Namespace struct {
- // Namespace options for Linux namespaces.
- Options *NamespaceOption `protobuf:"bytes,2,opt,name=options" json:"options,omitempty"`
-}
-
-func (m *Namespace) Reset() { *m = Namespace{} }
-func (*Namespace) ProtoMessage() {}
-func (*Namespace) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{19} }
-
-func (m *Namespace) GetOptions() *NamespaceOption {
- if m != nil {
- return m.Options
- }
- return nil
-}
-
-// LinuxSandboxStatus contains status specific to Linux sandboxes.
-type LinuxPodSandboxStatus struct {
- // Paths to the sandbox's namespaces.
- Namespaces *Namespace `protobuf:"bytes,1,opt,name=namespaces" json:"namespaces,omitempty"`
-}
-
-func (m *LinuxPodSandboxStatus) Reset() { *m = LinuxPodSandboxStatus{} }
-func (*LinuxPodSandboxStatus) ProtoMessage() {}
-func (*LinuxPodSandboxStatus) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{20} }
-
-func (m *LinuxPodSandboxStatus) GetNamespaces() *Namespace {
- if m != nil {
- return m.Namespaces
- }
- return nil
-}
-
-// PodSandboxStatus contains the status of the PodSandbox.
-type PodSandboxStatus struct {
- // ID of the sandbox.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // Metadata of the sandbox.
- Metadata *PodSandboxMetadata `protobuf:"bytes,2,opt,name=metadata" json:"metadata,omitempty"`
- // State of the sandbox.
- State PodSandboxState `protobuf:"varint,3,opt,name=state,proto3,enum=runtime.v1alpha2.PodSandboxState" json:"state,omitempty"`
- // Creation timestamp of the sandbox in nanoseconds. Must be > 0.
- CreatedAt int64 `protobuf:"varint,4,opt,name=created_at,json=createdAt,proto3" json:"created_at,omitempty"`
- // Network contains network status if network is handled by the runtime.
- Network *PodSandboxNetworkStatus `protobuf:"bytes,5,opt,name=network" json:"network,omitempty"`
- // Linux-specific status to a pod sandbox.
- Linux *LinuxPodSandboxStatus `protobuf:"bytes,6,opt,name=linux" json:"linux,omitempty"`
- // Labels are key-value pairs that may be used to scope and select individual resources.
- Labels map[string]string `protobuf:"bytes,7,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding PodSandboxConfig used to
- // instantiate the pod sandbox this status represents.
- Annotations map[string]string `protobuf:"bytes,8,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *PodSandboxStatus) Reset() { *m = PodSandboxStatus{} }
-func (*PodSandboxStatus) ProtoMessage() {}
-func (*PodSandboxStatus) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{21} }
-
-func (m *PodSandboxStatus) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *PodSandboxStatus) GetMetadata() *PodSandboxMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *PodSandboxStatus) GetState() PodSandboxState {
- if m != nil {
- return m.State
- }
- return PodSandboxState_SANDBOX_READY
-}
-
-func (m *PodSandboxStatus) GetCreatedAt() int64 {
- if m != nil {
- return m.CreatedAt
- }
- return 0
-}
-
-func (m *PodSandboxStatus) GetNetwork() *PodSandboxNetworkStatus {
- if m != nil {
- return m.Network
- }
- return nil
-}
-
-func (m *PodSandboxStatus) GetLinux() *LinuxPodSandboxStatus {
- if m != nil {
- return m.Linux
- }
- return nil
-}
-
-func (m *PodSandboxStatus) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *PodSandboxStatus) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-type PodSandboxStatusResponse struct {
- // Status of the PodSandbox.
- Status *PodSandboxStatus `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"`
- // Info is extra information of the PodSandbox. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. network namespace for linux container based container runtime.
- // It should only be returned non-empty when Verbose is true.
- Info map[string]string `protobuf:"bytes,2,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *PodSandboxStatusResponse) Reset() { *m = PodSandboxStatusResponse{} }
-func (*PodSandboxStatusResponse) ProtoMessage() {}
-func (*PodSandboxStatusResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{22} }
-
-func (m *PodSandboxStatusResponse) GetStatus() *PodSandboxStatus {
- if m != nil {
- return m.Status
- }
- return nil
-}
-
-func (m *PodSandboxStatusResponse) GetInfo() map[string]string {
- if m != nil {
- return m.Info
- }
- return nil
-}
-
-// PodSandboxStateValue is the wrapper of PodSandboxState.
-type PodSandboxStateValue struct {
- // State of the sandbox.
- State PodSandboxState `protobuf:"varint,1,opt,name=state,proto3,enum=runtime.v1alpha2.PodSandboxState" json:"state,omitempty"`
-}
-
-func (m *PodSandboxStateValue) Reset() { *m = PodSandboxStateValue{} }
-func (*PodSandboxStateValue) ProtoMessage() {}
-func (*PodSandboxStateValue) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{23} }
-
-func (m *PodSandboxStateValue) GetState() PodSandboxState {
- if m != nil {
- return m.State
- }
- return PodSandboxState_SANDBOX_READY
-}
-
-// PodSandboxFilter is used to filter a list of PodSandboxes.
-// All those fields are combined with 'AND'
-type PodSandboxFilter struct {
- // ID of the sandbox.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // State of the sandbox.
- State *PodSandboxStateValue `protobuf:"bytes,2,opt,name=state" json:"state,omitempty"`
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- LabelSelector map[string]string `protobuf:"bytes,3,rep,name=label_selector,json=labelSelector" json:"label_selector,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *PodSandboxFilter) Reset() { *m = PodSandboxFilter{} }
-func (*PodSandboxFilter) ProtoMessage() {}
-func (*PodSandboxFilter) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{24} }
-
-func (m *PodSandboxFilter) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *PodSandboxFilter) GetState() *PodSandboxStateValue {
- if m != nil {
- return m.State
- }
- return nil
-}
-
-func (m *PodSandboxFilter) GetLabelSelector() map[string]string {
- if m != nil {
- return m.LabelSelector
- }
- return nil
-}
-
-type ListPodSandboxRequest struct {
- // PodSandboxFilter to filter a list of PodSandboxes.
- Filter *PodSandboxFilter `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"`
-}
-
-func (m *ListPodSandboxRequest) Reset() { *m = ListPodSandboxRequest{} }
-func (*ListPodSandboxRequest) ProtoMessage() {}
-func (*ListPodSandboxRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{25} }
-
-func (m *ListPodSandboxRequest) GetFilter() *PodSandboxFilter {
- if m != nil {
- return m.Filter
- }
- return nil
-}
-
-// PodSandbox contains minimal information about a sandbox.
-type PodSandbox struct {
- // ID of the PodSandbox.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // Metadata of the PodSandbox.
- Metadata *PodSandboxMetadata `protobuf:"bytes,2,opt,name=metadata" json:"metadata,omitempty"`
- // State of the PodSandbox.
- State PodSandboxState `protobuf:"varint,3,opt,name=state,proto3,enum=runtime.v1alpha2.PodSandboxState" json:"state,omitempty"`
- // Creation timestamps of the PodSandbox in nanoseconds. Must be > 0.
- CreatedAt int64 `protobuf:"varint,4,opt,name=created_at,json=createdAt,proto3" json:"created_at,omitempty"`
- // Labels of the PodSandbox.
- Labels map[string]string `protobuf:"bytes,5,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding PodSandboxConfig used to
- // instantiate this PodSandbox.
- Annotations map[string]string `protobuf:"bytes,6,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *PodSandbox) Reset() { *m = PodSandbox{} }
-func (*PodSandbox) ProtoMessage() {}
-func (*PodSandbox) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{26} }
-
-func (m *PodSandbox) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *PodSandbox) GetMetadata() *PodSandboxMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *PodSandbox) GetState() PodSandboxState {
- if m != nil {
- return m.State
- }
- return PodSandboxState_SANDBOX_READY
-}
-
-func (m *PodSandbox) GetCreatedAt() int64 {
- if m != nil {
- return m.CreatedAt
- }
- return 0
-}
-
-func (m *PodSandbox) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *PodSandbox) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-type ListPodSandboxResponse struct {
- // List of PodSandboxes.
- Items []*PodSandbox `protobuf:"bytes,1,rep,name=items" json:"items,omitempty"`
-}
-
-func (m *ListPodSandboxResponse) Reset() { *m = ListPodSandboxResponse{} }
-func (*ListPodSandboxResponse) ProtoMessage() {}
-func (*ListPodSandboxResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{27} }
-
-func (m *ListPodSandboxResponse) GetItems() []*PodSandbox {
- if m != nil {
- return m.Items
- }
- return nil
-}
-
-// ImageSpec is an internal representation of an image. Currently, it wraps the
-// value of a Container's Image field (e.g. imageID or imageDigest), but in the
-// future it will include more detailed information about the different image types.
-type ImageSpec struct {
- Image string `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"`
-}
-
-func (m *ImageSpec) Reset() { *m = ImageSpec{} }
-func (*ImageSpec) ProtoMessage() {}
-func (*ImageSpec) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{28} }
-
-func (m *ImageSpec) GetImage() string {
- if m != nil {
- return m.Image
- }
- return ""
-}
-
-type KeyValue struct {
- Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"`
- Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"`
-}
-
-func (m *KeyValue) Reset() { *m = KeyValue{} }
-func (*KeyValue) ProtoMessage() {}
-func (*KeyValue) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{29} }
-
-func (m *KeyValue) GetKey() string {
- if m != nil {
- return m.Key
- }
- return ""
-}
-
-func (m *KeyValue) GetValue() string {
- if m != nil {
- return m.Value
- }
- return ""
-}
-
-// LinuxContainerResources specifies Linux specific configuration for
-// resources.
-// TODO: Consider using Resources from opencontainers/runtime-spec/specs-go
-// directly.
-type LinuxContainerResources struct {
- // CPU CFS (Completely Fair Scheduler) period. Default: 0 (not specified).
- CpuPeriod int64 `protobuf:"varint,1,opt,name=cpu_period,json=cpuPeriod,proto3" json:"cpu_period,omitempty"`
- // CPU CFS (Completely Fair Scheduler) quota. Default: 0 (not specified).
- CpuQuota int64 `protobuf:"varint,2,opt,name=cpu_quota,json=cpuQuota,proto3" json:"cpu_quota,omitempty"`
- // CPU shares (relative weight vs. other containers). Default: 0 (not specified).
- CpuShares int64 `protobuf:"varint,3,opt,name=cpu_shares,json=cpuShares,proto3" json:"cpu_shares,omitempty"`
- // Memory limit in bytes. Default: 0 (not specified).
- MemoryLimitInBytes int64 `protobuf:"varint,4,opt,name=memory_limit_in_bytes,json=memoryLimitInBytes,proto3" json:"memory_limit_in_bytes,omitempty"`
- // OOMScoreAdj adjusts the oom-killer score. Default: 0 (not specified).
- OomScoreAdj int64 `protobuf:"varint,5,opt,name=oom_score_adj,json=oomScoreAdj,proto3" json:"oom_score_adj,omitempty"`
- // CpusetCpus constrains the allowed set of logical CPUs. Default: "" (not specified).
- CpusetCpus string `protobuf:"bytes,6,opt,name=cpuset_cpus,json=cpusetCpus,proto3" json:"cpuset_cpus,omitempty"`
- // CpusetMems constrains the allowed set of memory nodes. Default: "" (not specified).
- CpusetMems string `protobuf:"bytes,7,opt,name=cpuset_mems,json=cpusetMems,proto3" json:"cpuset_mems,omitempty"`
-}
-
-func (m *LinuxContainerResources) Reset() { *m = LinuxContainerResources{} }
-func (*LinuxContainerResources) ProtoMessage() {}
-func (*LinuxContainerResources) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{30} }
-
-func (m *LinuxContainerResources) GetCpuPeriod() int64 {
- if m != nil {
- return m.CpuPeriod
- }
- return 0
-}
-
-func (m *LinuxContainerResources) GetCpuQuota() int64 {
- if m != nil {
- return m.CpuQuota
- }
- return 0
-}
-
-func (m *LinuxContainerResources) GetCpuShares() int64 {
- if m != nil {
- return m.CpuShares
- }
- return 0
-}
-
-func (m *LinuxContainerResources) GetMemoryLimitInBytes() int64 {
- if m != nil {
- return m.MemoryLimitInBytes
- }
- return 0
-}
-
-func (m *LinuxContainerResources) GetOomScoreAdj() int64 {
- if m != nil {
- return m.OomScoreAdj
- }
- return 0
-}
-
-func (m *LinuxContainerResources) GetCpusetCpus() string {
- if m != nil {
- return m.CpusetCpus
- }
- return ""
-}
-
-func (m *LinuxContainerResources) GetCpusetMems() string {
- if m != nil {
- return m.CpusetMems
- }
- return ""
-}
-
-// SELinuxOption are the labels to be applied to the container.
-type SELinuxOption struct {
- User string `protobuf:"bytes,1,opt,name=user,proto3" json:"user,omitempty"`
- Role string `protobuf:"bytes,2,opt,name=role,proto3" json:"role,omitempty"`
- Type string `protobuf:"bytes,3,opt,name=type,proto3" json:"type,omitempty"`
- Level string `protobuf:"bytes,4,opt,name=level,proto3" json:"level,omitempty"`
-}
-
-func (m *SELinuxOption) Reset() { *m = SELinuxOption{} }
-func (*SELinuxOption) ProtoMessage() {}
-func (*SELinuxOption) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{31} }
-
-func (m *SELinuxOption) GetUser() string {
- if m != nil {
- return m.User
- }
- return ""
-}
-
-func (m *SELinuxOption) GetRole() string {
- if m != nil {
- return m.Role
- }
- return ""
-}
-
-func (m *SELinuxOption) GetType() string {
- if m != nil {
- return m.Type
- }
- return ""
-}
-
-func (m *SELinuxOption) GetLevel() string {
- if m != nil {
- return m.Level
- }
- return ""
-}
-
-// Capability contains the container capabilities to add or drop
-type Capability struct {
- // List of capabilities to add.
- AddCapabilities []string `protobuf:"bytes,1,rep,name=add_capabilities,json=addCapabilities" json:"add_capabilities,omitempty"`
- // List of capabilities to drop.
- DropCapabilities []string `protobuf:"bytes,2,rep,name=drop_capabilities,json=dropCapabilities" json:"drop_capabilities,omitempty"`
-}
-
-func (m *Capability) Reset() { *m = Capability{} }
-func (*Capability) ProtoMessage() {}
-func (*Capability) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{32} }
-
-func (m *Capability) GetAddCapabilities() []string {
- if m != nil {
- return m.AddCapabilities
- }
- return nil
-}
-
-func (m *Capability) GetDropCapabilities() []string {
- if m != nil {
- return m.DropCapabilities
- }
- return nil
-}
-
-// LinuxContainerSecurityContext holds linux security configuration that will be applied to a container.
-type LinuxContainerSecurityContext struct {
- // Capabilities to add or drop.
- Capabilities *Capability `protobuf:"bytes,1,opt,name=capabilities" json:"capabilities,omitempty"`
- // If set, run container in privileged mode.
- // Privileged mode is incompatible with the following options. If
- // privileged is set, the following features MAY have no effect:
- // 1. capabilities
- // 2. selinux_options
- // 4. seccomp
- // 5. apparmor
- //
- // Privileged mode implies the following specific options are applied:
- // 1. All capabilities are added.
- // 2. Sensitive paths, such as kernel module paths within sysfs, are not masked.
- // 3. Any sysfs and procfs mounts are mounted RW.
- // 4. Apparmor confinement is not applied.
- // 5. Seccomp restrictions are not applied.
- // 6. The device cgroup does not restrict access to any devices.
- // 7. All devices from the host's /dev are available within the container.
- // 8. SELinux restrictions are not applied (e.g. label=disabled).
- Privileged bool `protobuf:"varint,2,opt,name=privileged,proto3" json:"privileged,omitempty"`
- // Configurations for the container's namespaces.
- // Only used if the container uses namespace for isolation.
- NamespaceOptions *NamespaceOption `protobuf:"bytes,3,opt,name=namespace_options,json=namespaceOptions" json:"namespace_options,omitempty"`
- // SELinux context to be optionally applied.
- SelinuxOptions *SELinuxOption `protobuf:"bytes,4,opt,name=selinux_options,json=selinuxOptions" json:"selinux_options,omitempty"`
- // UID to run the container process as. Only one of run_as_user and
- // run_as_username can be specified at a time.
- RunAsUser *Int64Value `protobuf:"bytes,5,opt,name=run_as_user,json=runAsUser" json:"run_as_user,omitempty"`
- // User name to run the container process as. If specified, the user MUST
- // exist in the container image (i.e. in the /etc/passwd inside the image),
- // and be resolved there by the runtime; otherwise, the runtime MUST error.
- RunAsUsername string `protobuf:"bytes,6,opt,name=run_as_username,json=runAsUsername,proto3" json:"run_as_username,omitempty"`
- // If set, the root filesystem of the container is read-only.
- ReadonlyRootfs bool `protobuf:"varint,7,opt,name=readonly_rootfs,json=readonlyRootfs,proto3" json:"readonly_rootfs,omitempty"`
- // List of groups applied to the first process run in the container, in
- // addition to the container's primary GID.
- SupplementalGroups []int64 `protobuf:"varint,8,rep,packed,name=supplemental_groups,json=supplementalGroups" json:"supplemental_groups,omitempty"`
- // AppArmor profile for the container, candidate values are:
- // * runtime/default: equivalent to not specifying a profile.
- // * unconfined: no profiles are loaded
- // * localhost/<profile_name>: profile loaded on the node
- // (localhost) by name. The possible profile names are detailed at
- // http://wiki.apparmor.net/index.php/AppArmor_Core_Policy_Reference
- ApparmorProfile string `protobuf:"bytes,9,opt,name=apparmor_profile,json=apparmorProfile,proto3" json:"apparmor_profile,omitempty"`
- // Seccomp profile for the container, candidate values are:
- // * docker/default: the default profile for the docker container runtime
- // * unconfined: unconfined profile, ie, no seccomp sandboxing
- // * localhost/<full-path-to-profile>: the profile installed on the node.
- // <full-path-to-profile> is the full path of the profile.
- // Default: "", which is identical with unconfined.
- SeccompProfilePath string `protobuf:"bytes,10,opt,name=seccomp_profile_path,json=seccompProfilePath,proto3" json:"seccomp_profile_path,omitempty"`
- // no_new_privs defines if the flag for no_new_privs should be set on the
- // container.
- NoNewPrivs bool `protobuf:"varint,11,opt,name=no_new_privs,json=noNewPrivs,proto3" json:"no_new_privs,omitempty"`
-}
-
-func (m *LinuxContainerSecurityContext) Reset() { *m = LinuxContainerSecurityContext{} }
-func (*LinuxContainerSecurityContext) ProtoMessage() {}
-func (*LinuxContainerSecurityContext) Descriptor() ([]byte, []int) {
- return fileDescriptorApi, []int{33}
-}
-
-func (m *LinuxContainerSecurityContext) GetCapabilities() *Capability {
- if m != nil {
- return m.Capabilities
- }
- return nil
-}
-
-func (m *LinuxContainerSecurityContext) GetPrivileged() bool {
- if m != nil {
- return m.Privileged
- }
- return false
-}
-
-func (m *LinuxContainerSecurityContext) GetNamespaceOptions() *NamespaceOption {
- if m != nil {
- return m.NamespaceOptions
- }
- return nil
-}
-
-func (m *LinuxContainerSecurityContext) GetSelinuxOptions() *SELinuxOption {
- if m != nil {
- return m.SelinuxOptions
- }
- return nil
-}
-
-func (m *LinuxContainerSecurityContext) GetRunAsUser() *Int64Value {
- if m != nil {
- return m.RunAsUser
- }
- return nil
-}
-
-func (m *LinuxContainerSecurityContext) GetRunAsUsername() string {
- if m != nil {
- return m.RunAsUsername
- }
- return ""
-}
-
-func (m *LinuxContainerSecurityContext) GetReadonlyRootfs() bool {
- if m != nil {
- return m.ReadonlyRootfs
- }
- return false
-}
-
-func (m *LinuxContainerSecurityContext) GetSupplementalGroups() []int64 {
- if m != nil {
- return m.SupplementalGroups
- }
- return nil
-}
-
-func (m *LinuxContainerSecurityContext) GetApparmorProfile() string {
- if m != nil {
- return m.ApparmorProfile
- }
- return ""
-}
-
-func (m *LinuxContainerSecurityContext) GetSeccompProfilePath() string {
- if m != nil {
- return m.SeccompProfilePath
- }
- return ""
-}
-
-func (m *LinuxContainerSecurityContext) GetNoNewPrivs() bool {
- if m != nil {
- return m.NoNewPrivs
- }
- return false
-}
-
-// LinuxContainerConfig contains platform-specific configuration for
-// Linux-based containers.
-type LinuxContainerConfig struct {
- // Resources specification for the container.
- Resources *LinuxContainerResources `protobuf:"bytes,1,opt,name=resources" json:"resources,omitempty"`
- // LinuxContainerSecurityContext configuration for the container.
- SecurityContext *LinuxContainerSecurityContext `protobuf:"bytes,2,opt,name=security_context,json=securityContext" json:"security_context,omitempty"`
-}
-
-func (m *LinuxContainerConfig) Reset() { *m = LinuxContainerConfig{} }
-func (*LinuxContainerConfig) ProtoMessage() {}
-func (*LinuxContainerConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{34} }
-
-func (m *LinuxContainerConfig) GetResources() *LinuxContainerResources {
- if m != nil {
- return m.Resources
- }
- return nil
-}
-
-func (m *LinuxContainerConfig) GetSecurityContext() *LinuxContainerSecurityContext {
- if m != nil {
- return m.SecurityContext
- }
- return nil
-}
-
-// WindowsContainerConfig contains platform-specific configuration for
-// Windows-based containers.
-type WindowsContainerConfig struct {
- // Resources specification for the container.
- Resources *WindowsContainerResources `protobuf:"bytes,1,opt,name=resources" json:"resources,omitempty"`
-}
-
-func (m *WindowsContainerConfig) Reset() { *m = WindowsContainerConfig{} }
-func (*WindowsContainerConfig) ProtoMessage() {}
-func (*WindowsContainerConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{35} }
-
-func (m *WindowsContainerConfig) GetResources() *WindowsContainerResources {
- if m != nil {
- return m.Resources
- }
- return nil
-}
-
-// WindowsContainerResources specifies Windows specific configuration for
-// resources.
-type WindowsContainerResources struct {
- // CPU shares (relative weight vs. other containers). Default: 0 (not specified).
- CpuShares int64 `protobuf:"varint,1,opt,name=cpu_shares,json=cpuShares,proto3" json:"cpu_shares,omitempty"`
- // Number of CPUs available to the container. Default: 0 (not specified).
- CpuCount int64 `protobuf:"varint,2,opt,name=cpu_count,json=cpuCount,proto3" json:"cpu_count,omitempty"`
- // Specifies the portion of processor cycles that this container can use as a percentage times 100.
- CpuMaximum int64 `protobuf:"varint,3,opt,name=cpu_maximum,json=cpuMaximum,proto3" json:"cpu_maximum,omitempty"`
- // Memory limit in bytes. Default: 0 (not specified).
- MemoryLimitInBytes int64 `protobuf:"varint,4,opt,name=memory_limit_in_bytes,json=memoryLimitInBytes,proto3" json:"memory_limit_in_bytes,omitempty"`
-}
-
-func (m *WindowsContainerResources) Reset() { *m = WindowsContainerResources{} }
-func (*WindowsContainerResources) ProtoMessage() {}
-func (*WindowsContainerResources) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{36} }
-
-func (m *WindowsContainerResources) GetCpuShares() int64 {
- if m != nil {
- return m.CpuShares
- }
- return 0
-}
-
-func (m *WindowsContainerResources) GetCpuCount() int64 {
- if m != nil {
- return m.CpuCount
- }
- return 0
-}
-
-func (m *WindowsContainerResources) GetCpuMaximum() int64 {
- if m != nil {
- return m.CpuMaximum
- }
- return 0
-}
-
-func (m *WindowsContainerResources) GetMemoryLimitInBytes() int64 {
- if m != nil {
- return m.MemoryLimitInBytes
- }
- return 0
-}
-
-// ContainerMetadata holds all necessary information for building the container
-// name. The container runtime is encouraged to expose the metadata in its user
-// interface for better user experience. E.g., runtime can construct a unique
-// container name based on the metadata. Note that (name, attempt) is unique
-// within a sandbox for the entire lifetime of the sandbox.
-type ContainerMetadata struct {
- // Name of the container. Same as the container name in the PodSpec.
- Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
- // Attempt number of creating the container. Default: 0.
- Attempt uint32 `protobuf:"varint,2,opt,name=attempt,proto3" json:"attempt,omitempty"`
-}
-
-func (m *ContainerMetadata) Reset() { *m = ContainerMetadata{} }
-func (*ContainerMetadata) ProtoMessage() {}
-func (*ContainerMetadata) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{37} }
-
-func (m *ContainerMetadata) GetName() string {
- if m != nil {
- return m.Name
- }
- return ""
-}
-
-func (m *ContainerMetadata) GetAttempt() uint32 {
- if m != nil {
- return m.Attempt
- }
- return 0
-}
-
-// Device specifies a host device to mount into a container.
-type Device struct {
- // Path of the device within the container.
- ContainerPath string `protobuf:"bytes,1,opt,name=container_path,json=containerPath,proto3" json:"container_path,omitempty"`
- // Path of the device on the host.
- HostPath string `protobuf:"bytes,2,opt,name=host_path,json=hostPath,proto3" json:"host_path,omitempty"`
- // Cgroups permissions of the device, candidates are one or more of
- // * r - allows container to read from the specified device.
- // * w - allows container to write to the specified device.
- // * m - allows container to create device files that do not yet exist.
- Permissions string `protobuf:"bytes,3,opt,name=permissions,proto3" json:"permissions,omitempty"`
-}
-
-func (m *Device) Reset() { *m = Device{} }
-func (*Device) ProtoMessage() {}
-func (*Device) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{38} }
-
-func (m *Device) GetContainerPath() string {
- if m != nil {
- return m.ContainerPath
- }
- return ""
-}
-
-func (m *Device) GetHostPath() string {
- if m != nil {
- return m.HostPath
- }
- return ""
-}
-
-func (m *Device) GetPermissions() string {
- if m != nil {
- return m.Permissions
- }
- return ""
-}
-
-// ContainerConfig holds all the required and optional fields for creating a
-// container.
-type ContainerConfig struct {
- // Metadata of the container. This information will uniquely identify the
- // container, and the runtime should leverage this to ensure correct
- // operation. The runtime may also use this information to improve UX, such
- // as by constructing a readable name.
- Metadata *ContainerMetadata `protobuf:"bytes,1,opt,name=metadata" json:"metadata,omitempty"`
- // Image to use.
- Image *ImageSpec `protobuf:"bytes,2,opt,name=image" json:"image,omitempty"`
- // Command to execute (i.e., entrypoint for docker)
- Command []string `protobuf:"bytes,3,rep,name=command" json:"command,omitempty"`
- // Args for the Command (i.e., command for docker)
- Args []string `protobuf:"bytes,4,rep,name=args" json:"args,omitempty"`
- // Current working directory of the command.
- WorkingDir string `protobuf:"bytes,5,opt,name=working_dir,json=workingDir,proto3" json:"working_dir,omitempty"`
- // List of environment variable to set in the container.
- Envs []*KeyValue `protobuf:"bytes,6,rep,name=envs" json:"envs,omitempty"`
- // Mounts for the container.
- Mounts []*Mount `protobuf:"bytes,7,rep,name=mounts" json:"mounts,omitempty"`
- // Devices for the container.
- Devices []*Device `protobuf:"bytes,8,rep,name=devices" json:"devices,omitempty"`
- // Key-value pairs that may be used to scope and select individual resources.
- // Label keys are of the form:
- // label-key ::= prefixed-name | name
- // prefixed-name ::= prefix '/' name
- // prefix ::= DNS_SUBDOMAIN
- // name ::= DNS_LABEL
- Labels map[string]string `protobuf:"bytes,9,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map that may be used by the kubelet to store and
- // retrieve arbitrary metadata.
- //
- // Annotations MUST NOT be altered by the runtime; the annotations stored
- // here MUST be returned in the ContainerStatus associated with the container
- // this ContainerConfig creates.
- //
- // In general, in order to preserve a well-defined interface between the
- // kubelet and the container runtime, annotations SHOULD NOT influence
- // runtime behaviour.
- Annotations map[string]string `protobuf:"bytes,10,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Path relative to PodSandboxConfig.LogDirectory for container to store
- // the log (STDOUT and STDERR) on the host.
- // E.g.,
- // PodSandboxConfig.LogDirectory = `/var/log/pods/<podUID>/`
- // ContainerConfig.LogPath = `containerName_Instance#.log`
- //
- // WARNING: Log management and how kubelet should interface with the
- // container logs are under active discussion in
- // https://issues.k8s.io/24677. There *may* be future change of direction
- // for logging as the discussion carries on.
- LogPath string `protobuf:"bytes,11,opt,name=log_path,json=logPath,proto3" json:"log_path,omitempty"`
- // Variables for interactive containers, these have very specialized
- // use-cases (e.g. debugging).
- // TODO: Determine if we need to continue supporting these fields that are
- // part of Kubernetes's Container Spec.
- Stdin bool `protobuf:"varint,12,opt,name=stdin,proto3" json:"stdin,omitempty"`
- StdinOnce bool `protobuf:"varint,13,opt,name=stdin_once,json=stdinOnce,proto3" json:"stdin_once,omitempty"`
- Tty bool `protobuf:"varint,14,opt,name=tty,proto3" json:"tty,omitempty"`
- // Configuration specific to Linux containers.
- Linux *LinuxContainerConfig `protobuf:"bytes,15,opt,name=linux" json:"linux,omitempty"`
- // Configuration specific to Windows containers.
- Windows *WindowsContainerConfig `protobuf:"bytes,16,opt,name=windows" json:"windows,omitempty"`
-}
-
-func (m *ContainerConfig) Reset() { *m = ContainerConfig{} }
-func (*ContainerConfig) ProtoMessage() {}
-func (*ContainerConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{39} }
-
-func (m *ContainerConfig) GetMetadata() *ContainerMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *ContainerConfig) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *ContainerConfig) GetCommand() []string {
- if m != nil {
- return m.Command
- }
- return nil
-}
-
-func (m *ContainerConfig) GetArgs() []string {
- if m != nil {
- return m.Args
- }
- return nil
-}
-
-func (m *ContainerConfig) GetWorkingDir() string {
- if m != nil {
- return m.WorkingDir
- }
- return ""
-}
-
-func (m *ContainerConfig) GetEnvs() []*KeyValue {
- if m != nil {
- return m.Envs
- }
- return nil
-}
-
-func (m *ContainerConfig) GetMounts() []*Mount {
- if m != nil {
- return m.Mounts
- }
- return nil
-}
-
-func (m *ContainerConfig) GetDevices() []*Device {
- if m != nil {
- return m.Devices
- }
- return nil
-}
-
-func (m *ContainerConfig) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *ContainerConfig) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-func (m *ContainerConfig) GetLogPath() string {
- if m != nil {
- return m.LogPath
- }
- return ""
-}
-
-func (m *ContainerConfig) GetStdin() bool {
- if m != nil {
- return m.Stdin
- }
- return false
-}
-
-func (m *ContainerConfig) GetStdinOnce() bool {
- if m != nil {
- return m.StdinOnce
- }
- return false
-}
-
-func (m *ContainerConfig) GetTty() bool {
- if m != nil {
- return m.Tty
- }
- return false
-}
-
-func (m *ContainerConfig) GetLinux() *LinuxContainerConfig {
- if m != nil {
- return m.Linux
- }
- return nil
-}
-
-func (m *ContainerConfig) GetWindows() *WindowsContainerConfig {
- if m != nil {
- return m.Windows
- }
- return nil
-}
-
-type CreateContainerRequest struct {
- // ID of the PodSandbox in which the container should be created.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // Config of the container.
- Config *ContainerConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"`
- // Config of the PodSandbox. This is the same config that was passed
- // to RunPodSandboxRequest to create the PodSandbox. It is passed again
- // here just for easy reference. The PodSandboxConfig is immutable and
- // remains the same throughout the lifetime of the pod.
- SandboxConfig *PodSandboxConfig `protobuf:"bytes,3,opt,name=sandbox_config,json=sandboxConfig" json:"sandbox_config,omitempty"`
-}
-
-func (m *CreateContainerRequest) Reset() { *m = CreateContainerRequest{} }
-func (*CreateContainerRequest) ProtoMessage() {}
-func (*CreateContainerRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{40} }
-
-func (m *CreateContainerRequest) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *CreateContainerRequest) GetConfig() *ContainerConfig {
- if m != nil {
- return m.Config
- }
- return nil
-}
-
-func (m *CreateContainerRequest) GetSandboxConfig() *PodSandboxConfig {
- if m != nil {
- return m.SandboxConfig
- }
- return nil
-}
-
-type CreateContainerResponse struct {
- // ID of the created container.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
-}
-
-func (m *CreateContainerResponse) Reset() { *m = CreateContainerResponse{} }
-func (*CreateContainerResponse) ProtoMessage() {}
-func (*CreateContainerResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{41} }
-
-func (m *CreateContainerResponse) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-type StartContainerRequest struct {
- // ID of the container to start.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
-}
-
-func (m *StartContainerRequest) Reset() { *m = StartContainerRequest{} }
-func (*StartContainerRequest) ProtoMessage() {}
-func (*StartContainerRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{42} }
-
-func (m *StartContainerRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-type StartContainerResponse struct {
-}
-
-func (m *StartContainerResponse) Reset() { *m = StartContainerResponse{} }
-func (*StartContainerResponse) ProtoMessage() {}
-func (*StartContainerResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{43} }
-
-type StopContainerRequest struct {
- // ID of the container to stop.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Timeout in seconds to wait for the container to stop before forcibly
- // terminating it. Default: 0 (forcibly terminate the container immediately)
- Timeout int64 `protobuf:"varint,2,opt,name=timeout,proto3" json:"timeout,omitempty"`
-}
-
-func (m *StopContainerRequest) Reset() { *m = StopContainerRequest{} }
-func (*StopContainerRequest) ProtoMessage() {}
-func (*StopContainerRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{44} }
-
-func (m *StopContainerRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *StopContainerRequest) GetTimeout() int64 {
- if m != nil {
- return m.Timeout
- }
- return 0
-}
-
-type StopContainerResponse struct {
-}
-
-func (m *StopContainerResponse) Reset() { *m = StopContainerResponse{} }
-func (*StopContainerResponse) ProtoMessage() {}
-func (*StopContainerResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{45} }
-
-type RemoveContainerRequest struct {
- // ID of the container to remove.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
-}
-
-func (m *RemoveContainerRequest) Reset() { *m = RemoveContainerRequest{} }
-func (*RemoveContainerRequest) ProtoMessage() {}
-func (*RemoveContainerRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{46} }
-
-func (m *RemoveContainerRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-type RemoveContainerResponse struct {
-}
-
-func (m *RemoveContainerResponse) Reset() { *m = RemoveContainerResponse{} }
-func (*RemoveContainerResponse) ProtoMessage() {}
-func (*RemoveContainerResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{47} }
-
-// ContainerStateValue is the wrapper of ContainerState.
-type ContainerStateValue struct {
- // State of the container.
- State ContainerState `protobuf:"varint,1,opt,name=state,proto3,enum=runtime.v1alpha2.ContainerState" json:"state,omitempty"`
-}
-
-func (m *ContainerStateValue) Reset() { *m = ContainerStateValue{} }
-func (*ContainerStateValue) ProtoMessage() {}
-func (*ContainerStateValue) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{48} }
-
-func (m *ContainerStateValue) GetState() ContainerState {
- if m != nil {
- return m.State
- }
- return ContainerState_CONTAINER_CREATED
-}
-
-// ContainerFilter is used to filter containers.
-// All those fields are combined with 'AND'
-type ContainerFilter struct {
- // ID of the container.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // State of the container.
- State *ContainerStateValue `protobuf:"bytes,2,opt,name=state" json:"state,omitempty"`
- // ID of the PodSandbox.
- PodSandboxId string `protobuf:"bytes,3,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- LabelSelector map[string]string `protobuf:"bytes,4,rep,name=label_selector,json=labelSelector" json:"label_selector,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *ContainerFilter) Reset() { *m = ContainerFilter{} }
-func (*ContainerFilter) ProtoMessage() {}
-func (*ContainerFilter) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{49} }
-
-func (m *ContainerFilter) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *ContainerFilter) GetState() *ContainerStateValue {
- if m != nil {
- return m.State
- }
- return nil
-}
-
-func (m *ContainerFilter) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *ContainerFilter) GetLabelSelector() map[string]string {
- if m != nil {
- return m.LabelSelector
- }
- return nil
-}
-
-type ListContainersRequest struct {
- Filter *ContainerFilter `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"`
-}
-
-func (m *ListContainersRequest) Reset() { *m = ListContainersRequest{} }
-func (*ListContainersRequest) ProtoMessage() {}
-func (*ListContainersRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{50} }
-
-func (m *ListContainersRequest) GetFilter() *ContainerFilter {
- if m != nil {
- return m.Filter
- }
- return nil
-}
-
-// Container provides the runtime information for a container, such as ID, hash,
-// state of the container.
-type Container struct {
- // ID of the container, used by the container runtime to identify
- // a container.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // ID of the sandbox to which this container belongs.
- PodSandboxId string `protobuf:"bytes,2,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // Metadata of the container.
- Metadata *ContainerMetadata `protobuf:"bytes,3,opt,name=metadata" json:"metadata,omitempty"`
- // Spec of the image.
- Image *ImageSpec `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"`
- // Reference to the image in use. For most runtimes, this should be an
- // image ID.
- ImageRef string `protobuf:"bytes,5,opt,name=image_ref,json=imageRef,proto3" json:"image_ref,omitempty"`
- // State of the container.
- State ContainerState `protobuf:"varint,6,opt,name=state,proto3,enum=runtime.v1alpha2.ContainerState" json:"state,omitempty"`
- // Creation time of the container in nanoseconds.
- CreatedAt int64 `protobuf:"varint,7,opt,name=created_at,json=createdAt,proto3" json:"created_at,omitempty"`
- // Key-value pairs that may be used to scope and select individual resources.
- Labels map[string]string `protobuf:"bytes,8,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate this Container.
- Annotations map[string]string `protobuf:"bytes,9,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *Container) Reset() { *m = Container{} }
-func (*Container) ProtoMessage() {}
-func (*Container) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{51} }
-
-func (m *Container) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *Container) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *Container) GetMetadata() *ContainerMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *Container) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *Container) GetImageRef() string {
- if m != nil {
- return m.ImageRef
- }
- return ""
-}
-
-func (m *Container) GetState() ContainerState {
- if m != nil {
- return m.State
- }
- return ContainerState_CONTAINER_CREATED
-}
-
-func (m *Container) GetCreatedAt() int64 {
- if m != nil {
- return m.CreatedAt
- }
- return 0
-}
-
-func (m *Container) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *Container) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-type ListContainersResponse struct {
- // List of containers.
- Containers []*Container `protobuf:"bytes,1,rep,name=containers" json:"containers,omitempty"`
-}
-
-func (m *ListContainersResponse) Reset() { *m = ListContainersResponse{} }
-func (*ListContainersResponse) ProtoMessage() {}
-func (*ListContainersResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{52} }
-
-func (m *ListContainersResponse) GetContainers() []*Container {
- if m != nil {
- return m.Containers
- }
- return nil
-}
-
-type ContainerStatusRequest struct {
- // ID of the container for which to retrieve status.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Verbose indicates whether to return extra information about the container.
- Verbose bool `protobuf:"varint,2,opt,name=verbose,proto3" json:"verbose,omitempty"`
-}
-
-func (m *ContainerStatusRequest) Reset() { *m = ContainerStatusRequest{} }
-func (*ContainerStatusRequest) ProtoMessage() {}
-func (*ContainerStatusRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{53} }
-
-func (m *ContainerStatusRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *ContainerStatusRequest) GetVerbose() bool {
- if m != nil {
- return m.Verbose
- }
- return false
-}
-
-// ContainerStatus represents the status of a container.
-type ContainerStatus struct {
- // ID of the container.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // Metadata of the container.
- Metadata *ContainerMetadata `protobuf:"bytes,2,opt,name=metadata" json:"metadata,omitempty"`
- // Status of the container.
- State ContainerState `protobuf:"varint,3,opt,name=state,proto3,enum=runtime.v1alpha2.ContainerState" json:"state,omitempty"`
- // Creation time of the container in nanoseconds.
- CreatedAt int64 `protobuf:"varint,4,opt,name=created_at,json=createdAt,proto3" json:"created_at,omitempty"`
- // Start time of the container in nanoseconds. Default: 0 (not specified).
- StartedAt int64 `protobuf:"varint,5,opt,name=started_at,json=startedAt,proto3" json:"started_at,omitempty"`
- // Finish time of the container in nanoseconds. Default: 0 (not specified).
- FinishedAt int64 `protobuf:"varint,6,opt,name=finished_at,json=finishedAt,proto3" json:"finished_at,omitempty"`
- // Exit code of the container. Only required when finished_at != 0. Default: 0.
- ExitCode int32 `protobuf:"varint,7,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"`
- // Spec of the image.
- Image *ImageSpec `protobuf:"bytes,8,opt,name=image" json:"image,omitempty"`
- // Reference to the image in use. For most runtimes, this should be an
- // image ID
- ImageRef string `protobuf:"bytes,9,opt,name=image_ref,json=imageRef,proto3" json:"image_ref,omitempty"`
- // Brief CamelCase string explaining why container is in its current state.
- Reason string `protobuf:"bytes,10,opt,name=reason,proto3" json:"reason,omitempty"`
- // Human-readable message indicating details about why container is in its
- // current state.
- Message string `protobuf:"bytes,11,opt,name=message,proto3" json:"message,omitempty"`
- // Key-value pairs that may be used to scope and select individual resources.
- Labels map[string]string `protobuf:"bytes,12,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate the Container this status represents.
- Annotations map[string]string `protobuf:"bytes,13,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Mounts for the container.
- Mounts []*Mount `protobuf:"bytes,14,rep,name=mounts" json:"mounts,omitempty"`
- // Log path of container.
- LogPath string `protobuf:"bytes,15,opt,name=log_path,json=logPath,proto3" json:"log_path,omitempty"`
-}
-
-func (m *ContainerStatus) Reset() { *m = ContainerStatus{} }
-func (*ContainerStatus) ProtoMessage() {}
-func (*ContainerStatus) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{54} }
-
-func (m *ContainerStatus) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *ContainerStatus) GetMetadata() *ContainerMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *ContainerStatus) GetState() ContainerState {
- if m != nil {
- return m.State
- }
- return ContainerState_CONTAINER_CREATED
-}
-
-func (m *ContainerStatus) GetCreatedAt() int64 {
- if m != nil {
- return m.CreatedAt
- }
- return 0
-}
-
-func (m *ContainerStatus) GetStartedAt() int64 {
- if m != nil {
- return m.StartedAt
- }
- return 0
-}
-
-func (m *ContainerStatus) GetFinishedAt() int64 {
- if m != nil {
- return m.FinishedAt
- }
- return 0
-}
-
-func (m *ContainerStatus) GetExitCode() int32 {
- if m != nil {
- return m.ExitCode
- }
- return 0
-}
-
-func (m *ContainerStatus) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *ContainerStatus) GetImageRef() string {
- if m != nil {
- return m.ImageRef
- }
- return ""
-}
-
-func (m *ContainerStatus) GetReason() string {
- if m != nil {
- return m.Reason
- }
- return ""
-}
-
-func (m *ContainerStatus) GetMessage() string {
- if m != nil {
- return m.Message
- }
- return ""
-}
-
-func (m *ContainerStatus) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *ContainerStatus) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-func (m *ContainerStatus) GetMounts() []*Mount {
- if m != nil {
- return m.Mounts
- }
- return nil
-}
-
-func (m *ContainerStatus) GetLogPath() string {
- if m != nil {
- return m.LogPath
- }
- return ""
-}
-
-type ContainerStatusResponse struct {
- // Status of the container.
- Status *ContainerStatus `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"`
- // Info is extra information of the Container. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. pid for linux container based container runtime.
- // It should only be returned non-empty when Verbose is true.
- Info map[string]string `protobuf:"bytes,2,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *ContainerStatusResponse) Reset() { *m = ContainerStatusResponse{} }
-func (*ContainerStatusResponse) ProtoMessage() {}
-func (*ContainerStatusResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{55} }
-
-func (m *ContainerStatusResponse) GetStatus() *ContainerStatus {
- if m != nil {
- return m.Status
- }
- return nil
-}
-
-func (m *ContainerStatusResponse) GetInfo() map[string]string {
- if m != nil {
- return m.Info
- }
- return nil
-}
-
-type UpdateContainerResourcesRequest struct {
- // ID of the container to update.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Resource configuration specific to Linux containers.
- Linux *LinuxContainerResources `protobuf:"bytes,2,opt,name=linux" json:"linux,omitempty"`
-}
-
-func (m *UpdateContainerResourcesRequest) Reset() { *m = UpdateContainerResourcesRequest{} }
-func (*UpdateContainerResourcesRequest) ProtoMessage() {}
-func (*UpdateContainerResourcesRequest) Descriptor() ([]byte, []int) {
- return fileDescriptorApi, []int{56}
-}
-
-func (m *UpdateContainerResourcesRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *UpdateContainerResourcesRequest) GetLinux() *LinuxContainerResources {
- if m != nil {
- return m.Linux
- }
- return nil
-}
-
-type UpdateContainerResourcesResponse struct {
-}
-
-func (m *UpdateContainerResourcesResponse) Reset() { *m = UpdateContainerResourcesResponse{} }
-func (*UpdateContainerResourcesResponse) ProtoMessage() {}
-func (*UpdateContainerResourcesResponse) Descriptor() ([]byte, []int) {
- return fileDescriptorApi, []int{57}
-}
-
-type ExecSyncRequest struct {
- // ID of the container.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Command to execute.
- Cmd []string `protobuf:"bytes,2,rep,name=cmd" json:"cmd,omitempty"`
- // Timeout in seconds to stop the command. Default: 0 (run forever).
- Timeout int64 `protobuf:"varint,3,opt,name=timeout,proto3" json:"timeout,omitempty"`
-}
-
-func (m *ExecSyncRequest) Reset() { *m = ExecSyncRequest{} }
-func (*ExecSyncRequest) ProtoMessage() {}
-func (*ExecSyncRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{58} }
-
-func (m *ExecSyncRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *ExecSyncRequest) GetCmd() []string {
- if m != nil {
- return m.Cmd
- }
- return nil
-}
-
-func (m *ExecSyncRequest) GetTimeout() int64 {
- if m != nil {
- return m.Timeout
- }
- return 0
-}
-
-type ExecSyncResponse struct {
- // Captured command stdout output.
- Stdout []byte `protobuf:"bytes,1,opt,name=stdout,proto3" json:"stdout,omitempty"`
- // Captured command stderr output.
- Stderr []byte `protobuf:"bytes,2,opt,name=stderr,proto3" json:"stderr,omitempty"`
- // Exit code the command finished with. Default: 0 (success).
- ExitCode int32 `protobuf:"varint,3,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"`
-}
-
-func (m *ExecSyncResponse) Reset() { *m = ExecSyncResponse{} }
-func (*ExecSyncResponse) ProtoMessage() {}
-func (*ExecSyncResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{59} }
-
-func (m *ExecSyncResponse) GetStdout() []byte {
- if m != nil {
- return m.Stdout
- }
- return nil
-}
-
-func (m *ExecSyncResponse) GetStderr() []byte {
- if m != nil {
- return m.Stderr
- }
- return nil
-}
-
-func (m *ExecSyncResponse) GetExitCode() int32 {
- if m != nil {
- return m.ExitCode
- }
- return 0
-}
-
-type ExecRequest struct {
- // ID of the container in which to execute the command.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Command to execute.
- Cmd []string `protobuf:"bytes,2,rep,name=cmd" json:"cmd,omitempty"`
- // Whether to exec the command in a TTY.
- Tty bool `protobuf:"varint,3,opt,name=tty,proto3" json:"tty,omitempty"`
- // Whether to stream stdin.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- Stdin bool `protobuf:"varint,4,opt,name=stdin,proto3" json:"stdin,omitempty"`
- // Whether to stream stdout.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- Stdout bool `protobuf:"varint,5,opt,name=stdout,proto3" json:"stdout,omitempty"`
- // Whether to stream stderr.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- // If `tty` is true, `stderr` MUST be false. Multiplexing is not supported
- // in this case. The output of stdout and stderr will be combined to a
- // single stream.
- Stderr bool `protobuf:"varint,6,opt,name=stderr,proto3" json:"stderr,omitempty"`
-}
-
-func (m *ExecRequest) Reset() { *m = ExecRequest{} }
-func (*ExecRequest) ProtoMessage() {}
-func (*ExecRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{60} }
-
-func (m *ExecRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *ExecRequest) GetCmd() []string {
- if m != nil {
- return m.Cmd
- }
- return nil
-}
-
-func (m *ExecRequest) GetTty() bool {
- if m != nil {
- return m.Tty
- }
- return false
-}
-
-func (m *ExecRequest) GetStdin() bool {
- if m != nil {
- return m.Stdin
- }
- return false
-}
-
-func (m *ExecRequest) GetStdout() bool {
- if m != nil {
- return m.Stdout
- }
- return false
-}
-
-func (m *ExecRequest) GetStderr() bool {
- if m != nil {
- return m.Stderr
- }
- return false
-}
-
-type ExecResponse struct {
- // Fully qualified URL of the exec streaming server.
- Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"`
-}
-
-func (m *ExecResponse) Reset() { *m = ExecResponse{} }
-func (*ExecResponse) ProtoMessage() {}
-func (*ExecResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{61} }
-
-func (m *ExecResponse) GetUrl() string {
- if m != nil {
- return m.Url
- }
- return ""
-}
-
-type AttachRequest struct {
- // ID of the container to which to attach.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
- // Whether to stream stdin.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- Stdin bool `protobuf:"varint,2,opt,name=stdin,proto3" json:"stdin,omitempty"`
- // Whether the process being attached is running in a TTY.
- // This must match the TTY setting in the ContainerConfig.
- Tty bool `protobuf:"varint,3,opt,name=tty,proto3" json:"tty,omitempty"`
- // Whether to stream stdout.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- Stdout bool `protobuf:"varint,4,opt,name=stdout,proto3" json:"stdout,omitempty"`
- // Whether to stream stderr.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- // If `tty` is true, `stderr` MUST be false. Multiplexing is not supported
- // in this case. The output of stdout and stderr will be combined to a
- // single stream.
- Stderr bool `protobuf:"varint,5,opt,name=stderr,proto3" json:"stderr,omitempty"`
-}
-
-func (m *AttachRequest) Reset() { *m = AttachRequest{} }
-func (*AttachRequest) ProtoMessage() {}
-func (*AttachRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{62} }
-
-func (m *AttachRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-func (m *AttachRequest) GetStdin() bool {
- if m != nil {
- return m.Stdin
- }
- return false
-}
-
-func (m *AttachRequest) GetTty() bool {
- if m != nil {
- return m.Tty
- }
- return false
-}
-
-func (m *AttachRequest) GetStdout() bool {
- if m != nil {
- return m.Stdout
- }
- return false
-}
-
-func (m *AttachRequest) GetStderr() bool {
- if m != nil {
- return m.Stderr
- }
- return false
-}
-
-type AttachResponse struct {
- // Fully qualified URL of the attach streaming server.
- Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"`
-}
-
-func (m *AttachResponse) Reset() { *m = AttachResponse{} }
-func (*AttachResponse) ProtoMessage() {}
-func (*AttachResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{63} }
-
-func (m *AttachResponse) GetUrl() string {
- if m != nil {
- return m.Url
- }
- return ""
-}
-
-type PortForwardRequest struct {
- // ID of the container to which to forward the port.
- PodSandboxId string `protobuf:"bytes,1,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // Port to forward.
- Port []int32 `protobuf:"varint,2,rep,packed,name=port" json:"port,omitempty"`
-}
-
-func (m *PortForwardRequest) Reset() { *m = PortForwardRequest{} }
-func (*PortForwardRequest) ProtoMessage() {}
-func (*PortForwardRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{64} }
-
-func (m *PortForwardRequest) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *PortForwardRequest) GetPort() []int32 {
- if m != nil {
- return m.Port
- }
- return nil
-}
-
-type PortForwardResponse struct {
- // Fully qualified URL of the port-forward streaming server.
- Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"`
-}
-
-func (m *PortForwardResponse) Reset() { *m = PortForwardResponse{} }
-func (*PortForwardResponse) ProtoMessage() {}
-func (*PortForwardResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{65} }
-
-func (m *PortForwardResponse) GetUrl() string {
- if m != nil {
- return m.Url
- }
- return ""
-}
-
-type ImageFilter struct {
- // Spec of the image.
- Image *ImageSpec `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"`
-}
-
-func (m *ImageFilter) Reset() { *m = ImageFilter{} }
-func (*ImageFilter) ProtoMessage() {}
-func (*ImageFilter) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{66} }
-
-func (m *ImageFilter) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-type ListImagesRequest struct {
- // Filter to list images.
- Filter *ImageFilter `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"`
-}
-
-func (m *ListImagesRequest) Reset() { *m = ListImagesRequest{} }
-func (*ListImagesRequest) ProtoMessage() {}
-func (*ListImagesRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{67} }
-
-func (m *ListImagesRequest) GetFilter() *ImageFilter {
- if m != nil {
- return m.Filter
- }
- return nil
-}
-
-// Basic information about a container image.
-type Image struct {
- // ID of the image.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // Other names by which this image is known.
- RepoTags []string `protobuf:"bytes,2,rep,name=repo_tags,json=repoTags" json:"repo_tags,omitempty"`
- // Digests by which this image is known.
- RepoDigests []string `protobuf:"bytes,3,rep,name=repo_digests,json=repoDigests" json:"repo_digests,omitempty"`
- // Size of the image in bytes. Must be > 0.
- Size_ uint64 `protobuf:"varint,4,opt,name=size,proto3" json:"size,omitempty"`
- // UID that will run the command(s). This is used as a default if no user is
- // specified when creating the container. UID and the following user name
- // are mutually exclusive.
- Uid *Int64Value `protobuf:"bytes,5,opt,name=uid" json:"uid,omitempty"`
- // User name that will run the command(s). This is used if UID is not set
- // and no user is specified when creating container.
- Username string `protobuf:"bytes,6,opt,name=username,proto3" json:"username,omitempty"`
-}
-
-func (m *Image) Reset() { *m = Image{} }
-func (*Image) ProtoMessage() {}
-func (*Image) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{68} }
-
-func (m *Image) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *Image) GetRepoTags() []string {
- if m != nil {
- return m.RepoTags
- }
- return nil
-}
-
-func (m *Image) GetRepoDigests() []string {
- if m != nil {
- return m.RepoDigests
- }
- return nil
-}
-
-func (m *Image) GetSize_() uint64 {
- if m != nil {
- return m.Size_
- }
- return 0
-}
-
-func (m *Image) GetUid() *Int64Value {
- if m != nil {
- return m.Uid
- }
- return nil
-}
-
-func (m *Image) GetUsername() string {
- if m != nil {
- return m.Username
- }
- return ""
-}
-
-type ListImagesResponse struct {
- // List of images.
- Images []*Image `protobuf:"bytes,1,rep,name=images" json:"images,omitempty"`
-}
-
-func (m *ListImagesResponse) Reset() { *m = ListImagesResponse{} }
-func (*ListImagesResponse) ProtoMessage() {}
-func (*ListImagesResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{69} }
-
-func (m *ListImagesResponse) GetImages() []*Image {
- if m != nil {
- return m.Images
- }
- return nil
-}
-
-type ImageStatusRequest struct {
- // Spec of the image.
- Image *ImageSpec `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"`
- // Verbose indicates whether to return extra information about the image.
- Verbose bool `protobuf:"varint,2,opt,name=verbose,proto3" json:"verbose,omitempty"`
-}
-
-func (m *ImageStatusRequest) Reset() { *m = ImageStatusRequest{} }
-func (*ImageStatusRequest) ProtoMessage() {}
-func (*ImageStatusRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{70} }
-
-func (m *ImageStatusRequest) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *ImageStatusRequest) GetVerbose() bool {
- if m != nil {
- return m.Verbose
- }
- return false
-}
-
-type ImageStatusResponse struct {
- // Status of the image.
- Image *Image `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"`
- // Info is extra information of the Image. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful
- // for debug, e.g. image config for oci image based container runtime.
- // It should only be returned non-empty when Verbose is true.
- Info map[string]string `protobuf:"bytes,2,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *ImageStatusResponse) Reset() { *m = ImageStatusResponse{} }
-func (*ImageStatusResponse) ProtoMessage() {}
-func (*ImageStatusResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{71} }
-
-func (m *ImageStatusResponse) GetImage() *Image {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *ImageStatusResponse) GetInfo() map[string]string {
- if m != nil {
- return m.Info
- }
- return nil
-}
-
-// AuthConfig contains authorization information for connecting to a registry.
-type AuthConfig struct {
- Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"`
- Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"`
- Auth string `protobuf:"bytes,3,opt,name=auth,proto3" json:"auth,omitempty"`
- ServerAddress string `protobuf:"bytes,4,opt,name=server_address,json=serverAddress,proto3" json:"server_address,omitempty"`
- // IdentityToken is used to authenticate the user and get
- // an access token for the registry.
- IdentityToken string `protobuf:"bytes,5,opt,name=identity_token,json=identityToken,proto3" json:"identity_token,omitempty"`
- // RegistryToken is a bearer token to be sent to a registry
- RegistryToken string `protobuf:"bytes,6,opt,name=registry_token,json=registryToken,proto3" json:"registry_token,omitempty"`
-}
-
-func (m *AuthConfig) Reset() { *m = AuthConfig{} }
-func (*AuthConfig) ProtoMessage() {}
-func (*AuthConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{72} }
-
-func (m *AuthConfig) GetUsername() string {
- if m != nil {
- return m.Username
- }
- return ""
-}
-
-func (m *AuthConfig) GetPassword() string {
- if m != nil {
- return m.Password
- }
- return ""
-}
-
-func (m *AuthConfig) GetAuth() string {
- if m != nil {
- return m.Auth
- }
- return ""
-}
-
-func (m *AuthConfig) GetServerAddress() string {
- if m != nil {
- return m.ServerAddress
- }
- return ""
-}
-
-func (m *AuthConfig) GetIdentityToken() string {
- if m != nil {
- return m.IdentityToken
- }
- return ""
-}
-
-func (m *AuthConfig) GetRegistryToken() string {
- if m != nil {
- return m.RegistryToken
- }
- return ""
-}
-
-type PullImageRequest struct {
- // Spec of the image.
- Image *ImageSpec `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"`
- // Authentication configuration for pulling the image.
- Auth *AuthConfig `protobuf:"bytes,2,opt,name=auth" json:"auth,omitempty"`
- // Config of the PodSandbox, which is used to pull image in PodSandbox context.
- SandboxConfig *PodSandboxConfig `protobuf:"bytes,3,opt,name=sandbox_config,json=sandboxConfig" json:"sandbox_config,omitempty"`
-}
-
-func (m *PullImageRequest) Reset() { *m = PullImageRequest{} }
-func (*PullImageRequest) ProtoMessage() {}
-func (*PullImageRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{73} }
-
-func (m *PullImageRequest) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-func (m *PullImageRequest) GetAuth() *AuthConfig {
- if m != nil {
- return m.Auth
- }
- return nil
-}
-
-func (m *PullImageRequest) GetSandboxConfig() *PodSandboxConfig {
- if m != nil {
- return m.SandboxConfig
- }
- return nil
-}
-
-type PullImageResponse struct {
- // Reference to the image in use. For most runtimes, this should be an
- // image ID or digest.
- ImageRef string `protobuf:"bytes,1,opt,name=image_ref,json=imageRef,proto3" json:"image_ref,omitempty"`
-}
-
-func (m *PullImageResponse) Reset() { *m = PullImageResponse{} }
-func (*PullImageResponse) ProtoMessage() {}
-func (*PullImageResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{74} }
-
-func (m *PullImageResponse) GetImageRef() string {
- if m != nil {
- return m.ImageRef
- }
- return ""
-}
-
-type RemoveImageRequest struct {
- // Spec of the image to remove.
- Image *ImageSpec `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"`
-}
-
-func (m *RemoveImageRequest) Reset() { *m = RemoveImageRequest{} }
-func (*RemoveImageRequest) ProtoMessage() {}
-func (*RemoveImageRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{75} }
-
-func (m *RemoveImageRequest) GetImage() *ImageSpec {
- if m != nil {
- return m.Image
- }
- return nil
-}
-
-type RemoveImageResponse struct {
-}
-
-func (m *RemoveImageResponse) Reset() { *m = RemoveImageResponse{} }
-func (*RemoveImageResponse) ProtoMessage() {}
-func (*RemoveImageResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{76} }
-
-type NetworkConfig struct {
- // CIDR to use for pod IP addresses.
- PodCidr string `protobuf:"bytes,1,opt,name=pod_cidr,json=podCidr,proto3" json:"pod_cidr,omitempty"`
-}
-
-func (m *NetworkConfig) Reset() { *m = NetworkConfig{} }
-func (*NetworkConfig) ProtoMessage() {}
-func (*NetworkConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{77} }
-
-func (m *NetworkConfig) GetPodCidr() string {
- if m != nil {
- return m.PodCidr
- }
- return ""
-}
-
-type RuntimeConfig struct {
- NetworkConfig *NetworkConfig `protobuf:"bytes,1,opt,name=network_config,json=networkConfig" json:"network_config,omitempty"`
-}
-
-func (m *RuntimeConfig) Reset() { *m = RuntimeConfig{} }
-func (*RuntimeConfig) ProtoMessage() {}
-func (*RuntimeConfig) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{78} }
-
-func (m *RuntimeConfig) GetNetworkConfig() *NetworkConfig {
- if m != nil {
- return m.NetworkConfig
- }
- return nil
-}
-
-type UpdateRuntimeConfigRequest struct {
- RuntimeConfig *RuntimeConfig `protobuf:"bytes,1,opt,name=runtime_config,json=runtimeConfig" json:"runtime_config,omitempty"`
-}
-
-func (m *UpdateRuntimeConfigRequest) Reset() { *m = UpdateRuntimeConfigRequest{} }
-func (*UpdateRuntimeConfigRequest) ProtoMessage() {}
-func (*UpdateRuntimeConfigRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{79} }
-
-func (m *UpdateRuntimeConfigRequest) GetRuntimeConfig() *RuntimeConfig {
- if m != nil {
- return m.RuntimeConfig
- }
- return nil
-}
-
-type UpdateRuntimeConfigResponse struct {
-}
-
-func (m *UpdateRuntimeConfigResponse) Reset() { *m = UpdateRuntimeConfigResponse{} }
-func (*UpdateRuntimeConfigResponse) ProtoMessage() {}
-func (*UpdateRuntimeConfigResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{80} }
-
-// RuntimeCondition contains condition information for the runtime.
-// There are 2 kinds of runtime conditions:
-// 1. Required conditions: Conditions are required for kubelet to work
-// properly. If any required condition is unmet, the node will be not ready.
-// The required conditions include:
-// * RuntimeReady: RuntimeReady means the runtime is up and ready to accept
-// basic containers e.g. container only needs host network.
-// * NetworkReady: NetworkReady means the runtime network is up and ready to
-// accept containers which require container network.
-// 2. Optional conditions: Conditions are informative to the user, but kubelet
-// will not rely on. Since condition type is an arbitrary string, all conditions
-// not required are optional. These conditions will be exposed to users to help
-// them understand the status of the system.
-type RuntimeCondition struct {
- // Type of runtime condition.
- Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"`
- // Status of the condition, one of true/false. Default: false.
- Status bool `protobuf:"varint,2,opt,name=status,proto3" json:"status,omitempty"`
- // Brief CamelCase string containing reason for the condition's last transition.
- Reason string `protobuf:"bytes,3,opt,name=reason,proto3" json:"reason,omitempty"`
- // Human-readable message indicating details about last transition.
- Message string `protobuf:"bytes,4,opt,name=message,proto3" json:"message,omitempty"`
-}
-
-func (m *RuntimeCondition) Reset() { *m = RuntimeCondition{} }
-func (*RuntimeCondition) ProtoMessage() {}
-func (*RuntimeCondition) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{81} }
-
-func (m *RuntimeCondition) GetType() string {
- if m != nil {
- return m.Type
- }
- return ""
-}
-
-func (m *RuntimeCondition) GetStatus() bool {
- if m != nil {
- return m.Status
- }
- return false
-}
-
-func (m *RuntimeCondition) GetReason() string {
- if m != nil {
- return m.Reason
- }
- return ""
-}
-
-func (m *RuntimeCondition) GetMessage() string {
- if m != nil {
- return m.Message
- }
- return ""
-}
-
-// RuntimeStatus is information about the current status of the runtime.
-type RuntimeStatus struct {
- // List of current observed runtime conditions.
- Conditions []*RuntimeCondition `protobuf:"bytes,1,rep,name=conditions" json:"conditions,omitempty"`
-}
-
-func (m *RuntimeStatus) Reset() { *m = RuntimeStatus{} }
-func (*RuntimeStatus) ProtoMessage() {}
-func (*RuntimeStatus) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{82} }
-
-func (m *RuntimeStatus) GetConditions() []*RuntimeCondition {
- if m != nil {
- return m.Conditions
- }
- return nil
-}
-
-type StatusRequest struct {
- // Verbose indicates whether to return extra information about the runtime.
- Verbose bool `protobuf:"varint,1,opt,name=verbose,proto3" json:"verbose,omitempty"`
-}
-
-func (m *StatusRequest) Reset() { *m = StatusRequest{} }
-func (*StatusRequest) ProtoMessage() {}
-func (*StatusRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{83} }
-
-func (m *StatusRequest) GetVerbose() bool {
- if m != nil {
- return m.Verbose
- }
- return false
-}
-
-type StatusResponse struct {
- // Status of the Runtime.
- Status *RuntimeStatus `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"`
- // Info is extra information of the Runtime. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. plugins used by the container runtime.
- // It should only be returned non-empty when Verbose is true.
- Info map[string]string `protobuf:"bytes,2,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *StatusResponse) Reset() { *m = StatusResponse{} }
-func (*StatusResponse) ProtoMessage() {}
-func (*StatusResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{84} }
-
-func (m *StatusResponse) GetStatus() *RuntimeStatus {
- if m != nil {
- return m.Status
- }
- return nil
-}
-
-func (m *StatusResponse) GetInfo() map[string]string {
- if m != nil {
- return m.Info
- }
- return nil
-}
-
-type ImageFsInfoRequest struct {
-}
-
-func (m *ImageFsInfoRequest) Reset() { *m = ImageFsInfoRequest{} }
-func (*ImageFsInfoRequest) ProtoMessage() {}
-func (*ImageFsInfoRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{85} }
-
-// UInt64Value is the wrapper of uint64.
-type UInt64Value struct {
- // The value.
- Value uint64 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"`
-}
-
-func (m *UInt64Value) Reset() { *m = UInt64Value{} }
-func (*UInt64Value) ProtoMessage() {}
-func (*UInt64Value) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{86} }
-
-func (m *UInt64Value) GetValue() uint64 {
- if m != nil {
- return m.Value
- }
- return 0
-}
-
-// FilesystemIdentifier uniquely identify the filesystem.
-type FilesystemIdentifier struct {
- // Mountpoint of a filesystem.
- Mountpoint string `protobuf:"bytes,1,opt,name=mountpoint,proto3" json:"mountpoint,omitempty"`
-}
-
-func (m *FilesystemIdentifier) Reset() { *m = FilesystemIdentifier{} }
-func (*FilesystemIdentifier) ProtoMessage() {}
-func (*FilesystemIdentifier) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{87} }
-
-func (m *FilesystemIdentifier) GetMountpoint() string {
- if m != nil {
- return m.Mountpoint
- }
- return ""
-}
-
-// FilesystemUsage provides the filesystem usage information.
-type FilesystemUsage struct {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- Timestamp int64 `protobuf:"varint,1,opt,name=timestamp,proto3" json:"timestamp,omitempty"`
- // The unique identifier of the filesystem.
- FsId *FilesystemIdentifier `protobuf:"bytes,2,opt,name=fs_id,json=fsId" json:"fs_id,omitempty"`
- // UsedBytes represents the bytes used for images on the filesystem.
- // This may differ from the total bytes used on the filesystem and may not
- // equal CapacityBytes - AvailableBytes.
- UsedBytes *UInt64Value `protobuf:"bytes,3,opt,name=used_bytes,json=usedBytes" json:"used_bytes,omitempty"`
- // InodesUsed represents the inodes used by the images.
- // This may not equal InodesCapacity - InodesAvailable because the underlying
- // filesystem may also be used for purposes other than storing images.
- InodesUsed *UInt64Value `protobuf:"bytes,4,opt,name=inodes_used,json=inodesUsed" json:"inodes_used,omitempty"`
-}
-
-func (m *FilesystemUsage) Reset() { *m = FilesystemUsage{} }
-func (*FilesystemUsage) ProtoMessage() {}
-func (*FilesystemUsage) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{88} }
-
-func (m *FilesystemUsage) GetTimestamp() int64 {
- if m != nil {
- return m.Timestamp
- }
- return 0
-}
-
-func (m *FilesystemUsage) GetFsId() *FilesystemIdentifier {
- if m != nil {
- return m.FsId
- }
- return nil
-}
-
-func (m *FilesystemUsage) GetUsedBytes() *UInt64Value {
- if m != nil {
- return m.UsedBytes
- }
- return nil
-}
-
-func (m *FilesystemUsage) GetInodesUsed() *UInt64Value {
- if m != nil {
- return m.InodesUsed
- }
- return nil
-}
-
-type ImageFsInfoResponse struct {
- // Information of image filesystem(s).
- ImageFilesystems []*FilesystemUsage `protobuf:"bytes,1,rep,name=image_filesystems,json=imageFilesystems" json:"image_filesystems,omitempty"`
-}
-
-func (m *ImageFsInfoResponse) Reset() { *m = ImageFsInfoResponse{} }
-func (*ImageFsInfoResponse) ProtoMessage() {}
-func (*ImageFsInfoResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{89} }
-
-func (m *ImageFsInfoResponse) GetImageFilesystems() []*FilesystemUsage {
- if m != nil {
- return m.ImageFilesystems
- }
- return nil
-}
-
-type ContainerStatsRequest struct {
- // ID of the container for which to retrieve stats.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
-}
-
-func (m *ContainerStatsRequest) Reset() { *m = ContainerStatsRequest{} }
-func (*ContainerStatsRequest) ProtoMessage() {}
-func (*ContainerStatsRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{90} }
-
-func (m *ContainerStatsRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-type ContainerStatsResponse struct {
- // Stats of the container.
- Stats *ContainerStats `protobuf:"bytes,1,opt,name=stats" json:"stats,omitempty"`
-}
-
-func (m *ContainerStatsResponse) Reset() { *m = ContainerStatsResponse{} }
-func (*ContainerStatsResponse) ProtoMessage() {}
-func (*ContainerStatsResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{91} }
-
-func (m *ContainerStatsResponse) GetStats() *ContainerStats {
- if m != nil {
- return m.Stats
- }
- return nil
-}
-
-type ListContainerStatsRequest struct {
- // Filter for the list request.
- Filter *ContainerStatsFilter `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"`
-}
-
-func (m *ListContainerStatsRequest) Reset() { *m = ListContainerStatsRequest{} }
-func (*ListContainerStatsRequest) ProtoMessage() {}
-func (*ListContainerStatsRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{92} }
-
-func (m *ListContainerStatsRequest) GetFilter() *ContainerStatsFilter {
- if m != nil {
- return m.Filter
- }
- return nil
-}
-
-// ContainerStatsFilter is used to filter containers.
-// All those fields are combined with 'AND'
-type ContainerStatsFilter struct {
- // ID of the container.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // ID of the PodSandbox.
- PodSandboxId string `protobuf:"bytes,2,opt,name=pod_sandbox_id,json=podSandboxId,proto3" json:"pod_sandbox_id,omitempty"`
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- LabelSelector map[string]string `protobuf:"bytes,3,rep,name=label_selector,json=labelSelector" json:"label_selector,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *ContainerStatsFilter) Reset() { *m = ContainerStatsFilter{} }
-func (*ContainerStatsFilter) ProtoMessage() {}
-func (*ContainerStatsFilter) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{93} }
-
-func (m *ContainerStatsFilter) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *ContainerStatsFilter) GetPodSandboxId() string {
- if m != nil {
- return m.PodSandboxId
- }
- return ""
-}
-
-func (m *ContainerStatsFilter) GetLabelSelector() map[string]string {
- if m != nil {
- return m.LabelSelector
- }
- return nil
-}
-
-type ListContainerStatsResponse struct {
- // Stats of the container.
- Stats []*ContainerStats `protobuf:"bytes,1,rep,name=stats" json:"stats,omitempty"`
-}
-
-func (m *ListContainerStatsResponse) Reset() { *m = ListContainerStatsResponse{} }
-func (*ListContainerStatsResponse) ProtoMessage() {}
-func (*ListContainerStatsResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{94} }
-
-func (m *ListContainerStatsResponse) GetStats() []*ContainerStats {
- if m != nil {
- return m.Stats
- }
- return nil
-}
-
-// ContainerAttributes provides basic information of the container.
-type ContainerAttributes struct {
- // ID of the container.
- Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"`
- // Metadata of the container.
- Metadata *ContainerMetadata `protobuf:"bytes,2,opt,name=metadata" json:"metadata,omitempty"`
- // Key-value pairs that may be used to scope and select individual resources.
- Labels map[string]string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate the Container this status represents.
- Annotations map[string]string `protobuf:"bytes,4,rep,name=annotations" json:"annotations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"`
-}
-
-func (m *ContainerAttributes) Reset() { *m = ContainerAttributes{} }
-func (*ContainerAttributes) ProtoMessage() {}
-func (*ContainerAttributes) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{95} }
-
-func (m *ContainerAttributes) GetId() string {
- if m != nil {
- return m.Id
- }
- return ""
-}
-
-func (m *ContainerAttributes) GetMetadata() *ContainerMetadata {
- if m != nil {
- return m.Metadata
- }
- return nil
-}
-
-func (m *ContainerAttributes) GetLabels() map[string]string {
- if m != nil {
- return m.Labels
- }
- return nil
-}
-
-func (m *ContainerAttributes) GetAnnotations() map[string]string {
- if m != nil {
- return m.Annotations
- }
- return nil
-}
-
-// ContainerStats provides the resource usage statistics for a container.
-type ContainerStats struct {
- // Information of the container.
- Attributes *ContainerAttributes `protobuf:"bytes,1,opt,name=attributes" json:"attributes,omitempty"`
- // CPU usage gathered from the container.
- Cpu *CpuUsage `protobuf:"bytes,2,opt,name=cpu" json:"cpu,omitempty"`
- // Memory usage gathered from the container.
- Memory *MemoryUsage `protobuf:"bytes,3,opt,name=memory" json:"memory,omitempty"`
- // Usage of the writeable layer.
- WritableLayer *FilesystemUsage `protobuf:"bytes,4,opt,name=writable_layer,json=writableLayer" json:"writable_layer,omitempty"`
-}
-
-func (m *ContainerStats) Reset() { *m = ContainerStats{} }
-func (*ContainerStats) ProtoMessage() {}
-func (*ContainerStats) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{96} }
-
-func (m *ContainerStats) GetAttributes() *ContainerAttributes {
- if m != nil {
- return m.Attributes
- }
- return nil
-}
-
-func (m *ContainerStats) GetCpu() *CpuUsage {
- if m != nil {
- return m.Cpu
- }
- return nil
-}
-
-func (m *ContainerStats) GetMemory() *MemoryUsage {
- if m != nil {
- return m.Memory
- }
- return nil
-}
-
-func (m *ContainerStats) GetWritableLayer() *FilesystemUsage {
- if m != nil {
- return m.WritableLayer
- }
- return nil
-}
-
-// CpuUsage provides the CPU usage information.
-type CpuUsage struct {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- Timestamp int64 `protobuf:"varint,1,opt,name=timestamp,proto3" json:"timestamp,omitempty"`
- // Cumulative CPU usage (sum across all cores) since object creation.
- UsageCoreNanoSeconds *UInt64Value `protobuf:"bytes,2,opt,name=usage_core_nano_seconds,json=usageCoreNanoSeconds" json:"usage_core_nano_seconds,omitempty"`
-}
-
-func (m *CpuUsage) Reset() { *m = CpuUsage{} }
-func (*CpuUsage) ProtoMessage() {}
-func (*CpuUsage) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{97} }
-
-func (m *CpuUsage) GetTimestamp() int64 {
- if m != nil {
- return m.Timestamp
- }
- return 0
-}
-
-func (m *CpuUsage) GetUsageCoreNanoSeconds() *UInt64Value {
- if m != nil {
- return m.UsageCoreNanoSeconds
- }
- return nil
-}
-
-// MemoryUsage provides the memory usage information.
-type MemoryUsage struct {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- Timestamp int64 `protobuf:"varint,1,opt,name=timestamp,proto3" json:"timestamp,omitempty"`
- // The amount of working set memory in bytes.
- WorkingSetBytes *UInt64Value `protobuf:"bytes,2,opt,name=working_set_bytes,json=workingSetBytes" json:"working_set_bytes,omitempty"`
-}
-
-func (m *MemoryUsage) Reset() { *m = MemoryUsage{} }
-func (*MemoryUsage) ProtoMessage() {}
-func (*MemoryUsage) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{98} }
-
-func (m *MemoryUsage) GetTimestamp() int64 {
- if m != nil {
- return m.Timestamp
- }
- return 0
-}
-
-func (m *MemoryUsage) GetWorkingSetBytes() *UInt64Value {
- if m != nil {
- return m.WorkingSetBytes
- }
- return nil
-}
-
-type ReopenContainerLogRequest struct {
- // ID of the container for which to reopen the log.
- ContainerId string `protobuf:"bytes,1,opt,name=container_id,json=containerId,proto3" json:"container_id,omitempty"`
-}
-
-func (m *ReopenContainerLogRequest) Reset() { *m = ReopenContainerLogRequest{} }
-func (*ReopenContainerLogRequest) ProtoMessage() {}
-func (*ReopenContainerLogRequest) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{99} }
-
-func (m *ReopenContainerLogRequest) GetContainerId() string {
- if m != nil {
- return m.ContainerId
- }
- return ""
-}
-
-type ReopenContainerLogResponse struct {
-}
-
-func (m *ReopenContainerLogResponse) Reset() { *m = ReopenContainerLogResponse{} }
-func (*ReopenContainerLogResponse) ProtoMessage() {}
-func (*ReopenContainerLogResponse) Descriptor() ([]byte, []int) { return fileDescriptorApi, []int{100} }
-
-func init() {
- proto.RegisterType((*VersionRequest)(nil), "runtime.v1alpha2.VersionRequest")
- proto.RegisterType((*VersionResponse)(nil), "runtime.v1alpha2.VersionResponse")
- proto.RegisterType((*DNSConfig)(nil), "runtime.v1alpha2.DNSConfig")
- proto.RegisterType((*PortMapping)(nil), "runtime.v1alpha2.PortMapping")
- proto.RegisterType((*Mount)(nil), "runtime.v1alpha2.Mount")
- proto.RegisterType((*NamespaceOption)(nil), "runtime.v1alpha2.NamespaceOption")
- proto.RegisterType((*Int64Value)(nil), "runtime.v1alpha2.Int64Value")
- proto.RegisterType((*LinuxSandboxSecurityContext)(nil), "runtime.v1alpha2.LinuxSandboxSecurityContext")
- proto.RegisterType((*LinuxPodSandboxConfig)(nil), "runtime.v1alpha2.LinuxPodSandboxConfig")
- proto.RegisterType((*PodSandboxMetadata)(nil), "runtime.v1alpha2.PodSandboxMetadata")
- proto.RegisterType((*PodSandboxConfig)(nil), "runtime.v1alpha2.PodSandboxConfig")
- proto.RegisterType((*RunPodSandboxRequest)(nil), "runtime.v1alpha2.RunPodSandboxRequest")
- proto.RegisterType((*RunPodSandboxResponse)(nil), "runtime.v1alpha2.RunPodSandboxResponse")
- proto.RegisterType((*StopPodSandboxRequest)(nil), "runtime.v1alpha2.StopPodSandboxRequest")
- proto.RegisterType((*StopPodSandboxResponse)(nil), "runtime.v1alpha2.StopPodSandboxResponse")
- proto.RegisterType((*RemovePodSandboxRequest)(nil), "runtime.v1alpha2.RemovePodSandboxRequest")
- proto.RegisterType((*RemovePodSandboxResponse)(nil), "runtime.v1alpha2.RemovePodSandboxResponse")
- proto.RegisterType((*PodSandboxStatusRequest)(nil), "runtime.v1alpha2.PodSandboxStatusRequest")
- proto.RegisterType((*PodSandboxNetworkStatus)(nil), "runtime.v1alpha2.PodSandboxNetworkStatus")
- proto.RegisterType((*Namespace)(nil), "runtime.v1alpha2.Namespace")
- proto.RegisterType((*LinuxPodSandboxStatus)(nil), "runtime.v1alpha2.LinuxPodSandboxStatus")
- proto.RegisterType((*PodSandboxStatus)(nil), "runtime.v1alpha2.PodSandboxStatus")
- proto.RegisterType((*PodSandboxStatusResponse)(nil), "runtime.v1alpha2.PodSandboxStatusResponse")
- proto.RegisterType((*PodSandboxStateValue)(nil), "runtime.v1alpha2.PodSandboxStateValue")
- proto.RegisterType((*PodSandboxFilter)(nil), "runtime.v1alpha2.PodSandboxFilter")
- proto.RegisterType((*ListPodSandboxRequest)(nil), "runtime.v1alpha2.ListPodSandboxRequest")
- proto.RegisterType((*PodSandbox)(nil), "runtime.v1alpha2.PodSandbox")
- proto.RegisterType((*ListPodSandboxResponse)(nil), "runtime.v1alpha2.ListPodSandboxResponse")
- proto.RegisterType((*ImageSpec)(nil), "runtime.v1alpha2.ImageSpec")
- proto.RegisterType((*KeyValue)(nil), "runtime.v1alpha2.KeyValue")
- proto.RegisterType((*LinuxContainerResources)(nil), "runtime.v1alpha2.LinuxContainerResources")
- proto.RegisterType((*SELinuxOption)(nil), "runtime.v1alpha2.SELinuxOption")
- proto.RegisterType((*Capability)(nil), "runtime.v1alpha2.Capability")
- proto.RegisterType((*LinuxContainerSecurityContext)(nil), "runtime.v1alpha2.LinuxContainerSecurityContext")
- proto.RegisterType((*LinuxContainerConfig)(nil), "runtime.v1alpha2.LinuxContainerConfig")
- proto.RegisterType((*WindowsContainerConfig)(nil), "runtime.v1alpha2.WindowsContainerConfig")
- proto.RegisterType((*WindowsContainerResources)(nil), "runtime.v1alpha2.WindowsContainerResources")
- proto.RegisterType((*ContainerMetadata)(nil), "runtime.v1alpha2.ContainerMetadata")
- proto.RegisterType((*Device)(nil), "runtime.v1alpha2.Device")
- proto.RegisterType((*ContainerConfig)(nil), "runtime.v1alpha2.ContainerConfig")
- proto.RegisterType((*CreateContainerRequest)(nil), "runtime.v1alpha2.CreateContainerRequest")
- proto.RegisterType((*CreateContainerResponse)(nil), "runtime.v1alpha2.CreateContainerResponse")
- proto.RegisterType((*StartContainerRequest)(nil), "runtime.v1alpha2.StartContainerRequest")
- proto.RegisterType((*StartContainerResponse)(nil), "runtime.v1alpha2.StartContainerResponse")
- proto.RegisterType((*StopContainerRequest)(nil), "runtime.v1alpha2.StopContainerRequest")
- proto.RegisterType((*StopContainerResponse)(nil), "runtime.v1alpha2.StopContainerResponse")
- proto.RegisterType((*RemoveContainerRequest)(nil), "runtime.v1alpha2.RemoveContainerRequest")
- proto.RegisterType((*RemoveContainerResponse)(nil), "runtime.v1alpha2.RemoveContainerResponse")
- proto.RegisterType((*ContainerStateValue)(nil), "runtime.v1alpha2.ContainerStateValue")
- proto.RegisterType((*ContainerFilter)(nil), "runtime.v1alpha2.ContainerFilter")
- proto.RegisterType((*ListContainersRequest)(nil), "runtime.v1alpha2.ListContainersRequest")
- proto.RegisterType((*Container)(nil), "runtime.v1alpha2.Container")
- proto.RegisterType((*ListContainersResponse)(nil), "runtime.v1alpha2.ListContainersResponse")
- proto.RegisterType((*ContainerStatusRequest)(nil), "runtime.v1alpha2.ContainerStatusRequest")
- proto.RegisterType((*ContainerStatus)(nil), "runtime.v1alpha2.ContainerStatus")
- proto.RegisterType((*ContainerStatusResponse)(nil), "runtime.v1alpha2.ContainerStatusResponse")
- proto.RegisterType((*UpdateContainerResourcesRequest)(nil), "runtime.v1alpha2.UpdateContainerResourcesRequest")
- proto.RegisterType((*UpdateContainerResourcesResponse)(nil), "runtime.v1alpha2.UpdateContainerResourcesResponse")
- proto.RegisterType((*ExecSyncRequest)(nil), "runtime.v1alpha2.ExecSyncRequest")
- proto.RegisterType((*ExecSyncResponse)(nil), "runtime.v1alpha2.ExecSyncResponse")
- proto.RegisterType((*ExecRequest)(nil), "runtime.v1alpha2.ExecRequest")
- proto.RegisterType((*ExecResponse)(nil), "runtime.v1alpha2.ExecResponse")
- proto.RegisterType((*AttachRequest)(nil), "runtime.v1alpha2.AttachRequest")
- proto.RegisterType((*AttachResponse)(nil), "runtime.v1alpha2.AttachResponse")
- proto.RegisterType((*PortForwardRequest)(nil), "runtime.v1alpha2.PortForwardRequest")
- proto.RegisterType((*PortForwardResponse)(nil), "runtime.v1alpha2.PortForwardResponse")
- proto.RegisterType((*ImageFilter)(nil), "runtime.v1alpha2.ImageFilter")
- proto.RegisterType((*ListImagesRequest)(nil), "runtime.v1alpha2.ListImagesRequest")
- proto.RegisterType((*Image)(nil), "runtime.v1alpha2.Image")
- proto.RegisterType((*ListImagesResponse)(nil), "runtime.v1alpha2.ListImagesResponse")
- proto.RegisterType((*ImageStatusRequest)(nil), "runtime.v1alpha2.ImageStatusRequest")
- proto.RegisterType((*ImageStatusResponse)(nil), "runtime.v1alpha2.ImageStatusResponse")
- proto.RegisterType((*AuthConfig)(nil), "runtime.v1alpha2.AuthConfig")
- proto.RegisterType((*PullImageRequest)(nil), "runtime.v1alpha2.PullImageRequest")
- proto.RegisterType((*PullImageResponse)(nil), "runtime.v1alpha2.PullImageResponse")
- proto.RegisterType((*RemoveImageRequest)(nil), "runtime.v1alpha2.RemoveImageRequest")
- proto.RegisterType((*RemoveImageResponse)(nil), "runtime.v1alpha2.RemoveImageResponse")
- proto.RegisterType((*NetworkConfig)(nil), "runtime.v1alpha2.NetworkConfig")
- proto.RegisterType((*RuntimeConfig)(nil), "runtime.v1alpha2.RuntimeConfig")
- proto.RegisterType((*UpdateRuntimeConfigRequest)(nil), "runtime.v1alpha2.UpdateRuntimeConfigRequest")
- proto.RegisterType((*UpdateRuntimeConfigResponse)(nil), "runtime.v1alpha2.UpdateRuntimeConfigResponse")
- proto.RegisterType((*RuntimeCondition)(nil), "runtime.v1alpha2.RuntimeCondition")
- proto.RegisterType((*RuntimeStatus)(nil), "runtime.v1alpha2.RuntimeStatus")
- proto.RegisterType((*StatusRequest)(nil), "runtime.v1alpha2.StatusRequest")
- proto.RegisterType((*StatusResponse)(nil), "runtime.v1alpha2.StatusResponse")
- proto.RegisterType((*ImageFsInfoRequest)(nil), "runtime.v1alpha2.ImageFsInfoRequest")
- proto.RegisterType((*UInt64Value)(nil), "runtime.v1alpha2.UInt64Value")
- proto.RegisterType((*FilesystemIdentifier)(nil), "runtime.v1alpha2.FilesystemIdentifier")
- proto.RegisterType((*FilesystemUsage)(nil), "runtime.v1alpha2.FilesystemUsage")
- proto.RegisterType((*ImageFsInfoResponse)(nil), "runtime.v1alpha2.ImageFsInfoResponse")
- proto.RegisterType((*ContainerStatsRequest)(nil), "runtime.v1alpha2.ContainerStatsRequest")
- proto.RegisterType((*ContainerStatsResponse)(nil), "runtime.v1alpha2.ContainerStatsResponse")
- proto.RegisterType((*ListContainerStatsRequest)(nil), "runtime.v1alpha2.ListContainerStatsRequest")
- proto.RegisterType((*ContainerStatsFilter)(nil), "runtime.v1alpha2.ContainerStatsFilter")
- proto.RegisterType((*ListContainerStatsResponse)(nil), "runtime.v1alpha2.ListContainerStatsResponse")
- proto.RegisterType((*ContainerAttributes)(nil), "runtime.v1alpha2.ContainerAttributes")
- proto.RegisterType((*ContainerStats)(nil), "runtime.v1alpha2.ContainerStats")
- proto.RegisterType((*CpuUsage)(nil), "runtime.v1alpha2.CpuUsage")
- proto.RegisterType((*MemoryUsage)(nil), "runtime.v1alpha2.MemoryUsage")
- proto.RegisterType((*ReopenContainerLogRequest)(nil), "runtime.v1alpha2.ReopenContainerLogRequest")
- proto.RegisterType((*ReopenContainerLogResponse)(nil), "runtime.v1alpha2.ReopenContainerLogResponse")
- proto.RegisterEnum("runtime.v1alpha2.Protocol", Protocol_name, Protocol_value)
- proto.RegisterEnum("runtime.v1alpha2.MountPropagation", MountPropagation_name, MountPropagation_value)
- proto.RegisterEnum("runtime.v1alpha2.NamespaceMode", NamespaceMode_name, NamespaceMode_value)
- proto.RegisterEnum("runtime.v1alpha2.PodSandboxState", PodSandboxState_name, PodSandboxState_value)
- proto.RegisterEnum("runtime.v1alpha2.ContainerState", ContainerState_name, ContainerState_value)
-}
-
-// Reference imports to suppress errors if they are not otherwise used.
-var _ context.Context
-var _ grpc.ClientConn
-
-// This is a compile-time assertion to ensure that this generated file
-// is compatible with the grpc package it is being compiled against.
-const _ = grpc.SupportPackageIsVersion4
-
-// Client API for RuntimeService service
-
-type RuntimeServiceClient interface {
- // Version returns the runtime name, runtime version, and runtime API version.
- Version(ctx context.Context, in *VersionRequest, opts ...grpc.CallOption) (*VersionResponse, error)
- // RunPodSandbox creates and starts a pod-level sandbox. Runtimes must ensure
- // the sandbox is in the ready state on success.
- RunPodSandbox(ctx context.Context, in *RunPodSandboxRequest, opts ...grpc.CallOption) (*RunPodSandboxResponse, error)
- // StopPodSandbox stops any running process that is part of the sandbox and
- // reclaims network resources (e.g., IP addresses) allocated to the sandbox.
- // If there are any running containers in the sandbox, they must be forcibly
- // terminated.
- // This call is idempotent, and must not return an error if all relevant
- // resources have already been reclaimed. kubelet will call StopPodSandbox
- // at least once before calling RemovePodSandbox. It will also attempt to
- // reclaim resources eagerly, as soon as a sandbox is not needed. Hence,
- // multiple StopPodSandbox calls are expected.
- StopPodSandbox(ctx context.Context, in *StopPodSandboxRequest, opts ...grpc.CallOption) (*StopPodSandboxResponse, error)
- // RemovePodSandbox removes the sandbox. If there are any running containers
- // in the sandbox, they must be forcibly terminated and removed.
- // This call is idempotent, and must not return an error if the sandbox has
- // already been removed.
- RemovePodSandbox(ctx context.Context, in *RemovePodSandboxRequest, opts ...grpc.CallOption) (*RemovePodSandboxResponse, error)
- // PodSandboxStatus returns the status of the PodSandbox. If the PodSandbox is not
- // present, returns an error.
- PodSandboxStatus(ctx context.Context, in *PodSandboxStatusRequest, opts ...grpc.CallOption) (*PodSandboxStatusResponse, error)
- // ListPodSandbox returns a list of PodSandboxes.
- ListPodSandbox(ctx context.Context, in *ListPodSandboxRequest, opts ...grpc.CallOption) (*ListPodSandboxResponse, error)
- // CreateContainer creates a new container in specified PodSandbox
- CreateContainer(ctx context.Context, in *CreateContainerRequest, opts ...grpc.CallOption) (*CreateContainerResponse, error)
- // StartContainer starts the container.
- StartContainer(ctx context.Context, in *StartContainerRequest, opts ...grpc.CallOption) (*StartContainerResponse, error)
- // StopContainer stops a running container with a grace period (i.e., timeout).
- // This call is idempotent, and must not return an error if the container has
- // already been stopped.
- // TODO: what must the runtime do after the grace period is reached?
- StopContainer(ctx context.Context, in *StopContainerRequest, opts ...grpc.CallOption) (*StopContainerResponse, error)
- // RemoveContainer removes the container. If the container is running, the
- // container must be forcibly removed.
- // This call is idempotent, and must not return an error if the container has
- // already been removed.
- RemoveContainer(ctx context.Context, in *RemoveContainerRequest, opts ...grpc.CallOption) (*RemoveContainerResponse, error)
- // ListContainers lists all containers by filters.
- ListContainers(ctx context.Context, in *ListContainersRequest, opts ...grpc.CallOption) (*ListContainersResponse, error)
- // ContainerStatus returns status of the container. If the container is not
- // present, returns an error.
- ContainerStatus(ctx context.Context, in *ContainerStatusRequest, opts ...grpc.CallOption) (*ContainerStatusResponse, error)
- // UpdateContainerResources updates ContainerConfig of the container.
- UpdateContainerResources(ctx context.Context, in *UpdateContainerResourcesRequest, opts ...grpc.CallOption) (*UpdateContainerResourcesResponse, error)
- // ReopenContainerLog asks runtime to reopen the stdout/stderr log file
- // for the container. This is often called after the log file has been
- // rotated. If the container is not running, container runtime can choose
- // to either create a new log file and return nil, or return an error.
- // Once it returns error, new container log file MUST NOT be created.
- ReopenContainerLog(ctx context.Context, in *ReopenContainerLogRequest, opts ...grpc.CallOption) (*ReopenContainerLogResponse, error)
- // ExecSync runs a command in a container synchronously.
- ExecSync(ctx context.Context, in *ExecSyncRequest, opts ...grpc.CallOption) (*ExecSyncResponse, error)
- // Exec prepares a streaming endpoint to execute a command in the container.
- Exec(ctx context.Context, in *ExecRequest, opts ...grpc.CallOption) (*ExecResponse, error)
- // Attach prepares a streaming endpoint to attach to a running container.
- Attach(ctx context.Context, in *AttachRequest, opts ...grpc.CallOption) (*AttachResponse, error)
- // PortForward prepares a streaming endpoint to forward ports from a PodSandbox.
- PortForward(ctx context.Context, in *PortForwardRequest, opts ...grpc.CallOption) (*PortForwardResponse, error)
- // ContainerStats returns stats of the container. If the container does not
- // exist, the call returns an error.
- ContainerStats(ctx context.Context, in *ContainerStatsRequest, opts ...grpc.CallOption) (*ContainerStatsResponse, error)
- // ListContainerStats returns stats of all running containers.
- ListContainerStats(ctx context.Context, in *ListContainerStatsRequest, opts ...grpc.CallOption) (*ListContainerStatsResponse, error)
- // UpdateRuntimeConfig updates the runtime configuration based on the given request.
- UpdateRuntimeConfig(ctx context.Context, in *UpdateRuntimeConfigRequest, opts ...grpc.CallOption) (*UpdateRuntimeConfigResponse, error)
- // Status returns the status of the runtime.
- Status(ctx context.Context, in *StatusRequest, opts ...grpc.CallOption) (*StatusResponse, error)
-}
-
-type runtimeServiceClient struct {
- cc *grpc.ClientConn
-}
-
-func NewRuntimeServiceClient(cc *grpc.ClientConn) RuntimeServiceClient {
- return &runtimeServiceClient{cc}
-}
-
-func (c *runtimeServiceClient) Version(ctx context.Context, in *VersionRequest, opts ...grpc.CallOption) (*VersionResponse, error) {
- out := new(VersionResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/Version", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) RunPodSandbox(ctx context.Context, in *RunPodSandboxRequest, opts ...grpc.CallOption) (*RunPodSandboxResponse, error) {
- out := new(RunPodSandboxResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/RunPodSandbox", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) StopPodSandbox(ctx context.Context, in *StopPodSandboxRequest, opts ...grpc.CallOption) (*StopPodSandboxResponse, error) {
- out := new(StopPodSandboxResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/StopPodSandbox", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) RemovePodSandbox(ctx context.Context, in *RemovePodSandboxRequest, opts ...grpc.CallOption) (*RemovePodSandboxResponse, error) {
- out := new(RemovePodSandboxResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/RemovePodSandbox", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) PodSandboxStatus(ctx context.Context, in *PodSandboxStatusRequest, opts ...grpc.CallOption) (*PodSandboxStatusResponse, error) {
- out := new(PodSandboxStatusResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/PodSandboxStatus", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ListPodSandbox(ctx context.Context, in *ListPodSandboxRequest, opts ...grpc.CallOption) (*ListPodSandboxResponse, error) {
- out := new(ListPodSandboxResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ListPodSandbox", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) CreateContainer(ctx context.Context, in *CreateContainerRequest, opts ...grpc.CallOption) (*CreateContainerResponse, error) {
- out := new(CreateContainerResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/CreateContainer", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) StartContainer(ctx context.Context, in *StartContainerRequest, opts ...grpc.CallOption) (*StartContainerResponse, error) {
- out := new(StartContainerResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/StartContainer", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) StopContainer(ctx context.Context, in *StopContainerRequest, opts ...grpc.CallOption) (*StopContainerResponse, error) {
- out := new(StopContainerResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/StopContainer", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) RemoveContainer(ctx context.Context, in *RemoveContainerRequest, opts ...grpc.CallOption) (*RemoveContainerResponse, error) {
- out := new(RemoveContainerResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/RemoveContainer", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ListContainers(ctx context.Context, in *ListContainersRequest, opts ...grpc.CallOption) (*ListContainersResponse, error) {
- out := new(ListContainersResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ListContainers", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ContainerStatus(ctx context.Context, in *ContainerStatusRequest, opts ...grpc.CallOption) (*ContainerStatusResponse, error) {
- out := new(ContainerStatusResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ContainerStatus", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) UpdateContainerResources(ctx context.Context, in *UpdateContainerResourcesRequest, opts ...grpc.CallOption) (*UpdateContainerResourcesResponse, error) {
- out := new(UpdateContainerResourcesResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/UpdateContainerResources", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ReopenContainerLog(ctx context.Context, in *ReopenContainerLogRequest, opts ...grpc.CallOption) (*ReopenContainerLogResponse, error) {
- out := new(ReopenContainerLogResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ReopenContainerLog", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ExecSync(ctx context.Context, in *ExecSyncRequest, opts ...grpc.CallOption) (*ExecSyncResponse, error) {
- out := new(ExecSyncResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ExecSync", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) Exec(ctx context.Context, in *ExecRequest, opts ...grpc.CallOption) (*ExecResponse, error) {
- out := new(ExecResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/Exec", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) Attach(ctx context.Context, in *AttachRequest, opts ...grpc.CallOption) (*AttachResponse, error) {
- out := new(AttachResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/Attach", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) PortForward(ctx context.Context, in *PortForwardRequest, opts ...grpc.CallOption) (*PortForwardResponse, error) {
- out := new(PortForwardResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/PortForward", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ContainerStats(ctx context.Context, in *ContainerStatsRequest, opts ...grpc.CallOption) (*ContainerStatsResponse, error) {
- out := new(ContainerStatsResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ContainerStats", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) ListContainerStats(ctx context.Context, in *ListContainerStatsRequest, opts ...grpc.CallOption) (*ListContainerStatsResponse, error) {
- out := new(ListContainerStatsResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/ListContainerStats", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) UpdateRuntimeConfig(ctx context.Context, in *UpdateRuntimeConfigRequest, opts ...grpc.CallOption) (*UpdateRuntimeConfigResponse, error) {
- out := new(UpdateRuntimeConfigResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/UpdateRuntimeConfig", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *runtimeServiceClient) Status(ctx context.Context, in *StatusRequest, opts ...grpc.CallOption) (*StatusResponse, error) {
- out := new(StatusResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.RuntimeService/Status", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-// Server API for RuntimeService service
-
-type RuntimeServiceServer interface {
- // Version returns the runtime name, runtime version, and runtime API version.
- Version(context.Context, *VersionRequest) (*VersionResponse, error)
- // RunPodSandbox creates and starts a pod-level sandbox. Runtimes must ensure
- // the sandbox is in the ready state on success.
- RunPodSandbox(context.Context, *RunPodSandboxRequest) (*RunPodSandboxResponse, error)
- // StopPodSandbox stops any running process that is part of the sandbox and
- // reclaims network resources (e.g., IP addresses) allocated to the sandbox.
- // If there are any running containers in the sandbox, they must be forcibly
- // terminated.
- // This call is idempotent, and must not return an error if all relevant
- // resources have already been reclaimed. kubelet will call StopPodSandbox
- // at least once before calling RemovePodSandbox. It will also attempt to
- // reclaim resources eagerly, as soon as a sandbox is not needed. Hence,
- // multiple StopPodSandbox calls are expected.
- StopPodSandbox(context.Context, *StopPodSandboxRequest) (*StopPodSandboxResponse, error)
- // RemovePodSandbox removes the sandbox. If there are any running containers
- // in the sandbox, they must be forcibly terminated and removed.
- // This call is idempotent, and must not return an error if the sandbox has
- // already been removed.
- RemovePodSandbox(context.Context, *RemovePodSandboxRequest) (*RemovePodSandboxResponse, error)
- // PodSandboxStatus returns the status of the PodSandbox. If the PodSandbox is not
- // present, returns an error.
- PodSandboxStatus(context.Context, *PodSandboxStatusRequest) (*PodSandboxStatusResponse, error)
- // ListPodSandbox returns a list of PodSandboxes.
- ListPodSandbox(context.Context, *ListPodSandboxRequest) (*ListPodSandboxResponse, error)
- // CreateContainer creates a new container in specified PodSandbox
- CreateContainer(context.Context, *CreateContainerRequest) (*CreateContainerResponse, error)
- // StartContainer starts the container.
- StartContainer(context.Context, *StartContainerRequest) (*StartContainerResponse, error)
- // StopContainer stops a running container with a grace period (i.e., timeout).
- // This call is idempotent, and must not return an error if the container has
- // already been stopped.
- // TODO: what must the runtime do after the grace period is reached?
- StopContainer(context.Context, *StopContainerRequest) (*StopContainerResponse, error)
- // RemoveContainer removes the container. If the container is running, the
- // container must be forcibly removed.
- // This call is idempotent, and must not return an error if the container has
- // already been removed.
- RemoveContainer(context.Context, *RemoveContainerRequest) (*RemoveContainerResponse, error)
- // ListContainers lists all containers by filters.
- ListContainers(context.Context, *ListContainersRequest) (*ListContainersResponse, error)
- // ContainerStatus returns status of the container. If the container is not
- // present, returns an error.
- ContainerStatus(context.Context, *ContainerStatusRequest) (*ContainerStatusResponse, error)
- // UpdateContainerResources updates ContainerConfig of the container.
- UpdateContainerResources(context.Context, *UpdateContainerResourcesRequest) (*UpdateContainerResourcesResponse, error)
- // ReopenContainerLog asks runtime to reopen the stdout/stderr log file
- // for the container. This is often called after the log file has been
- // rotated. If the container is not running, container runtime can choose
- // to either create a new log file and return nil, or return an error.
- // Once it returns error, new container log file MUST NOT be created.
- ReopenContainerLog(context.Context, *ReopenContainerLogRequest) (*ReopenContainerLogResponse, error)
- // ExecSync runs a command in a container synchronously.
- ExecSync(context.Context, *ExecSyncRequest) (*ExecSyncResponse, error)
- // Exec prepares a streaming endpoint to execute a command in the container.
- Exec(context.Context, *ExecRequest) (*ExecResponse, error)
- // Attach prepares a streaming endpoint to attach to a running container.
- Attach(context.Context, *AttachRequest) (*AttachResponse, error)
- // PortForward prepares a streaming endpoint to forward ports from a PodSandbox.
- PortForward(context.Context, *PortForwardRequest) (*PortForwardResponse, error)
- // ContainerStats returns stats of the container. If the container does not
- // exist, the call returns an error.
- ContainerStats(context.Context, *ContainerStatsRequest) (*ContainerStatsResponse, error)
- // ListContainerStats returns stats of all running containers.
- ListContainerStats(context.Context, *ListContainerStatsRequest) (*ListContainerStatsResponse, error)
- // UpdateRuntimeConfig updates the runtime configuration based on the given request.
- UpdateRuntimeConfig(context.Context, *UpdateRuntimeConfigRequest) (*UpdateRuntimeConfigResponse, error)
- // Status returns the status of the runtime.
- Status(context.Context, *StatusRequest) (*StatusResponse, error)
-}
-
-func RegisterRuntimeServiceServer(s *grpc.Server, srv RuntimeServiceServer) {
- s.RegisterService(&_RuntimeService_serviceDesc, srv)
-}
-
-func _RuntimeService_Version_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(VersionRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).Version(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/Version",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).Version(ctx, req.(*VersionRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_RunPodSandbox_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(RunPodSandboxRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).RunPodSandbox(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/RunPodSandbox",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).RunPodSandbox(ctx, req.(*RunPodSandboxRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_StopPodSandbox_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(StopPodSandboxRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).StopPodSandbox(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/StopPodSandbox",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).StopPodSandbox(ctx, req.(*StopPodSandboxRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_RemovePodSandbox_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(RemovePodSandboxRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).RemovePodSandbox(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/RemovePodSandbox",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).RemovePodSandbox(ctx, req.(*RemovePodSandboxRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_PodSandboxStatus_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(PodSandboxStatusRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).PodSandboxStatus(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/PodSandboxStatus",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).PodSandboxStatus(ctx, req.(*PodSandboxStatusRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ListPodSandbox_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ListPodSandboxRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ListPodSandbox(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ListPodSandbox",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ListPodSandbox(ctx, req.(*ListPodSandboxRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_CreateContainer_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(CreateContainerRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).CreateContainer(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/CreateContainer",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).CreateContainer(ctx, req.(*CreateContainerRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_StartContainer_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(StartContainerRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).StartContainer(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/StartContainer",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).StartContainer(ctx, req.(*StartContainerRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_StopContainer_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(StopContainerRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).StopContainer(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/StopContainer",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).StopContainer(ctx, req.(*StopContainerRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_RemoveContainer_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(RemoveContainerRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).RemoveContainer(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/RemoveContainer",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).RemoveContainer(ctx, req.(*RemoveContainerRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ListContainers_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ListContainersRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ListContainers(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ListContainers",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ListContainers(ctx, req.(*ListContainersRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ContainerStatus_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ContainerStatusRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ContainerStatus(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ContainerStatus",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ContainerStatus(ctx, req.(*ContainerStatusRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_UpdateContainerResources_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(UpdateContainerResourcesRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).UpdateContainerResources(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/UpdateContainerResources",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).UpdateContainerResources(ctx, req.(*UpdateContainerResourcesRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ReopenContainerLog_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ReopenContainerLogRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ReopenContainerLog(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ReopenContainerLog",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ReopenContainerLog(ctx, req.(*ReopenContainerLogRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ExecSync_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ExecSyncRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ExecSync(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ExecSync",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ExecSync(ctx, req.(*ExecSyncRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_Exec_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ExecRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).Exec(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/Exec",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).Exec(ctx, req.(*ExecRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_Attach_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(AttachRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).Attach(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/Attach",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).Attach(ctx, req.(*AttachRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_PortForward_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(PortForwardRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).PortForward(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/PortForward",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).PortForward(ctx, req.(*PortForwardRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ContainerStats_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ContainerStatsRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ContainerStats(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ContainerStats",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ContainerStats(ctx, req.(*ContainerStatsRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_ListContainerStats_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ListContainerStatsRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).ListContainerStats(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/ListContainerStats",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).ListContainerStats(ctx, req.(*ListContainerStatsRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_UpdateRuntimeConfig_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(UpdateRuntimeConfigRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).UpdateRuntimeConfig(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/UpdateRuntimeConfig",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).UpdateRuntimeConfig(ctx, req.(*UpdateRuntimeConfigRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _RuntimeService_Status_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(StatusRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(RuntimeServiceServer).Status(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.RuntimeService/Status",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(RuntimeServiceServer).Status(ctx, req.(*StatusRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-var _RuntimeService_serviceDesc = grpc.ServiceDesc{
- ServiceName: "runtime.v1alpha2.RuntimeService",
- HandlerType: (*RuntimeServiceServer)(nil),
- Methods: []grpc.MethodDesc{
- {
- MethodName: "Version",
- Handler: _RuntimeService_Version_Handler,
- },
- {
- MethodName: "RunPodSandbox",
- Handler: _RuntimeService_RunPodSandbox_Handler,
- },
- {
- MethodName: "StopPodSandbox",
- Handler: _RuntimeService_StopPodSandbox_Handler,
- },
- {
- MethodName: "RemovePodSandbox",
- Handler: _RuntimeService_RemovePodSandbox_Handler,
- },
- {
- MethodName: "PodSandboxStatus",
- Handler: _RuntimeService_PodSandboxStatus_Handler,
- },
- {
- MethodName: "ListPodSandbox",
- Handler: _RuntimeService_ListPodSandbox_Handler,
- },
- {
- MethodName: "CreateContainer",
- Handler: _RuntimeService_CreateContainer_Handler,
- },
- {
- MethodName: "StartContainer",
- Handler: _RuntimeService_StartContainer_Handler,
- },
- {
- MethodName: "StopContainer",
- Handler: _RuntimeService_StopContainer_Handler,
- },
- {
- MethodName: "RemoveContainer",
- Handler: _RuntimeService_RemoveContainer_Handler,
- },
- {
- MethodName: "ListContainers",
- Handler: _RuntimeService_ListContainers_Handler,
- },
- {
- MethodName: "ContainerStatus",
- Handler: _RuntimeService_ContainerStatus_Handler,
- },
- {
- MethodName: "UpdateContainerResources",
- Handler: _RuntimeService_UpdateContainerResources_Handler,
- },
- {
- MethodName: "ReopenContainerLog",
- Handler: _RuntimeService_ReopenContainerLog_Handler,
- },
- {
- MethodName: "ExecSync",
- Handler: _RuntimeService_ExecSync_Handler,
- },
- {
- MethodName: "Exec",
- Handler: _RuntimeService_Exec_Handler,
- },
- {
- MethodName: "Attach",
- Handler: _RuntimeService_Attach_Handler,
- },
- {
- MethodName: "PortForward",
- Handler: _RuntimeService_PortForward_Handler,
- },
- {
- MethodName: "ContainerStats",
- Handler: _RuntimeService_ContainerStats_Handler,
- },
- {
- MethodName: "ListContainerStats",
- Handler: _RuntimeService_ListContainerStats_Handler,
- },
- {
- MethodName: "UpdateRuntimeConfig",
- Handler: _RuntimeService_UpdateRuntimeConfig_Handler,
- },
- {
- MethodName: "Status",
- Handler: _RuntimeService_Status_Handler,
- },
- },
- Streams: []grpc.StreamDesc{},
- Metadata: "api.proto",
-}
-
-// Client API for ImageService service
-
-type ImageServiceClient interface {
- // ListImages lists existing images.
- ListImages(ctx context.Context, in *ListImagesRequest, opts ...grpc.CallOption) (*ListImagesResponse, error)
- // ImageStatus returns the status of the image. If the image is not
- // present, returns a response with ImageStatusResponse.Image set to
- // nil.
- ImageStatus(ctx context.Context, in *ImageStatusRequest, opts ...grpc.CallOption) (*ImageStatusResponse, error)
- // PullImage pulls an image with authentication config.
- PullImage(ctx context.Context, in *PullImageRequest, opts ...grpc.CallOption) (*PullImageResponse, error)
- // RemoveImage removes the image.
- // This call is idempotent, and must not return an error if the image has
- // already been removed.
- RemoveImage(ctx context.Context, in *RemoveImageRequest, opts ...grpc.CallOption) (*RemoveImageResponse, error)
- // ImageFSInfo returns information of the filesystem that is used to store images.
- ImageFsInfo(ctx context.Context, in *ImageFsInfoRequest, opts ...grpc.CallOption) (*ImageFsInfoResponse, error)
-}
-
-type imageServiceClient struct {
- cc *grpc.ClientConn
-}
-
-func NewImageServiceClient(cc *grpc.ClientConn) ImageServiceClient {
- return &imageServiceClient{cc}
-}
-
-func (c *imageServiceClient) ListImages(ctx context.Context, in *ListImagesRequest, opts ...grpc.CallOption) (*ListImagesResponse, error) {
- out := new(ListImagesResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.ImageService/ListImages", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *imageServiceClient) ImageStatus(ctx context.Context, in *ImageStatusRequest, opts ...grpc.CallOption) (*ImageStatusResponse, error) {
- out := new(ImageStatusResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.ImageService/ImageStatus", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *imageServiceClient) PullImage(ctx context.Context, in *PullImageRequest, opts ...grpc.CallOption) (*PullImageResponse, error) {
- out := new(PullImageResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.ImageService/PullImage", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *imageServiceClient) RemoveImage(ctx context.Context, in *RemoveImageRequest, opts ...grpc.CallOption) (*RemoveImageResponse, error) {
- out := new(RemoveImageResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.ImageService/RemoveImage", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-func (c *imageServiceClient) ImageFsInfo(ctx context.Context, in *ImageFsInfoRequest, opts ...grpc.CallOption) (*ImageFsInfoResponse, error) {
- out := new(ImageFsInfoResponse)
- err := grpc.Invoke(ctx, "/runtime.v1alpha2.ImageService/ImageFsInfo", in, out, c.cc, opts...)
- if err != nil {
- return nil, err
- }
- return out, nil
-}
-
-// Server API for ImageService service
-
-type ImageServiceServer interface {
- // ListImages lists existing images.
- ListImages(context.Context, *ListImagesRequest) (*ListImagesResponse, error)
- // ImageStatus returns the status of the image. If the image is not
- // present, returns a response with ImageStatusResponse.Image set to
- // nil.
- ImageStatus(context.Context, *ImageStatusRequest) (*ImageStatusResponse, error)
- // PullImage pulls an image with authentication config.
- PullImage(context.Context, *PullImageRequest) (*PullImageResponse, error)
- // RemoveImage removes the image.
- // This call is idempotent, and must not return an error if the image has
- // already been removed.
- RemoveImage(context.Context, *RemoveImageRequest) (*RemoveImageResponse, error)
- // ImageFSInfo returns information of the filesystem that is used to store images.
- ImageFsInfo(context.Context, *ImageFsInfoRequest) (*ImageFsInfoResponse, error)
-}
-
-func RegisterImageServiceServer(s *grpc.Server, srv ImageServiceServer) {
- s.RegisterService(&_ImageService_serviceDesc, srv)
-}
-
-func _ImageService_ListImages_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ListImagesRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(ImageServiceServer).ListImages(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.ImageService/ListImages",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(ImageServiceServer).ListImages(ctx, req.(*ListImagesRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _ImageService_ImageStatus_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ImageStatusRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(ImageServiceServer).ImageStatus(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.ImageService/ImageStatus",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(ImageServiceServer).ImageStatus(ctx, req.(*ImageStatusRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _ImageService_PullImage_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(PullImageRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(ImageServiceServer).PullImage(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.ImageService/PullImage",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(ImageServiceServer).PullImage(ctx, req.(*PullImageRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _ImageService_RemoveImage_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(RemoveImageRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(ImageServiceServer).RemoveImage(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.ImageService/RemoveImage",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(ImageServiceServer).RemoveImage(ctx, req.(*RemoveImageRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-func _ImageService_ImageFsInfo_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
- in := new(ImageFsInfoRequest)
- if err := dec(in); err != nil {
- return nil, err
- }
- if interceptor == nil {
- return srv.(ImageServiceServer).ImageFsInfo(ctx, in)
- }
- info := &grpc.UnaryServerInfo{
- Server: srv,
- FullMethod: "/runtime.v1alpha2.ImageService/ImageFsInfo",
- }
- handler := func(ctx context.Context, req interface{}) (interface{}, error) {
- return srv.(ImageServiceServer).ImageFsInfo(ctx, req.(*ImageFsInfoRequest))
- }
- return interceptor(ctx, in, info, handler)
-}
-
-var _ImageService_serviceDesc = grpc.ServiceDesc{
- ServiceName: "runtime.v1alpha2.ImageService",
- HandlerType: (*ImageServiceServer)(nil),
- Methods: []grpc.MethodDesc{
- {
- MethodName: "ListImages",
- Handler: _ImageService_ListImages_Handler,
- },
- {
- MethodName: "ImageStatus",
- Handler: _ImageService_ImageStatus_Handler,
- },
- {
- MethodName: "PullImage",
- Handler: _ImageService_PullImage_Handler,
- },
- {
- MethodName: "RemoveImage",
- Handler: _ImageService_RemoveImage_Handler,
- },
- {
- MethodName: "ImageFsInfo",
- Handler: _ImageService_ImageFsInfo_Handler,
- },
- },
- Streams: []grpc.StreamDesc{},
- Metadata: "api.proto",
-}
-
-func (m *VersionRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *VersionRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Version) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Version)))
- i += copy(dAtA[i:], m.Version)
- }
- return i, nil
-}
-
-func (m *VersionResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *VersionResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Version) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Version)))
- i += copy(dAtA[i:], m.Version)
- }
- if len(m.RuntimeName) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.RuntimeName)))
- i += copy(dAtA[i:], m.RuntimeName)
- }
- if len(m.RuntimeVersion) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.RuntimeVersion)))
- i += copy(dAtA[i:], m.RuntimeVersion)
- }
- if len(m.RuntimeApiVersion) > 0 {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.RuntimeApiVersion)))
- i += copy(dAtA[i:], m.RuntimeApiVersion)
- }
- return i, nil
-}
-
-func (m *DNSConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *DNSConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Servers) > 0 {
- for _, s := range m.Servers {
- dAtA[i] = 0xa
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.Searches) > 0 {
- for _, s := range m.Searches {
- dAtA[i] = 0x12
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.Options) > 0 {
- for _, s := range m.Options {
- dAtA[i] = 0x1a
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- return i, nil
-}
-
-func (m *PortMapping) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PortMapping) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Protocol != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Protocol))
- }
- if m.ContainerPort != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.ContainerPort))
- }
- if m.HostPort != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.HostPort))
- }
- if len(m.HostIp) > 0 {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.HostIp)))
- i += copy(dAtA[i:], m.HostIp)
- }
- return i, nil
-}
-
-func (m *Mount) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Mount) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerPath) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerPath)))
- i += copy(dAtA[i:], m.ContainerPath)
- }
- if len(m.HostPath) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.HostPath)))
- i += copy(dAtA[i:], m.HostPath)
- }
- if m.Readonly {
- dAtA[i] = 0x18
- i++
- if m.Readonly {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.SelinuxRelabel {
- dAtA[i] = 0x20
- i++
- if m.SelinuxRelabel {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Propagation != 0 {
- dAtA[i] = 0x28
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Propagation))
- }
- return i, nil
-}
-
-func (m *NamespaceOption) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *NamespaceOption) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Network != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Network))
- }
- if m.Pid != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Pid))
- }
- if m.Ipc != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Ipc))
- }
- return i, nil
-}
-
-func (m *Int64Value) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Int64Value) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Value != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Value))
- }
- return i, nil
-}
-
-func (m *LinuxSandboxSecurityContext) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxSandboxSecurityContext) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.NamespaceOptions != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.NamespaceOptions.Size()))
- n1, err := m.NamespaceOptions.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n1
- }
- if m.SelinuxOptions != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SelinuxOptions.Size()))
- n2, err := m.SelinuxOptions.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n2
- }
- if m.RunAsUser != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.RunAsUser.Size()))
- n3, err := m.RunAsUser.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n3
- }
- if m.ReadonlyRootfs {
- dAtA[i] = 0x20
- i++
- if m.ReadonlyRootfs {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if len(m.SupplementalGroups) > 0 {
- dAtA5 := make([]byte, len(m.SupplementalGroups)*10)
- var j4 int
- for _, num1 := range m.SupplementalGroups {
- num := uint64(num1)
- for num >= 1<<7 {
- dAtA5[j4] = uint8(uint64(num)&0x7f | 0x80)
- num >>= 7
- j4++
- }
- dAtA5[j4] = uint8(num)
- j4++
- }
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(j4))
- i += copy(dAtA[i:], dAtA5[:j4])
- }
- if m.Privileged {
- dAtA[i] = 0x30
- i++
- if m.Privileged {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if len(m.SeccompProfilePath) > 0 {
- dAtA[i] = 0x3a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.SeccompProfilePath)))
- i += copy(dAtA[i:], m.SeccompProfilePath)
- }
- return i, nil
-}
-
-func (m *LinuxPodSandboxConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxPodSandboxConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.CgroupParent) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.CgroupParent)))
- i += copy(dAtA[i:], m.CgroupParent)
- }
- if m.SecurityContext != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SecurityContext.Size()))
- n6, err := m.SecurityContext.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n6
- }
- if len(m.Sysctls) > 0 {
- for k := range m.Sysctls {
- dAtA[i] = 0x1a
- i++
- v := m.Sysctls[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *PodSandboxMetadata) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxMetadata) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Name) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Name)))
- i += copy(dAtA[i:], m.Name)
- }
- if len(m.Uid) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Uid)))
- i += copy(dAtA[i:], m.Uid)
- }
- if len(m.Namespace) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Namespace)))
- i += copy(dAtA[i:], m.Namespace)
- }
- if m.Attempt != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Attempt))
- }
- return i, nil
-}
-
-func (m *PodSandboxConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Metadata != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n7, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n7
- }
- if len(m.Hostname) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Hostname)))
- i += copy(dAtA[i:], m.Hostname)
- }
- if len(m.LogDirectory) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.LogDirectory)))
- i += copy(dAtA[i:], m.LogDirectory)
- }
- if m.DnsConfig != nil {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.DnsConfig.Size()))
- n8, err := m.DnsConfig.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n8
- }
- if len(m.PortMappings) > 0 {
- for _, msg := range m.PortMappings {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x32
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x3a
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if m.Linux != nil {
- dAtA[i] = 0x42
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Linux.Size()))
- n9, err := m.Linux.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n9
- }
- return i, nil
-}
-
-func (m *RunPodSandboxRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RunPodSandboxRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Config != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Config.Size()))
- n10, err := m.Config.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n10
- }
- return i, nil
-}
-
-func (m *RunPodSandboxResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RunPodSandboxResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- return i, nil
-}
-
-func (m *StopPodSandboxRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StopPodSandboxRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- return i, nil
-}
-
-func (m *StopPodSandboxResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StopPodSandboxResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *RemovePodSandboxRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemovePodSandboxRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- return i, nil
-}
-
-func (m *RemovePodSandboxResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemovePodSandboxResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *PodSandboxStatusRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxStatusRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if m.Verbose {
- dAtA[i] = 0x10
- i++
- if m.Verbose {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *PodSandboxNetworkStatus) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxNetworkStatus) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Ip) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Ip)))
- i += copy(dAtA[i:], m.Ip)
- }
- return i, nil
-}
-
-func (m *Namespace) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Namespace) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Options != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Options.Size()))
- n11, err := m.Options.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n11
- }
- return i, nil
-}
-
-func (m *LinuxPodSandboxStatus) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxPodSandboxStatus) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Namespaces != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Namespaces.Size()))
- n12, err := m.Namespaces.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n12
- }
- return i, nil
-}
-
-func (m *PodSandboxStatus) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxStatus) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.Metadata != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n13, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n13
- }
- if m.State != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- if m.CreatedAt != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CreatedAt))
- }
- if m.Network != nil {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Network.Size()))
- n14, err := m.Network.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n14
- }
- if m.Linux != nil {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Linux.Size()))
- n15, err := m.Linux.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n15
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x3a
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x42
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *PodSandboxStatusResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxStatusResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Status != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Status.Size()))
- n16, err := m.Status.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n16
- }
- if len(m.Info) > 0 {
- for k := range m.Info {
- dAtA[i] = 0x12
- i++
- v := m.Info[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *PodSandboxStateValue) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxStateValue) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.State != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- return i, nil
-}
-
-func (m *PodSandboxFilter) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandboxFilter) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.State != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State.Size()))
- n17, err := m.State.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n17
- }
- if len(m.LabelSelector) > 0 {
- for k := range m.LabelSelector {
- dAtA[i] = 0x1a
- i++
- v := m.LabelSelector[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ListPodSandboxRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListPodSandboxRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Filter != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Filter.Size()))
- n18, err := m.Filter.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n18
- }
- return i, nil
-}
-
-func (m *PodSandbox) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PodSandbox) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.Metadata != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n19, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n19
- }
- if m.State != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- if m.CreatedAt != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CreatedAt))
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x2a
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x32
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ListPodSandboxResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListPodSandboxResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Items) > 0 {
- for _, msg := range m.Items {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *ImageSpec) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageSpec) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Image) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Image)))
- i += copy(dAtA[i:], m.Image)
- }
- return i, nil
-}
-
-func (m *KeyValue) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *KeyValue) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Key) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Key)))
- i += copy(dAtA[i:], m.Key)
- }
- if len(m.Value) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Value)))
- i += copy(dAtA[i:], m.Value)
- }
- return i, nil
-}
-
-func (m *LinuxContainerResources) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxContainerResources) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.CpuPeriod != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuPeriod))
- }
- if m.CpuQuota != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuQuota))
- }
- if m.CpuShares != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuShares))
- }
- if m.MemoryLimitInBytes != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.MemoryLimitInBytes))
- }
- if m.OomScoreAdj != 0 {
- dAtA[i] = 0x28
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.OomScoreAdj))
- }
- if len(m.CpusetCpus) > 0 {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.CpusetCpus)))
- i += copy(dAtA[i:], m.CpusetCpus)
- }
- if len(m.CpusetMems) > 0 {
- dAtA[i] = 0x3a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.CpusetMems)))
- i += copy(dAtA[i:], m.CpusetMems)
- }
- return i, nil
-}
-
-func (m *SELinuxOption) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *SELinuxOption) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.User) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.User)))
- i += copy(dAtA[i:], m.User)
- }
- if len(m.Role) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Role)))
- i += copy(dAtA[i:], m.Role)
- }
- if len(m.Type) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Type)))
- i += copy(dAtA[i:], m.Type)
- }
- if len(m.Level) > 0 {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Level)))
- i += copy(dAtA[i:], m.Level)
- }
- return i, nil
-}
-
-func (m *Capability) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Capability) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.AddCapabilities) > 0 {
- for _, s := range m.AddCapabilities {
- dAtA[i] = 0xa
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.DropCapabilities) > 0 {
- for _, s := range m.DropCapabilities {
- dAtA[i] = 0x12
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- return i, nil
-}
-
-func (m *LinuxContainerSecurityContext) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxContainerSecurityContext) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Capabilities != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Capabilities.Size()))
- n20, err := m.Capabilities.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n20
- }
- if m.Privileged {
- dAtA[i] = 0x10
- i++
- if m.Privileged {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.NamespaceOptions != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.NamespaceOptions.Size()))
- n21, err := m.NamespaceOptions.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n21
- }
- if m.SelinuxOptions != nil {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SelinuxOptions.Size()))
- n22, err := m.SelinuxOptions.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n22
- }
- if m.RunAsUser != nil {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.RunAsUser.Size()))
- n23, err := m.RunAsUser.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n23
- }
- if len(m.RunAsUsername) > 0 {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.RunAsUsername)))
- i += copy(dAtA[i:], m.RunAsUsername)
- }
- if m.ReadonlyRootfs {
- dAtA[i] = 0x38
- i++
- if m.ReadonlyRootfs {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if len(m.SupplementalGroups) > 0 {
- dAtA25 := make([]byte, len(m.SupplementalGroups)*10)
- var j24 int
- for _, num1 := range m.SupplementalGroups {
- num := uint64(num1)
- for num >= 1<<7 {
- dAtA25[j24] = uint8(uint64(num)&0x7f | 0x80)
- num >>= 7
- j24++
- }
- dAtA25[j24] = uint8(num)
- j24++
- }
- dAtA[i] = 0x42
- i++
- i = encodeVarintApi(dAtA, i, uint64(j24))
- i += copy(dAtA[i:], dAtA25[:j24])
- }
- if len(m.ApparmorProfile) > 0 {
- dAtA[i] = 0x4a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ApparmorProfile)))
- i += copy(dAtA[i:], m.ApparmorProfile)
- }
- if len(m.SeccompProfilePath) > 0 {
- dAtA[i] = 0x52
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.SeccompProfilePath)))
- i += copy(dAtA[i:], m.SeccompProfilePath)
- }
- if m.NoNewPrivs {
- dAtA[i] = 0x58
- i++
- if m.NoNewPrivs {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *LinuxContainerConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *LinuxContainerConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Resources != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Resources.Size()))
- n26, err := m.Resources.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n26
- }
- if m.SecurityContext != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SecurityContext.Size()))
- n27, err := m.SecurityContext.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n27
- }
- return i, nil
-}
-
-func (m *WindowsContainerConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *WindowsContainerConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Resources != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Resources.Size()))
- n28, err := m.Resources.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n28
- }
- return i, nil
-}
-
-func (m *WindowsContainerResources) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *WindowsContainerResources) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.CpuShares != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuShares))
- }
- if m.CpuCount != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuCount))
- }
- if m.CpuMaximum != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CpuMaximum))
- }
- if m.MemoryLimitInBytes != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.MemoryLimitInBytes))
- }
- return i, nil
-}
-
-func (m *ContainerMetadata) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerMetadata) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Name) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Name)))
- i += copy(dAtA[i:], m.Name)
- }
- if m.Attempt != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Attempt))
- }
- return i, nil
-}
-
-func (m *Device) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Device) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerPath) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerPath)))
- i += copy(dAtA[i:], m.ContainerPath)
- }
- if len(m.HostPath) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.HostPath)))
- i += copy(dAtA[i:], m.HostPath)
- }
- if len(m.Permissions) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Permissions)))
- i += copy(dAtA[i:], m.Permissions)
- }
- return i, nil
-}
-
-func (m *ContainerConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Metadata != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n29, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n29
- }
- if m.Image != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n30, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n30
- }
- if len(m.Command) > 0 {
- for _, s := range m.Command {
- dAtA[i] = 0x1a
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.Args) > 0 {
- for _, s := range m.Args {
- dAtA[i] = 0x22
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.WorkingDir) > 0 {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.WorkingDir)))
- i += copy(dAtA[i:], m.WorkingDir)
- }
- if len(m.Envs) > 0 {
- for _, msg := range m.Envs {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- if len(m.Mounts) > 0 {
- for _, msg := range m.Mounts {
- dAtA[i] = 0x3a
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- if len(m.Devices) > 0 {
- for _, msg := range m.Devices {
- dAtA[i] = 0x42
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x4a
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x52
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.LogPath) > 0 {
- dAtA[i] = 0x5a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.LogPath)))
- i += copy(dAtA[i:], m.LogPath)
- }
- if m.Stdin {
- dAtA[i] = 0x60
- i++
- if m.Stdin {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.StdinOnce {
- dAtA[i] = 0x68
- i++
- if m.StdinOnce {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Tty {
- dAtA[i] = 0x70
- i++
- if m.Tty {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Linux != nil {
- dAtA[i] = 0x7a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Linux.Size()))
- n31, err := m.Linux.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n31
- }
- if m.Windows != nil {
- dAtA[i] = 0x82
- i++
- dAtA[i] = 0x1
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Windows.Size()))
- n32, err := m.Windows.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n32
- }
- return i, nil
-}
-
-func (m *CreateContainerRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *CreateContainerRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if m.Config != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Config.Size()))
- n33, err := m.Config.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n33
- }
- if m.SandboxConfig != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SandboxConfig.Size()))
- n34, err := m.SandboxConfig.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n34
- }
- return i, nil
-}
-
-func (m *CreateContainerResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *CreateContainerResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- return i, nil
-}
-
-func (m *StartContainerRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StartContainerRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- return i, nil
-}
-
-func (m *StartContainerResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StartContainerResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *StopContainerRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StopContainerRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if m.Timeout != 0 {
- dAtA[i] = 0x10
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Timeout))
- }
- return i, nil
-}
-
-func (m *StopContainerResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StopContainerResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *RemoveContainerRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemoveContainerRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- return i, nil
-}
-
-func (m *RemoveContainerResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemoveContainerResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *ContainerStateValue) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStateValue) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.State != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- return i, nil
-}
-
-func (m *ContainerFilter) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerFilter) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.State != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State.Size()))
- n35, err := m.State.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n35
- }
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if len(m.LabelSelector) > 0 {
- for k := range m.LabelSelector {
- dAtA[i] = 0x22
- i++
- v := m.LabelSelector[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ListContainersRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListContainersRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Filter != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Filter.Size()))
- n36, err := m.Filter.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n36
- }
- return i, nil
-}
-
-func (m *Container) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Container) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if m.Metadata != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n37, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n37
- }
- if m.Image != nil {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n38, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n38
- }
- if len(m.ImageRef) > 0 {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ImageRef)))
- i += copy(dAtA[i:], m.ImageRef)
- }
- if m.State != 0 {
- dAtA[i] = 0x30
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- if m.CreatedAt != 0 {
- dAtA[i] = 0x38
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CreatedAt))
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x42
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x4a
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ListContainersResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListContainersResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Containers) > 0 {
- for _, msg := range m.Containers {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *ContainerStatusRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatusRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if m.Verbose {
- dAtA[i] = 0x10
- i++
- if m.Verbose {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *ContainerStatus) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatus) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.Metadata != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n39, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n39
- }
- if m.State != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.State))
- }
- if m.CreatedAt != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.CreatedAt))
- }
- if m.StartedAt != 0 {
- dAtA[i] = 0x28
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.StartedAt))
- }
- if m.FinishedAt != 0 {
- dAtA[i] = 0x30
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.FinishedAt))
- }
- if m.ExitCode != 0 {
- dAtA[i] = 0x38
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.ExitCode))
- }
- if m.Image != nil {
- dAtA[i] = 0x42
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n40, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n40
- }
- if len(m.ImageRef) > 0 {
- dAtA[i] = 0x4a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ImageRef)))
- i += copy(dAtA[i:], m.ImageRef)
- }
- if len(m.Reason) > 0 {
- dAtA[i] = 0x52
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Reason)))
- i += copy(dAtA[i:], m.Reason)
- }
- if len(m.Message) > 0 {
- dAtA[i] = 0x5a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Message)))
- i += copy(dAtA[i:], m.Message)
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x62
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x6a
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Mounts) > 0 {
- for _, msg := range m.Mounts {
- dAtA[i] = 0x72
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- if len(m.LogPath) > 0 {
- dAtA[i] = 0x7a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.LogPath)))
- i += copy(dAtA[i:], m.LogPath)
- }
- return i, nil
-}
-
-func (m *ContainerStatusResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatusResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Status != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Status.Size()))
- n41, err := m.Status.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n41
- }
- if len(m.Info) > 0 {
- for k := range m.Info {
- dAtA[i] = 0x12
- i++
- v := m.Info[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *UpdateContainerResourcesRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *UpdateContainerResourcesRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if m.Linux != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Linux.Size()))
- n42, err := m.Linux.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n42
- }
- return i, nil
-}
-
-func (m *UpdateContainerResourcesResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *UpdateContainerResourcesResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *ExecSyncRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ExecSyncRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if len(m.Cmd) > 0 {
- for _, s := range m.Cmd {
- dAtA[i] = 0x12
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if m.Timeout != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Timeout))
- }
- return i, nil
-}
-
-func (m *ExecSyncResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ExecSyncResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Stdout) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Stdout)))
- i += copy(dAtA[i:], m.Stdout)
- }
- if len(m.Stderr) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Stderr)))
- i += copy(dAtA[i:], m.Stderr)
- }
- if m.ExitCode != 0 {
- dAtA[i] = 0x18
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.ExitCode))
- }
- return i, nil
-}
-
-func (m *ExecRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ExecRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if len(m.Cmd) > 0 {
- for _, s := range m.Cmd {
- dAtA[i] = 0x12
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if m.Tty {
- dAtA[i] = 0x18
- i++
- if m.Tty {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Stdin {
- dAtA[i] = 0x20
- i++
- if m.Stdin {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Stdout {
- dAtA[i] = 0x28
- i++
- if m.Stdout {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Stderr {
- dAtA[i] = 0x30
- i++
- if m.Stderr {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *ExecResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ExecResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Url) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Url)))
- i += copy(dAtA[i:], m.Url)
- }
- return i, nil
-}
-
-func (m *AttachRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *AttachRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- if m.Stdin {
- dAtA[i] = 0x10
- i++
- if m.Stdin {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Tty {
- dAtA[i] = 0x18
- i++
- if m.Tty {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Stdout {
- dAtA[i] = 0x20
- i++
- if m.Stdout {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if m.Stderr {
- dAtA[i] = 0x28
- i++
- if m.Stderr {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *AttachResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *AttachResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Url) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Url)))
- i += copy(dAtA[i:], m.Url)
- }
- return i, nil
-}
-
-func (m *PortForwardRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PortForwardRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if len(m.Port) > 0 {
- dAtA44 := make([]byte, len(m.Port)*10)
- var j43 int
- for _, num1 := range m.Port {
- num := uint64(num1)
- for num >= 1<<7 {
- dAtA44[j43] = uint8(uint64(num)&0x7f | 0x80)
- num >>= 7
- j43++
- }
- dAtA44[j43] = uint8(num)
- j43++
- }
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(j43))
- i += copy(dAtA[i:], dAtA44[:j43])
- }
- return i, nil
-}
-
-func (m *PortForwardResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PortForwardResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Url) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Url)))
- i += copy(dAtA[i:], m.Url)
- }
- return i, nil
-}
-
-func (m *ImageFilter) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageFilter) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Image != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n45, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n45
- }
- return i, nil
-}
-
-func (m *ListImagesRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListImagesRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Filter != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Filter.Size()))
- n46, err := m.Filter.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n46
- }
- return i, nil
-}
-
-func (m *Image) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *Image) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if len(m.RepoTags) > 0 {
- for _, s := range m.RepoTags {
- dAtA[i] = 0x12
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if len(m.RepoDigests) > 0 {
- for _, s := range m.RepoDigests {
- dAtA[i] = 0x1a
- i++
- l = len(s)
- for l >= 1<<7 {
- dAtA[i] = uint8(uint64(l)&0x7f | 0x80)
- l >>= 7
- i++
- }
- dAtA[i] = uint8(l)
- i++
- i += copy(dAtA[i:], s)
- }
- }
- if m.Size_ != 0 {
- dAtA[i] = 0x20
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Size_))
- }
- if m.Uid != nil {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Uid.Size()))
- n47, err := m.Uid.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n47
- }
- if len(m.Username) > 0 {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Username)))
- i += copy(dAtA[i:], m.Username)
- }
- return i, nil
-}
-
-func (m *ListImagesResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListImagesResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Images) > 0 {
- for _, msg := range m.Images {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *ImageStatusRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageStatusRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Image != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n48, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n48
- }
- if m.Verbose {
- dAtA[i] = 0x10
- i++
- if m.Verbose {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *ImageStatusResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageStatusResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Image != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n49, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n49
- }
- if len(m.Info) > 0 {
- for k := range m.Info {
- dAtA[i] = 0x12
- i++
- v := m.Info[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *AuthConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *AuthConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Username) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Username)))
- i += copy(dAtA[i:], m.Username)
- }
- if len(m.Password) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Password)))
- i += copy(dAtA[i:], m.Password)
- }
- if len(m.Auth) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Auth)))
- i += copy(dAtA[i:], m.Auth)
- }
- if len(m.ServerAddress) > 0 {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ServerAddress)))
- i += copy(dAtA[i:], m.ServerAddress)
- }
- if len(m.IdentityToken) > 0 {
- dAtA[i] = 0x2a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.IdentityToken)))
- i += copy(dAtA[i:], m.IdentityToken)
- }
- if len(m.RegistryToken) > 0 {
- dAtA[i] = 0x32
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.RegistryToken)))
- i += copy(dAtA[i:], m.RegistryToken)
- }
- return i, nil
-}
-
-func (m *PullImageRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PullImageRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Image != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n50, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n50
- }
- if m.Auth != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Auth.Size()))
- n51, err := m.Auth.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n51
- }
- if m.SandboxConfig != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.SandboxConfig.Size()))
- n52, err := m.SandboxConfig.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n52
- }
- return i, nil
-}
-
-func (m *PullImageResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *PullImageResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ImageRef) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ImageRef)))
- i += copy(dAtA[i:], m.ImageRef)
- }
- return i, nil
-}
-
-func (m *RemoveImageRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemoveImageRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Image != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Image.Size()))
- n53, err := m.Image.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n53
- }
- return i, nil
-}
-
-func (m *RemoveImageResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RemoveImageResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *NetworkConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *NetworkConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.PodCidr) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodCidr)))
- i += copy(dAtA[i:], m.PodCidr)
- }
- return i, nil
-}
-
-func (m *RuntimeConfig) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RuntimeConfig) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.NetworkConfig != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.NetworkConfig.Size()))
- n54, err := m.NetworkConfig.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n54
- }
- return i, nil
-}
-
-func (m *UpdateRuntimeConfigRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *UpdateRuntimeConfigRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.RuntimeConfig != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.RuntimeConfig.Size()))
- n55, err := m.RuntimeConfig.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n55
- }
- return i, nil
-}
-
-func (m *UpdateRuntimeConfigResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *UpdateRuntimeConfigResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *RuntimeCondition) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RuntimeCondition) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Type) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Type)))
- i += copy(dAtA[i:], m.Type)
- }
- if m.Status {
- dAtA[i] = 0x10
- i++
- if m.Status {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- if len(m.Reason) > 0 {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Reason)))
- i += copy(dAtA[i:], m.Reason)
- }
- if len(m.Message) > 0 {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Message)))
- i += copy(dAtA[i:], m.Message)
- }
- return i, nil
-}
-
-func (m *RuntimeStatus) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *RuntimeStatus) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Conditions) > 0 {
- for _, msg := range m.Conditions {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *StatusRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StatusRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Verbose {
- dAtA[i] = 0x8
- i++
- if m.Verbose {
- dAtA[i] = 1
- } else {
- dAtA[i] = 0
- }
- i++
- }
- return i, nil
-}
-
-func (m *StatusResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *StatusResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Status != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Status.Size()))
- n56, err := m.Status.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n56
- }
- if len(m.Info) > 0 {
- for k := range m.Info {
- dAtA[i] = 0x12
- i++
- v := m.Info[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ImageFsInfoRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageFsInfoRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func (m *UInt64Value) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *UInt64Value) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Value != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Value))
- }
- return i, nil
-}
-
-func (m *FilesystemIdentifier) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *FilesystemIdentifier) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Mountpoint) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Mountpoint)))
- i += copy(dAtA[i:], m.Mountpoint)
- }
- return i, nil
-}
-
-func (m *FilesystemUsage) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *FilesystemUsage) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Timestamp != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Timestamp))
- }
- if m.FsId != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.FsId.Size()))
- n57, err := m.FsId.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n57
- }
- if m.UsedBytes != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.UsedBytes.Size()))
- n58, err := m.UsedBytes.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n58
- }
- if m.InodesUsed != nil {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.InodesUsed.Size()))
- n59, err := m.InodesUsed.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n59
- }
- return i, nil
-}
-
-func (m *ImageFsInfoResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ImageFsInfoResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ImageFilesystems) > 0 {
- for _, msg := range m.ImageFilesystems {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *ContainerStatsRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatsRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- return i, nil
-}
-
-func (m *ContainerStatsResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatsResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Stats != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Stats.Size()))
- n60, err := m.Stats.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n60
- }
- return i, nil
-}
-
-func (m *ListContainerStatsRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListContainerStatsRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Filter != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Filter.Size()))
- n61, err := m.Filter.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n61
- }
- return i, nil
-}
-
-func (m *ContainerStatsFilter) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStatsFilter) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if len(m.PodSandboxId) > 0 {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.PodSandboxId)))
- i += copy(dAtA[i:], m.PodSandboxId)
- }
- if len(m.LabelSelector) > 0 {
- for k := range m.LabelSelector {
- dAtA[i] = 0x1a
- i++
- v := m.LabelSelector[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ListContainerStatsResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ListContainerStatsResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Stats) > 0 {
- for _, msg := range m.Stats {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(msg.Size()))
- n, err := msg.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n
- }
- }
- return i, nil
-}
-
-func (m *ContainerAttributes) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerAttributes) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.Id) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.Id)))
- i += copy(dAtA[i:], m.Id)
- }
- if m.Metadata != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Metadata.Size()))
- n62, err := m.Metadata.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n62
- }
- if len(m.Labels) > 0 {
- for k := range m.Labels {
- dAtA[i] = 0x1a
- i++
- v := m.Labels[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- if len(m.Annotations) > 0 {
- for k := range m.Annotations {
- dAtA[i] = 0x22
- i++
- v := m.Annotations[k]
- mapSize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- i = encodeVarintApi(dAtA, i, uint64(mapSize))
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(k)))
- i += copy(dAtA[i:], k)
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(v)))
- i += copy(dAtA[i:], v)
- }
- }
- return i, nil
-}
-
-func (m *ContainerStats) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ContainerStats) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Attributes != nil {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Attributes.Size()))
- n63, err := m.Attributes.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n63
- }
- if m.Cpu != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Cpu.Size()))
- n64, err := m.Cpu.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n64
- }
- if m.Memory != nil {
- dAtA[i] = 0x1a
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Memory.Size()))
- n65, err := m.Memory.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n65
- }
- if m.WritableLayer != nil {
- dAtA[i] = 0x22
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.WritableLayer.Size()))
- n66, err := m.WritableLayer.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n66
- }
- return i, nil
-}
-
-func (m *CpuUsage) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *CpuUsage) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Timestamp != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Timestamp))
- }
- if m.UsageCoreNanoSeconds != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.UsageCoreNanoSeconds.Size()))
- n67, err := m.UsageCoreNanoSeconds.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n67
- }
- return i, nil
-}
-
-func (m *MemoryUsage) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *MemoryUsage) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if m.Timestamp != 0 {
- dAtA[i] = 0x8
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.Timestamp))
- }
- if m.WorkingSetBytes != nil {
- dAtA[i] = 0x12
- i++
- i = encodeVarintApi(dAtA, i, uint64(m.WorkingSetBytes.Size()))
- n68, err := m.WorkingSetBytes.MarshalTo(dAtA[i:])
- if err != nil {
- return 0, err
- }
- i += n68
- }
- return i, nil
-}
-
-func (m *ReopenContainerLogRequest) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ReopenContainerLogRequest) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- if len(m.ContainerId) > 0 {
- dAtA[i] = 0xa
- i++
- i = encodeVarintApi(dAtA, i, uint64(len(m.ContainerId)))
- i += copy(dAtA[i:], m.ContainerId)
- }
- return i, nil
-}
-
-func (m *ReopenContainerLogResponse) Marshal() (dAtA []byte, err error) {
- size := m.Size()
- dAtA = make([]byte, size)
- n, err := m.MarshalTo(dAtA)
- if err != nil {
- return nil, err
- }
- return dAtA[:n], nil
-}
-
-func (m *ReopenContainerLogResponse) MarshalTo(dAtA []byte) (int, error) {
- var i int
- _ = i
- var l int
- _ = l
- return i, nil
-}
-
-func encodeFixed64Api(dAtA []byte, offset int, v uint64) int {
- dAtA[offset] = uint8(v)
- dAtA[offset+1] = uint8(v >> 8)
- dAtA[offset+2] = uint8(v >> 16)
- dAtA[offset+3] = uint8(v >> 24)
- dAtA[offset+4] = uint8(v >> 32)
- dAtA[offset+5] = uint8(v >> 40)
- dAtA[offset+6] = uint8(v >> 48)
- dAtA[offset+7] = uint8(v >> 56)
- return offset + 8
-}
-func encodeFixed32Api(dAtA []byte, offset int, v uint32) int {
- dAtA[offset] = uint8(v)
- dAtA[offset+1] = uint8(v >> 8)
- dAtA[offset+2] = uint8(v >> 16)
- dAtA[offset+3] = uint8(v >> 24)
- return offset + 4
-}
-func encodeVarintApi(dAtA []byte, offset int, v uint64) int {
- for v >= 1<<7 {
- dAtA[offset] = uint8(v&0x7f | 0x80)
- v >>= 7
- offset++
- }
- dAtA[offset] = uint8(v)
- return offset + 1
-}
-func (m *VersionRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.Version)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *VersionResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.Version)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.RuntimeName)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.RuntimeVersion)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.RuntimeApiVersion)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *DNSConfig) Size() (n int) {
- var l int
- _ = l
- if len(m.Servers) > 0 {
- for _, s := range m.Servers {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Searches) > 0 {
- for _, s := range m.Searches {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Options) > 0 {
- for _, s := range m.Options {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *PortMapping) Size() (n int) {
- var l int
- _ = l
- if m.Protocol != 0 {
- n += 1 + sovApi(uint64(m.Protocol))
- }
- if m.ContainerPort != 0 {
- n += 1 + sovApi(uint64(m.ContainerPort))
- }
- if m.HostPort != 0 {
- n += 1 + sovApi(uint64(m.HostPort))
- }
- l = len(m.HostIp)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *Mount) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.HostPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Readonly {
- n += 2
- }
- if m.SelinuxRelabel {
- n += 2
- }
- if m.Propagation != 0 {
- n += 1 + sovApi(uint64(m.Propagation))
- }
- return n
-}
-
-func (m *NamespaceOption) Size() (n int) {
- var l int
- _ = l
- if m.Network != 0 {
- n += 1 + sovApi(uint64(m.Network))
- }
- if m.Pid != 0 {
- n += 1 + sovApi(uint64(m.Pid))
- }
- if m.Ipc != 0 {
- n += 1 + sovApi(uint64(m.Ipc))
- }
- return n
-}
-
-func (m *Int64Value) Size() (n int) {
- var l int
- _ = l
- if m.Value != 0 {
- n += 1 + sovApi(uint64(m.Value))
- }
- return n
-}
-
-func (m *LinuxSandboxSecurityContext) Size() (n int) {
- var l int
- _ = l
- if m.NamespaceOptions != nil {
- l = m.NamespaceOptions.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SelinuxOptions != nil {
- l = m.SelinuxOptions.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.RunAsUser != nil {
- l = m.RunAsUser.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.ReadonlyRootfs {
- n += 2
- }
- if len(m.SupplementalGroups) > 0 {
- l = 0
- for _, e := range m.SupplementalGroups {
- l += sovApi(uint64(e))
- }
- n += 1 + sovApi(uint64(l)) + l
- }
- if m.Privileged {
- n += 2
- }
- l = len(m.SeccompProfilePath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *LinuxPodSandboxConfig) Size() (n int) {
- var l int
- _ = l
- l = len(m.CgroupParent)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SecurityContext != nil {
- l = m.SecurityContext.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Sysctls) > 0 {
- for k, v := range m.Sysctls {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *PodSandboxMetadata) Size() (n int) {
- var l int
- _ = l
- l = len(m.Name)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Uid)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Namespace)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Attempt != 0 {
- n += 1 + sovApi(uint64(m.Attempt))
- }
- return n
-}
-
-func (m *PodSandboxConfig) Size() (n int) {
- var l int
- _ = l
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Hostname)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.LogDirectory)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.DnsConfig != nil {
- l = m.DnsConfig.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.PortMappings) > 0 {
- for _, e := range m.PortMappings {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if m.Linux != nil {
- l = m.Linux.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RunPodSandboxRequest) Size() (n int) {
- var l int
- _ = l
- if m.Config != nil {
- l = m.Config.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RunPodSandboxResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *StopPodSandboxRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *StopPodSandboxResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *RemovePodSandboxRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RemovePodSandboxResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *PodSandboxStatusRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Verbose {
- n += 2
- }
- return n
-}
-
-func (m *PodSandboxNetworkStatus) Size() (n int) {
- var l int
- _ = l
- l = len(m.Ip)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *Namespace) Size() (n int) {
- var l int
- _ = l
- if m.Options != nil {
- l = m.Options.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *LinuxPodSandboxStatus) Size() (n int) {
- var l int
- _ = l
- if m.Namespaces != nil {
- l = m.Namespaces.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *PodSandboxStatus) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- if m.CreatedAt != 0 {
- n += 1 + sovApi(uint64(m.CreatedAt))
- }
- if m.Network != nil {
- l = m.Network.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Linux != nil {
- l = m.Linux.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *PodSandboxStatusResponse) Size() (n int) {
- var l int
- _ = l
- if m.Status != nil {
- l = m.Status.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Info) > 0 {
- for k, v := range m.Info {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *PodSandboxStateValue) Size() (n int) {
- var l int
- _ = l
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- return n
-}
-
-func (m *PodSandboxFilter) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != nil {
- l = m.State.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.LabelSelector) > 0 {
- for k, v := range m.LabelSelector {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ListPodSandboxRequest) Size() (n int) {
- var l int
- _ = l
- if m.Filter != nil {
- l = m.Filter.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *PodSandbox) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- if m.CreatedAt != 0 {
- n += 1 + sovApi(uint64(m.CreatedAt))
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ListPodSandboxResponse) Size() (n int) {
- var l int
- _ = l
- if len(m.Items) > 0 {
- for _, e := range m.Items {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *ImageSpec) Size() (n int) {
- var l int
- _ = l
- l = len(m.Image)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *KeyValue) Size() (n int) {
- var l int
- _ = l
- l = len(m.Key)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Value)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *LinuxContainerResources) Size() (n int) {
- var l int
- _ = l
- if m.CpuPeriod != 0 {
- n += 1 + sovApi(uint64(m.CpuPeriod))
- }
- if m.CpuQuota != 0 {
- n += 1 + sovApi(uint64(m.CpuQuota))
- }
- if m.CpuShares != 0 {
- n += 1 + sovApi(uint64(m.CpuShares))
- }
- if m.MemoryLimitInBytes != 0 {
- n += 1 + sovApi(uint64(m.MemoryLimitInBytes))
- }
- if m.OomScoreAdj != 0 {
- n += 1 + sovApi(uint64(m.OomScoreAdj))
- }
- l = len(m.CpusetCpus)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.CpusetMems)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *SELinuxOption) Size() (n int) {
- var l int
- _ = l
- l = len(m.User)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Role)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Type)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Level)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *Capability) Size() (n int) {
- var l int
- _ = l
- if len(m.AddCapabilities) > 0 {
- for _, s := range m.AddCapabilities {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.DropCapabilities) > 0 {
- for _, s := range m.DropCapabilities {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *LinuxContainerSecurityContext) Size() (n int) {
- var l int
- _ = l
- if m.Capabilities != nil {
- l = m.Capabilities.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Privileged {
- n += 2
- }
- if m.NamespaceOptions != nil {
- l = m.NamespaceOptions.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SelinuxOptions != nil {
- l = m.SelinuxOptions.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.RunAsUser != nil {
- l = m.RunAsUser.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.RunAsUsername)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.ReadonlyRootfs {
- n += 2
- }
- if len(m.SupplementalGroups) > 0 {
- l = 0
- for _, e := range m.SupplementalGroups {
- l += sovApi(uint64(e))
- }
- n += 1 + sovApi(uint64(l)) + l
- }
- l = len(m.ApparmorProfile)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.SeccompProfilePath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.NoNewPrivs {
- n += 2
- }
- return n
-}
-
-func (m *LinuxContainerConfig) Size() (n int) {
- var l int
- _ = l
- if m.Resources != nil {
- l = m.Resources.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SecurityContext != nil {
- l = m.SecurityContext.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *WindowsContainerConfig) Size() (n int) {
- var l int
- _ = l
- if m.Resources != nil {
- l = m.Resources.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *WindowsContainerResources) Size() (n int) {
- var l int
- _ = l
- if m.CpuShares != 0 {
- n += 1 + sovApi(uint64(m.CpuShares))
- }
- if m.CpuCount != 0 {
- n += 1 + sovApi(uint64(m.CpuCount))
- }
- if m.CpuMaximum != 0 {
- n += 1 + sovApi(uint64(m.CpuMaximum))
- }
- if m.MemoryLimitInBytes != 0 {
- n += 1 + sovApi(uint64(m.MemoryLimitInBytes))
- }
- return n
-}
-
-func (m *ContainerMetadata) Size() (n int) {
- var l int
- _ = l
- l = len(m.Name)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Attempt != 0 {
- n += 1 + sovApi(uint64(m.Attempt))
- }
- return n
-}
-
-func (m *Device) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.HostPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Permissions)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ContainerConfig) Size() (n int) {
- var l int
- _ = l
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Command) > 0 {
- for _, s := range m.Command {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Args) > 0 {
- for _, s := range m.Args {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- l = len(m.WorkingDir)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Envs) > 0 {
- for _, e := range m.Envs {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Mounts) > 0 {
- for _, e := range m.Mounts {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Devices) > 0 {
- for _, e := range m.Devices {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- l = len(m.LogPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Stdin {
- n += 2
- }
- if m.StdinOnce {
- n += 2
- }
- if m.Tty {
- n += 2
- }
- if m.Linux != nil {
- l = m.Linux.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Windows != nil {
- l = m.Windows.Size()
- n += 2 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *CreateContainerRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Config != nil {
- l = m.Config.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SandboxConfig != nil {
- l = m.SandboxConfig.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *CreateContainerResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *StartContainerRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *StartContainerResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *StopContainerRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Timeout != 0 {
- n += 1 + sovApi(uint64(m.Timeout))
- }
- return n
-}
-
-func (m *StopContainerResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *RemoveContainerRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RemoveContainerResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *ContainerStateValue) Size() (n int) {
- var l int
- _ = l
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- return n
-}
-
-func (m *ContainerFilter) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != nil {
- l = m.State.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.LabelSelector) > 0 {
- for k, v := range m.LabelSelector {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ListContainersRequest) Size() (n int) {
- var l int
- _ = l
- if m.Filter != nil {
- l = m.Filter.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *Container) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.ImageRef)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- if m.CreatedAt != 0 {
- n += 1 + sovApi(uint64(m.CreatedAt))
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ListContainersResponse) Size() (n int) {
- var l int
- _ = l
- if len(m.Containers) > 0 {
- for _, e := range m.Containers {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *ContainerStatusRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Verbose {
- n += 2
- }
- return n
-}
-
-func (m *ContainerStatus) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.State != 0 {
- n += 1 + sovApi(uint64(m.State))
- }
- if m.CreatedAt != 0 {
- n += 1 + sovApi(uint64(m.CreatedAt))
- }
- if m.StartedAt != 0 {
- n += 1 + sovApi(uint64(m.StartedAt))
- }
- if m.FinishedAt != 0 {
- n += 1 + sovApi(uint64(m.FinishedAt))
- }
- if m.ExitCode != 0 {
- n += 1 + sovApi(uint64(m.ExitCode))
- }
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.ImageRef)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Reason)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Message)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Mounts) > 0 {
- for _, e := range m.Mounts {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- l = len(m.LogPath)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ContainerStatusResponse) Size() (n int) {
- var l int
- _ = l
- if m.Status != nil {
- l = m.Status.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Info) > 0 {
- for k, v := range m.Info {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *UpdateContainerResourcesRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Linux != nil {
- l = m.Linux.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *UpdateContainerResourcesResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *ExecSyncRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Cmd) > 0 {
- for _, s := range m.Cmd {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if m.Timeout != 0 {
- n += 1 + sovApi(uint64(m.Timeout))
- }
- return n
-}
-
-func (m *ExecSyncResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.Stdout)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Stderr)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.ExitCode != 0 {
- n += 1 + sovApi(uint64(m.ExitCode))
- }
- return n
-}
-
-func (m *ExecRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Cmd) > 0 {
- for _, s := range m.Cmd {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if m.Tty {
- n += 2
- }
- if m.Stdin {
- n += 2
- }
- if m.Stdout {
- n += 2
- }
- if m.Stderr {
- n += 2
- }
- return n
-}
-
-func (m *ExecResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.Url)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *AttachRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Stdin {
- n += 2
- }
- if m.Tty {
- n += 2
- }
- if m.Stdout {
- n += 2
- }
- if m.Stderr {
- n += 2
- }
- return n
-}
-
-func (m *AttachResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.Url)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *PortForwardRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Port) > 0 {
- l = 0
- for _, e := range m.Port {
- l += sovApi(uint64(e))
- }
- n += 1 + sovApi(uint64(l)) + l
- }
- return n
-}
-
-func (m *PortForwardResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.Url)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ImageFilter) Size() (n int) {
- var l int
- _ = l
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ListImagesRequest) Size() (n int) {
- var l int
- _ = l
- if m.Filter != nil {
- l = m.Filter.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *Image) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.RepoTags) > 0 {
- for _, s := range m.RepoTags {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if len(m.RepoDigests) > 0 {
- for _, s := range m.RepoDigests {
- l = len(s)
- n += 1 + l + sovApi(uint64(l))
- }
- }
- if m.Size_ != 0 {
- n += 1 + sovApi(uint64(m.Size_))
- }
- if m.Uid != nil {
- l = m.Uid.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Username)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ListImagesResponse) Size() (n int) {
- var l int
- _ = l
- if len(m.Images) > 0 {
- for _, e := range m.Images {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *ImageStatusRequest) Size() (n int) {
- var l int
- _ = l
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Verbose {
- n += 2
- }
- return n
-}
-
-func (m *ImageStatusResponse) Size() (n int) {
- var l int
- _ = l
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Info) > 0 {
- for k, v := range m.Info {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *AuthConfig) Size() (n int) {
- var l int
- _ = l
- l = len(m.Username)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Password)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Auth)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.ServerAddress)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.IdentityToken)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.RegistryToken)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *PullImageRequest) Size() (n int) {
- var l int
- _ = l
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Auth != nil {
- l = m.Auth.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.SandboxConfig != nil {
- l = m.SandboxConfig.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *PullImageResponse) Size() (n int) {
- var l int
- _ = l
- l = len(m.ImageRef)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RemoveImageRequest) Size() (n int) {
- var l int
- _ = l
- if m.Image != nil {
- l = m.Image.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RemoveImageResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *NetworkConfig) Size() (n int) {
- var l int
- _ = l
- l = len(m.PodCidr)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RuntimeConfig) Size() (n int) {
- var l int
- _ = l
- if m.NetworkConfig != nil {
- l = m.NetworkConfig.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *UpdateRuntimeConfigRequest) Size() (n int) {
- var l int
- _ = l
- if m.RuntimeConfig != nil {
- l = m.RuntimeConfig.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *UpdateRuntimeConfigResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *RuntimeCondition) Size() (n int) {
- var l int
- _ = l
- l = len(m.Type)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Status {
- n += 2
- }
- l = len(m.Reason)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.Message)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *RuntimeStatus) Size() (n int) {
- var l int
- _ = l
- if len(m.Conditions) > 0 {
- for _, e := range m.Conditions {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *StatusRequest) Size() (n int) {
- var l int
- _ = l
- if m.Verbose {
- n += 2
- }
- return n
-}
-
-func (m *StatusResponse) Size() (n int) {
- var l int
- _ = l
- if m.Status != nil {
- l = m.Status.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Info) > 0 {
- for k, v := range m.Info {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ImageFsInfoRequest) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func (m *UInt64Value) Size() (n int) {
- var l int
- _ = l
- if m.Value != 0 {
- n += 1 + sovApi(uint64(m.Value))
- }
- return n
-}
-
-func (m *FilesystemIdentifier) Size() (n int) {
- var l int
- _ = l
- l = len(m.Mountpoint)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *FilesystemUsage) Size() (n int) {
- var l int
- _ = l
- if m.Timestamp != 0 {
- n += 1 + sovApi(uint64(m.Timestamp))
- }
- if m.FsId != nil {
- l = m.FsId.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.UsedBytes != nil {
- l = m.UsedBytes.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.InodesUsed != nil {
- l = m.InodesUsed.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ImageFsInfoResponse) Size() (n int) {
- var l int
- _ = l
- if len(m.ImageFilesystems) > 0 {
- for _, e := range m.ImageFilesystems {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *ContainerStatsRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ContainerStatsResponse) Size() (n int) {
- var l int
- _ = l
- if m.Stats != nil {
- l = m.Stats.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ListContainerStatsRequest) Size() (n int) {
- var l int
- _ = l
- if m.Filter != nil {
- l = m.Filter.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ContainerStatsFilter) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- l = len(m.PodSandboxId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.LabelSelector) > 0 {
- for k, v := range m.LabelSelector {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ListContainerStatsResponse) Size() (n int) {
- var l int
- _ = l
- if len(m.Stats) > 0 {
- for _, e := range m.Stats {
- l = e.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- }
- return n
-}
-
-func (m *ContainerAttributes) Size() (n int) {
- var l int
- _ = l
- l = len(m.Id)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Metadata != nil {
- l = m.Metadata.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if len(m.Labels) > 0 {
- for k, v := range m.Labels {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- if len(m.Annotations) > 0 {
- for k, v := range m.Annotations {
- _ = k
- _ = v
- mapEntrySize := 1 + len(k) + sovApi(uint64(len(k))) + 1 + len(v) + sovApi(uint64(len(v)))
- n += mapEntrySize + 1 + sovApi(uint64(mapEntrySize))
- }
- }
- return n
-}
-
-func (m *ContainerStats) Size() (n int) {
- var l int
- _ = l
- if m.Attributes != nil {
- l = m.Attributes.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Cpu != nil {
- l = m.Cpu.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.Memory != nil {
- l = m.Memory.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- if m.WritableLayer != nil {
- l = m.WritableLayer.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *CpuUsage) Size() (n int) {
- var l int
- _ = l
- if m.Timestamp != 0 {
- n += 1 + sovApi(uint64(m.Timestamp))
- }
- if m.UsageCoreNanoSeconds != nil {
- l = m.UsageCoreNanoSeconds.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *MemoryUsage) Size() (n int) {
- var l int
- _ = l
- if m.Timestamp != 0 {
- n += 1 + sovApi(uint64(m.Timestamp))
- }
- if m.WorkingSetBytes != nil {
- l = m.WorkingSetBytes.Size()
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ReopenContainerLogRequest) Size() (n int) {
- var l int
- _ = l
- l = len(m.ContainerId)
- if l > 0 {
- n += 1 + l + sovApi(uint64(l))
- }
- return n
-}
-
-func (m *ReopenContainerLogResponse) Size() (n int) {
- var l int
- _ = l
- return n
-}
-
-func sovApi(x uint64) (n int) {
- for {
- n++
- x >>= 7
- if x == 0 {
- break
- }
- }
- return n
-}
-func sozApi(x uint64) (n int) {
- return sovApi(uint64((x << 1) ^ uint64((int64(x) >> 63))))
-}
-func (this *VersionRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&VersionRequest{`,
- `Version:` + fmt.Sprintf("%v", this.Version) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *VersionResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&VersionResponse{`,
- `Version:` + fmt.Sprintf("%v", this.Version) + `,`,
- `RuntimeName:` + fmt.Sprintf("%v", this.RuntimeName) + `,`,
- `RuntimeVersion:` + fmt.Sprintf("%v", this.RuntimeVersion) + `,`,
- `RuntimeApiVersion:` + fmt.Sprintf("%v", this.RuntimeApiVersion) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *DNSConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&DNSConfig{`,
- `Servers:` + fmt.Sprintf("%v", this.Servers) + `,`,
- `Searches:` + fmt.Sprintf("%v", this.Searches) + `,`,
- `Options:` + fmt.Sprintf("%v", this.Options) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PortMapping) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PortMapping{`,
- `Protocol:` + fmt.Sprintf("%v", this.Protocol) + `,`,
- `ContainerPort:` + fmt.Sprintf("%v", this.ContainerPort) + `,`,
- `HostPort:` + fmt.Sprintf("%v", this.HostPort) + `,`,
- `HostIp:` + fmt.Sprintf("%v", this.HostIp) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Mount) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Mount{`,
- `ContainerPath:` + fmt.Sprintf("%v", this.ContainerPath) + `,`,
- `HostPath:` + fmt.Sprintf("%v", this.HostPath) + `,`,
- `Readonly:` + fmt.Sprintf("%v", this.Readonly) + `,`,
- `SelinuxRelabel:` + fmt.Sprintf("%v", this.SelinuxRelabel) + `,`,
- `Propagation:` + fmt.Sprintf("%v", this.Propagation) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *NamespaceOption) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&NamespaceOption{`,
- `Network:` + fmt.Sprintf("%v", this.Network) + `,`,
- `Pid:` + fmt.Sprintf("%v", this.Pid) + `,`,
- `Ipc:` + fmt.Sprintf("%v", this.Ipc) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Int64Value) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Int64Value{`,
- `Value:` + fmt.Sprintf("%v", this.Value) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxSandboxSecurityContext) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&LinuxSandboxSecurityContext{`,
- `NamespaceOptions:` + strings.Replace(fmt.Sprintf("%v", this.NamespaceOptions), "NamespaceOption", "NamespaceOption", 1) + `,`,
- `SelinuxOptions:` + strings.Replace(fmt.Sprintf("%v", this.SelinuxOptions), "SELinuxOption", "SELinuxOption", 1) + `,`,
- `RunAsUser:` + strings.Replace(fmt.Sprintf("%v", this.RunAsUser), "Int64Value", "Int64Value", 1) + `,`,
- `ReadonlyRootfs:` + fmt.Sprintf("%v", this.ReadonlyRootfs) + `,`,
- `SupplementalGroups:` + fmt.Sprintf("%v", this.SupplementalGroups) + `,`,
- `Privileged:` + fmt.Sprintf("%v", this.Privileged) + `,`,
- `SeccompProfilePath:` + fmt.Sprintf("%v", this.SeccompProfilePath) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxPodSandboxConfig) String() string {
- if this == nil {
- return "nil"
- }
- keysForSysctls := make([]string, 0, len(this.Sysctls))
- for k := range this.Sysctls {
- keysForSysctls = append(keysForSysctls, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForSysctls)
- mapStringForSysctls := "map[string]string{"
- for _, k := range keysForSysctls {
- mapStringForSysctls += fmt.Sprintf("%v: %v,", k, this.Sysctls[k])
- }
- mapStringForSysctls += "}"
- s := strings.Join([]string{`&LinuxPodSandboxConfig{`,
- `CgroupParent:` + fmt.Sprintf("%v", this.CgroupParent) + `,`,
- `SecurityContext:` + strings.Replace(fmt.Sprintf("%v", this.SecurityContext), "LinuxSandboxSecurityContext", "LinuxSandboxSecurityContext", 1) + `,`,
- `Sysctls:` + mapStringForSysctls + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxMetadata) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PodSandboxMetadata{`,
- `Name:` + fmt.Sprintf("%v", this.Name) + `,`,
- `Uid:` + fmt.Sprintf("%v", this.Uid) + `,`,
- `Namespace:` + fmt.Sprintf("%v", this.Namespace) + `,`,
- `Attempt:` + fmt.Sprintf("%v", this.Attempt) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxConfig) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&PodSandboxConfig{`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "PodSandboxMetadata", "PodSandboxMetadata", 1) + `,`,
- `Hostname:` + fmt.Sprintf("%v", this.Hostname) + `,`,
- `LogDirectory:` + fmt.Sprintf("%v", this.LogDirectory) + `,`,
- `DnsConfig:` + strings.Replace(fmt.Sprintf("%v", this.DnsConfig), "DNSConfig", "DNSConfig", 1) + `,`,
- `PortMappings:` + strings.Replace(fmt.Sprintf("%v", this.PortMappings), "PortMapping", "PortMapping", 1) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `Linux:` + strings.Replace(fmt.Sprintf("%v", this.Linux), "LinuxPodSandboxConfig", "LinuxPodSandboxConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RunPodSandboxRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RunPodSandboxRequest{`,
- `Config:` + strings.Replace(fmt.Sprintf("%v", this.Config), "PodSandboxConfig", "PodSandboxConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RunPodSandboxResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RunPodSandboxResponse{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StopPodSandboxRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StopPodSandboxRequest{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StopPodSandboxResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StopPodSandboxResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *RemovePodSandboxRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemovePodSandboxRequest{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RemovePodSandboxResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemovePodSandboxResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxStatusRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PodSandboxStatusRequest{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `Verbose:` + fmt.Sprintf("%v", this.Verbose) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxNetworkStatus) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PodSandboxNetworkStatus{`,
- `Ip:` + fmt.Sprintf("%v", this.Ip) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Namespace) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Namespace{`,
- `Options:` + strings.Replace(fmt.Sprintf("%v", this.Options), "NamespaceOption", "NamespaceOption", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxPodSandboxStatus) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&LinuxPodSandboxStatus{`,
- `Namespaces:` + strings.Replace(fmt.Sprintf("%v", this.Namespaces), "Namespace", "Namespace", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxStatus) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&PodSandboxStatus{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "PodSandboxMetadata", "PodSandboxMetadata", 1) + `,`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `CreatedAt:` + fmt.Sprintf("%v", this.CreatedAt) + `,`,
- `Network:` + strings.Replace(fmt.Sprintf("%v", this.Network), "PodSandboxNetworkStatus", "PodSandboxNetworkStatus", 1) + `,`,
- `Linux:` + strings.Replace(fmt.Sprintf("%v", this.Linux), "LinuxPodSandboxStatus", "LinuxPodSandboxStatus", 1) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxStatusResponse) String() string {
- if this == nil {
- return "nil"
- }
- keysForInfo := make([]string, 0, len(this.Info))
- for k := range this.Info {
- keysForInfo = append(keysForInfo, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForInfo)
- mapStringForInfo := "map[string]string{"
- for _, k := range keysForInfo {
- mapStringForInfo += fmt.Sprintf("%v: %v,", k, this.Info[k])
- }
- mapStringForInfo += "}"
- s := strings.Join([]string{`&PodSandboxStatusResponse{`,
- `Status:` + strings.Replace(fmt.Sprintf("%v", this.Status), "PodSandboxStatus", "PodSandboxStatus", 1) + `,`,
- `Info:` + mapStringForInfo + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxStateValue) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PodSandboxStateValue{`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandboxFilter) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabelSelector := make([]string, 0, len(this.LabelSelector))
- for k := range this.LabelSelector {
- keysForLabelSelector = append(keysForLabelSelector, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabelSelector)
- mapStringForLabelSelector := "map[string]string{"
- for _, k := range keysForLabelSelector {
- mapStringForLabelSelector += fmt.Sprintf("%v: %v,", k, this.LabelSelector[k])
- }
- mapStringForLabelSelector += "}"
- s := strings.Join([]string{`&PodSandboxFilter{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `State:` + strings.Replace(fmt.Sprintf("%v", this.State), "PodSandboxStateValue", "PodSandboxStateValue", 1) + `,`,
- `LabelSelector:` + mapStringForLabelSelector + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListPodSandboxRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListPodSandboxRequest{`,
- `Filter:` + strings.Replace(fmt.Sprintf("%v", this.Filter), "PodSandboxFilter", "PodSandboxFilter", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PodSandbox) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&PodSandbox{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "PodSandboxMetadata", "PodSandboxMetadata", 1) + `,`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `CreatedAt:` + fmt.Sprintf("%v", this.CreatedAt) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListPodSandboxResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListPodSandboxResponse{`,
- `Items:` + strings.Replace(fmt.Sprintf("%v", this.Items), "PodSandbox", "PodSandbox", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageSpec) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ImageSpec{`,
- `Image:` + fmt.Sprintf("%v", this.Image) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *KeyValue) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&KeyValue{`,
- `Key:` + fmt.Sprintf("%v", this.Key) + `,`,
- `Value:` + fmt.Sprintf("%v", this.Value) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxContainerResources) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&LinuxContainerResources{`,
- `CpuPeriod:` + fmt.Sprintf("%v", this.CpuPeriod) + `,`,
- `CpuQuota:` + fmt.Sprintf("%v", this.CpuQuota) + `,`,
- `CpuShares:` + fmt.Sprintf("%v", this.CpuShares) + `,`,
- `MemoryLimitInBytes:` + fmt.Sprintf("%v", this.MemoryLimitInBytes) + `,`,
- `OomScoreAdj:` + fmt.Sprintf("%v", this.OomScoreAdj) + `,`,
- `CpusetCpus:` + fmt.Sprintf("%v", this.CpusetCpus) + `,`,
- `CpusetMems:` + fmt.Sprintf("%v", this.CpusetMems) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *SELinuxOption) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&SELinuxOption{`,
- `User:` + fmt.Sprintf("%v", this.User) + `,`,
- `Role:` + fmt.Sprintf("%v", this.Role) + `,`,
- `Type:` + fmt.Sprintf("%v", this.Type) + `,`,
- `Level:` + fmt.Sprintf("%v", this.Level) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Capability) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Capability{`,
- `AddCapabilities:` + fmt.Sprintf("%v", this.AddCapabilities) + `,`,
- `DropCapabilities:` + fmt.Sprintf("%v", this.DropCapabilities) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxContainerSecurityContext) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&LinuxContainerSecurityContext{`,
- `Capabilities:` + strings.Replace(fmt.Sprintf("%v", this.Capabilities), "Capability", "Capability", 1) + `,`,
- `Privileged:` + fmt.Sprintf("%v", this.Privileged) + `,`,
- `NamespaceOptions:` + strings.Replace(fmt.Sprintf("%v", this.NamespaceOptions), "NamespaceOption", "NamespaceOption", 1) + `,`,
- `SelinuxOptions:` + strings.Replace(fmt.Sprintf("%v", this.SelinuxOptions), "SELinuxOption", "SELinuxOption", 1) + `,`,
- `RunAsUser:` + strings.Replace(fmt.Sprintf("%v", this.RunAsUser), "Int64Value", "Int64Value", 1) + `,`,
- `RunAsUsername:` + fmt.Sprintf("%v", this.RunAsUsername) + `,`,
- `ReadonlyRootfs:` + fmt.Sprintf("%v", this.ReadonlyRootfs) + `,`,
- `SupplementalGroups:` + fmt.Sprintf("%v", this.SupplementalGroups) + `,`,
- `ApparmorProfile:` + fmt.Sprintf("%v", this.ApparmorProfile) + `,`,
- `SeccompProfilePath:` + fmt.Sprintf("%v", this.SeccompProfilePath) + `,`,
- `NoNewPrivs:` + fmt.Sprintf("%v", this.NoNewPrivs) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *LinuxContainerConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&LinuxContainerConfig{`,
- `Resources:` + strings.Replace(fmt.Sprintf("%v", this.Resources), "LinuxContainerResources", "LinuxContainerResources", 1) + `,`,
- `SecurityContext:` + strings.Replace(fmt.Sprintf("%v", this.SecurityContext), "LinuxContainerSecurityContext", "LinuxContainerSecurityContext", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *WindowsContainerConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&WindowsContainerConfig{`,
- `Resources:` + strings.Replace(fmt.Sprintf("%v", this.Resources), "WindowsContainerResources", "WindowsContainerResources", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *WindowsContainerResources) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&WindowsContainerResources{`,
- `CpuShares:` + fmt.Sprintf("%v", this.CpuShares) + `,`,
- `CpuCount:` + fmt.Sprintf("%v", this.CpuCount) + `,`,
- `CpuMaximum:` + fmt.Sprintf("%v", this.CpuMaximum) + `,`,
- `MemoryLimitInBytes:` + fmt.Sprintf("%v", this.MemoryLimitInBytes) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerMetadata) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerMetadata{`,
- `Name:` + fmt.Sprintf("%v", this.Name) + `,`,
- `Attempt:` + fmt.Sprintf("%v", this.Attempt) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Device) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Device{`,
- `ContainerPath:` + fmt.Sprintf("%v", this.ContainerPath) + `,`,
- `HostPath:` + fmt.Sprintf("%v", this.HostPath) + `,`,
- `Permissions:` + fmt.Sprintf("%v", this.Permissions) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerConfig) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&ContainerConfig{`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "ContainerMetadata", "ContainerMetadata", 1) + `,`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `Command:` + fmt.Sprintf("%v", this.Command) + `,`,
- `Args:` + fmt.Sprintf("%v", this.Args) + `,`,
- `WorkingDir:` + fmt.Sprintf("%v", this.WorkingDir) + `,`,
- `Envs:` + strings.Replace(fmt.Sprintf("%v", this.Envs), "KeyValue", "KeyValue", 1) + `,`,
- `Mounts:` + strings.Replace(fmt.Sprintf("%v", this.Mounts), "Mount", "Mount", 1) + `,`,
- `Devices:` + strings.Replace(fmt.Sprintf("%v", this.Devices), "Device", "Device", 1) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `LogPath:` + fmt.Sprintf("%v", this.LogPath) + `,`,
- `Stdin:` + fmt.Sprintf("%v", this.Stdin) + `,`,
- `StdinOnce:` + fmt.Sprintf("%v", this.StdinOnce) + `,`,
- `Tty:` + fmt.Sprintf("%v", this.Tty) + `,`,
- `Linux:` + strings.Replace(fmt.Sprintf("%v", this.Linux), "LinuxContainerConfig", "LinuxContainerConfig", 1) + `,`,
- `Windows:` + strings.Replace(fmt.Sprintf("%v", this.Windows), "WindowsContainerConfig", "WindowsContainerConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *CreateContainerRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&CreateContainerRequest{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `Config:` + strings.Replace(fmt.Sprintf("%v", this.Config), "ContainerConfig", "ContainerConfig", 1) + `,`,
- `SandboxConfig:` + strings.Replace(fmt.Sprintf("%v", this.SandboxConfig), "PodSandboxConfig", "PodSandboxConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *CreateContainerResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&CreateContainerResponse{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StartContainerRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StartContainerRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StartContainerResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StartContainerResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *StopContainerRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StopContainerRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Timeout:` + fmt.Sprintf("%v", this.Timeout) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StopContainerResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StopContainerResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *RemoveContainerRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemoveContainerRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RemoveContainerResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemoveContainerResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStateValue) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerStateValue{`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerFilter) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabelSelector := make([]string, 0, len(this.LabelSelector))
- for k := range this.LabelSelector {
- keysForLabelSelector = append(keysForLabelSelector, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabelSelector)
- mapStringForLabelSelector := "map[string]string{"
- for _, k := range keysForLabelSelector {
- mapStringForLabelSelector += fmt.Sprintf("%v: %v,", k, this.LabelSelector[k])
- }
- mapStringForLabelSelector += "}"
- s := strings.Join([]string{`&ContainerFilter{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `State:` + strings.Replace(fmt.Sprintf("%v", this.State), "ContainerStateValue", "ContainerStateValue", 1) + `,`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `LabelSelector:` + mapStringForLabelSelector + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListContainersRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListContainersRequest{`,
- `Filter:` + strings.Replace(fmt.Sprintf("%v", this.Filter), "ContainerFilter", "ContainerFilter", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Container) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&Container{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "ContainerMetadata", "ContainerMetadata", 1) + `,`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `ImageRef:` + fmt.Sprintf("%v", this.ImageRef) + `,`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `CreatedAt:` + fmt.Sprintf("%v", this.CreatedAt) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListContainersResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListContainersResponse{`,
- `Containers:` + strings.Replace(fmt.Sprintf("%v", this.Containers), "Container", "Container", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatusRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerStatusRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Verbose:` + fmt.Sprintf("%v", this.Verbose) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatus) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&ContainerStatus{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "ContainerMetadata", "ContainerMetadata", 1) + `,`,
- `State:` + fmt.Sprintf("%v", this.State) + `,`,
- `CreatedAt:` + fmt.Sprintf("%v", this.CreatedAt) + `,`,
- `StartedAt:` + fmt.Sprintf("%v", this.StartedAt) + `,`,
- `FinishedAt:` + fmt.Sprintf("%v", this.FinishedAt) + `,`,
- `ExitCode:` + fmt.Sprintf("%v", this.ExitCode) + `,`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `ImageRef:` + fmt.Sprintf("%v", this.ImageRef) + `,`,
- `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`,
- `Message:` + fmt.Sprintf("%v", this.Message) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `Mounts:` + strings.Replace(fmt.Sprintf("%v", this.Mounts), "Mount", "Mount", 1) + `,`,
- `LogPath:` + fmt.Sprintf("%v", this.LogPath) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatusResponse) String() string {
- if this == nil {
- return "nil"
- }
- keysForInfo := make([]string, 0, len(this.Info))
- for k := range this.Info {
- keysForInfo = append(keysForInfo, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForInfo)
- mapStringForInfo := "map[string]string{"
- for _, k := range keysForInfo {
- mapStringForInfo += fmt.Sprintf("%v: %v,", k, this.Info[k])
- }
- mapStringForInfo += "}"
- s := strings.Join([]string{`&ContainerStatusResponse{`,
- `Status:` + strings.Replace(fmt.Sprintf("%v", this.Status), "ContainerStatus", "ContainerStatus", 1) + `,`,
- `Info:` + mapStringForInfo + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *UpdateContainerResourcesRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&UpdateContainerResourcesRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Linux:` + strings.Replace(fmt.Sprintf("%v", this.Linux), "LinuxContainerResources", "LinuxContainerResources", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *UpdateContainerResourcesResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&UpdateContainerResourcesResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *ExecSyncRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ExecSyncRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Cmd:` + fmt.Sprintf("%v", this.Cmd) + `,`,
- `Timeout:` + fmt.Sprintf("%v", this.Timeout) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ExecSyncResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ExecSyncResponse{`,
- `Stdout:` + fmt.Sprintf("%v", this.Stdout) + `,`,
- `Stderr:` + fmt.Sprintf("%v", this.Stderr) + `,`,
- `ExitCode:` + fmt.Sprintf("%v", this.ExitCode) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ExecRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ExecRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Cmd:` + fmt.Sprintf("%v", this.Cmd) + `,`,
- `Tty:` + fmt.Sprintf("%v", this.Tty) + `,`,
- `Stdin:` + fmt.Sprintf("%v", this.Stdin) + `,`,
- `Stdout:` + fmt.Sprintf("%v", this.Stdout) + `,`,
- `Stderr:` + fmt.Sprintf("%v", this.Stderr) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ExecResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ExecResponse{`,
- `Url:` + fmt.Sprintf("%v", this.Url) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *AttachRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&AttachRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `Stdin:` + fmt.Sprintf("%v", this.Stdin) + `,`,
- `Tty:` + fmt.Sprintf("%v", this.Tty) + `,`,
- `Stdout:` + fmt.Sprintf("%v", this.Stdout) + `,`,
- `Stderr:` + fmt.Sprintf("%v", this.Stderr) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *AttachResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&AttachResponse{`,
- `Url:` + fmt.Sprintf("%v", this.Url) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PortForwardRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PortForwardRequest{`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `Port:` + fmt.Sprintf("%v", this.Port) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PortForwardResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PortForwardResponse{`,
- `Url:` + fmt.Sprintf("%v", this.Url) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageFilter) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ImageFilter{`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListImagesRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListImagesRequest{`,
- `Filter:` + strings.Replace(fmt.Sprintf("%v", this.Filter), "ImageFilter", "ImageFilter", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *Image) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&Image{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `RepoTags:` + fmt.Sprintf("%v", this.RepoTags) + `,`,
- `RepoDigests:` + fmt.Sprintf("%v", this.RepoDigests) + `,`,
- `Size_:` + fmt.Sprintf("%v", this.Size_) + `,`,
- `Uid:` + strings.Replace(fmt.Sprintf("%v", this.Uid), "Int64Value", "Int64Value", 1) + `,`,
- `Username:` + fmt.Sprintf("%v", this.Username) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListImagesResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListImagesResponse{`,
- `Images:` + strings.Replace(fmt.Sprintf("%v", this.Images), "Image", "Image", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageStatusRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ImageStatusRequest{`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `Verbose:` + fmt.Sprintf("%v", this.Verbose) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageStatusResponse) String() string {
- if this == nil {
- return "nil"
- }
- keysForInfo := make([]string, 0, len(this.Info))
- for k := range this.Info {
- keysForInfo = append(keysForInfo, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForInfo)
- mapStringForInfo := "map[string]string{"
- for _, k := range keysForInfo {
- mapStringForInfo += fmt.Sprintf("%v: %v,", k, this.Info[k])
- }
- mapStringForInfo += "}"
- s := strings.Join([]string{`&ImageStatusResponse{`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "Image", "Image", 1) + `,`,
- `Info:` + mapStringForInfo + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *AuthConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&AuthConfig{`,
- `Username:` + fmt.Sprintf("%v", this.Username) + `,`,
- `Password:` + fmt.Sprintf("%v", this.Password) + `,`,
- `Auth:` + fmt.Sprintf("%v", this.Auth) + `,`,
- `ServerAddress:` + fmt.Sprintf("%v", this.ServerAddress) + `,`,
- `IdentityToken:` + fmt.Sprintf("%v", this.IdentityToken) + `,`,
- `RegistryToken:` + fmt.Sprintf("%v", this.RegistryToken) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PullImageRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PullImageRequest{`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `Auth:` + strings.Replace(fmt.Sprintf("%v", this.Auth), "AuthConfig", "AuthConfig", 1) + `,`,
- `SandboxConfig:` + strings.Replace(fmt.Sprintf("%v", this.SandboxConfig), "PodSandboxConfig", "PodSandboxConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *PullImageResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&PullImageResponse{`,
- `ImageRef:` + fmt.Sprintf("%v", this.ImageRef) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RemoveImageRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemoveImageRequest{`,
- `Image:` + strings.Replace(fmt.Sprintf("%v", this.Image), "ImageSpec", "ImageSpec", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RemoveImageResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RemoveImageResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *NetworkConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&NetworkConfig{`,
- `PodCidr:` + fmt.Sprintf("%v", this.PodCidr) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RuntimeConfig) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RuntimeConfig{`,
- `NetworkConfig:` + strings.Replace(fmt.Sprintf("%v", this.NetworkConfig), "NetworkConfig", "NetworkConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *UpdateRuntimeConfigRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&UpdateRuntimeConfigRequest{`,
- `RuntimeConfig:` + strings.Replace(fmt.Sprintf("%v", this.RuntimeConfig), "RuntimeConfig", "RuntimeConfig", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *UpdateRuntimeConfigResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&UpdateRuntimeConfigResponse{`,
- `}`,
- }, "")
- return s
-}
-func (this *RuntimeCondition) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RuntimeCondition{`,
- `Type:` + fmt.Sprintf("%v", this.Type) + `,`,
- `Status:` + fmt.Sprintf("%v", this.Status) + `,`,
- `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`,
- `Message:` + fmt.Sprintf("%v", this.Message) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *RuntimeStatus) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&RuntimeStatus{`,
- `Conditions:` + strings.Replace(fmt.Sprintf("%v", this.Conditions), "RuntimeCondition", "RuntimeCondition", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StatusRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&StatusRequest{`,
- `Verbose:` + fmt.Sprintf("%v", this.Verbose) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *StatusResponse) String() string {
- if this == nil {
- return "nil"
- }
- keysForInfo := make([]string, 0, len(this.Info))
- for k := range this.Info {
- keysForInfo = append(keysForInfo, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForInfo)
- mapStringForInfo := "map[string]string{"
- for _, k := range keysForInfo {
- mapStringForInfo += fmt.Sprintf("%v: %v,", k, this.Info[k])
- }
- mapStringForInfo += "}"
- s := strings.Join([]string{`&StatusResponse{`,
- `Status:` + strings.Replace(fmt.Sprintf("%v", this.Status), "RuntimeStatus", "RuntimeStatus", 1) + `,`,
- `Info:` + mapStringForInfo + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageFsInfoRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ImageFsInfoRequest{`,
- `}`,
- }, "")
- return s
-}
-func (this *UInt64Value) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&UInt64Value{`,
- `Value:` + fmt.Sprintf("%v", this.Value) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *FilesystemIdentifier) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&FilesystemIdentifier{`,
- `Mountpoint:` + fmt.Sprintf("%v", this.Mountpoint) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *FilesystemUsage) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&FilesystemUsage{`,
- `Timestamp:` + fmt.Sprintf("%v", this.Timestamp) + `,`,
- `FsId:` + strings.Replace(fmt.Sprintf("%v", this.FsId), "FilesystemIdentifier", "FilesystemIdentifier", 1) + `,`,
- `UsedBytes:` + strings.Replace(fmt.Sprintf("%v", this.UsedBytes), "UInt64Value", "UInt64Value", 1) + `,`,
- `InodesUsed:` + strings.Replace(fmt.Sprintf("%v", this.InodesUsed), "UInt64Value", "UInt64Value", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ImageFsInfoResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ImageFsInfoResponse{`,
- `ImageFilesystems:` + strings.Replace(fmt.Sprintf("%v", this.ImageFilesystems), "FilesystemUsage", "FilesystemUsage", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatsRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerStatsRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatsResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerStatsResponse{`,
- `Stats:` + strings.Replace(fmt.Sprintf("%v", this.Stats), "ContainerStats", "ContainerStats", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListContainerStatsRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListContainerStatsRequest{`,
- `Filter:` + strings.Replace(fmt.Sprintf("%v", this.Filter), "ContainerStatsFilter", "ContainerStatsFilter", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStatsFilter) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabelSelector := make([]string, 0, len(this.LabelSelector))
- for k := range this.LabelSelector {
- keysForLabelSelector = append(keysForLabelSelector, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabelSelector)
- mapStringForLabelSelector := "map[string]string{"
- for _, k := range keysForLabelSelector {
- mapStringForLabelSelector += fmt.Sprintf("%v: %v,", k, this.LabelSelector[k])
- }
- mapStringForLabelSelector += "}"
- s := strings.Join([]string{`&ContainerStatsFilter{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `PodSandboxId:` + fmt.Sprintf("%v", this.PodSandboxId) + `,`,
- `LabelSelector:` + mapStringForLabelSelector + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ListContainerStatsResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ListContainerStatsResponse{`,
- `Stats:` + strings.Replace(fmt.Sprintf("%v", this.Stats), "ContainerStats", "ContainerStats", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerAttributes) String() string {
- if this == nil {
- return "nil"
- }
- keysForLabels := make([]string, 0, len(this.Labels))
- for k := range this.Labels {
- keysForLabels = append(keysForLabels, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForLabels)
- mapStringForLabels := "map[string]string{"
- for _, k := range keysForLabels {
- mapStringForLabels += fmt.Sprintf("%v: %v,", k, this.Labels[k])
- }
- mapStringForLabels += "}"
- keysForAnnotations := make([]string, 0, len(this.Annotations))
- for k := range this.Annotations {
- keysForAnnotations = append(keysForAnnotations, k)
- }
- github_com_gogo_protobuf_sortkeys.Strings(keysForAnnotations)
- mapStringForAnnotations := "map[string]string{"
- for _, k := range keysForAnnotations {
- mapStringForAnnotations += fmt.Sprintf("%v: %v,", k, this.Annotations[k])
- }
- mapStringForAnnotations += "}"
- s := strings.Join([]string{`&ContainerAttributes{`,
- `Id:` + fmt.Sprintf("%v", this.Id) + `,`,
- `Metadata:` + strings.Replace(fmt.Sprintf("%v", this.Metadata), "ContainerMetadata", "ContainerMetadata", 1) + `,`,
- `Labels:` + mapStringForLabels + `,`,
- `Annotations:` + mapStringForAnnotations + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ContainerStats) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ContainerStats{`,
- `Attributes:` + strings.Replace(fmt.Sprintf("%v", this.Attributes), "ContainerAttributes", "ContainerAttributes", 1) + `,`,
- `Cpu:` + strings.Replace(fmt.Sprintf("%v", this.Cpu), "CpuUsage", "CpuUsage", 1) + `,`,
- `Memory:` + strings.Replace(fmt.Sprintf("%v", this.Memory), "MemoryUsage", "MemoryUsage", 1) + `,`,
- `WritableLayer:` + strings.Replace(fmt.Sprintf("%v", this.WritableLayer), "FilesystemUsage", "FilesystemUsage", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *CpuUsage) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&CpuUsage{`,
- `Timestamp:` + fmt.Sprintf("%v", this.Timestamp) + `,`,
- `UsageCoreNanoSeconds:` + strings.Replace(fmt.Sprintf("%v", this.UsageCoreNanoSeconds), "UInt64Value", "UInt64Value", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *MemoryUsage) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&MemoryUsage{`,
- `Timestamp:` + fmt.Sprintf("%v", this.Timestamp) + `,`,
- `WorkingSetBytes:` + strings.Replace(fmt.Sprintf("%v", this.WorkingSetBytes), "UInt64Value", "UInt64Value", 1) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ReopenContainerLogRequest) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ReopenContainerLogRequest{`,
- `ContainerId:` + fmt.Sprintf("%v", this.ContainerId) + `,`,
- `}`,
- }, "")
- return s
-}
-func (this *ReopenContainerLogResponse) String() string {
- if this == nil {
- return "nil"
- }
- s := strings.Join([]string{`&ReopenContainerLogResponse{`,
- `}`,
- }, "")
- return s
-}
-func valueToStringApi(v interface{}) string {
- rv := reflect.ValueOf(v)
- if rv.IsNil() {
- return "nil"
- }
- pv := reflect.Indirect(rv).Interface()
- return fmt.Sprintf("*%v", pv)
-}
-func (m *VersionRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: VersionRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: VersionRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Version", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Version = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *VersionResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: VersionResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: VersionResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Version", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Version = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RuntimeName", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RuntimeName = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RuntimeVersion", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RuntimeVersion = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RuntimeApiVersion", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RuntimeApiVersion = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *DNSConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: DNSConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: DNSConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Servers", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Servers = append(m.Servers, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Searches", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Searches = append(m.Searches, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Options", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Options = append(m.Options, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PortMapping) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PortMapping: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PortMapping: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Protocol", wireType)
- }
- m.Protocol = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Protocol |= (Protocol(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerPort", wireType)
- }
- m.ContainerPort = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.ContainerPort |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field HostPort", wireType)
- }
- m.HostPort = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.HostPort |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field HostIp", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.HostIp = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Mount) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Mount: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Mount: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field HostPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.HostPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Readonly", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Readonly = bool(v != 0)
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field SelinuxRelabel", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.SelinuxRelabel = bool(v != 0)
- case 5:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Propagation", wireType)
- }
- m.Propagation = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Propagation |= (MountPropagation(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *NamespaceOption) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: NamespaceOption: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: NamespaceOption: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Network", wireType)
- }
- m.Network = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Network |= (NamespaceMode(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Pid", wireType)
- }
- m.Pid = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Pid |= (NamespaceMode(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Ipc", wireType)
- }
- m.Ipc = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Ipc |= (NamespaceMode(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Int64Value) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Int64Value: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Int64Value: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Value", wireType)
- }
- m.Value = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Value |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxSandboxSecurityContext) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxSandboxSecurityContext: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxSandboxSecurityContext: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field NamespaceOptions", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.NamespaceOptions == nil {
- m.NamespaceOptions = &NamespaceOption{}
- }
- if err := m.NamespaceOptions.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SelinuxOptions", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SelinuxOptions == nil {
- m.SelinuxOptions = &SELinuxOption{}
- }
- if err := m.SelinuxOptions.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RunAsUser", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.RunAsUser == nil {
- m.RunAsUser = &Int64Value{}
- }
- if err := m.RunAsUser.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field ReadonlyRootfs", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.ReadonlyRootfs = bool(v != 0)
- case 5:
- if wireType == 0 {
- var v int64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.SupplementalGroups = append(m.SupplementalGroups, v)
- } else if wireType == 2 {
- var packedLen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- packedLen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if packedLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + packedLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- for iNdEx < postIndex {
- var v int64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.SupplementalGroups = append(m.SupplementalGroups, v)
- }
- } else {
- return fmt.Errorf("proto: wrong wireType = %d for field SupplementalGroups", wireType)
- }
- case 6:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Privileged", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Privileged = bool(v != 0)
- case 7:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SeccompProfilePath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.SeccompProfilePath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxPodSandboxConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxPodSandboxConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxPodSandboxConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field CgroupParent", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.CgroupParent = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SecurityContext", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SecurityContext == nil {
- m.SecurityContext = &LinuxSandboxSecurityContext{}
- }
- if err := m.SecurityContext.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Sysctls", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Sysctls == nil {
- m.Sysctls = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Sysctls[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Sysctls[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxMetadata) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxMetadata: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxMetadata: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Name = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Uid", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Uid = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Namespace", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Namespace = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Attempt", wireType)
- }
- m.Attempt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Attempt |= (uint32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &PodSandboxMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Hostname", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Hostname = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LogDirectory", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.LogDirectory = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field DnsConfig", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.DnsConfig == nil {
- m.DnsConfig = &DNSConfig{}
- }
- if err := m.DnsConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PortMappings", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PortMappings = append(m.PortMappings, &PortMapping{})
- if err := m.PortMappings[len(m.PortMappings)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 7:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 8:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Linux", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Linux == nil {
- m.Linux = &LinuxPodSandboxConfig{}
- }
- if err := m.Linux.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RunPodSandboxRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RunPodSandboxRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RunPodSandboxRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Config", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Config == nil {
- m.Config = &PodSandboxConfig{}
- }
- if err := m.Config.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RunPodSandboxResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RunPodSandboxResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RunPodSandboxResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StopPodSandboxRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StopPodSandboxRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StopPodSandboxRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StopPodSandboxResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StopPodSandboxResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StopPodSandboxResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemovePodSandboxRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemovePodSandboxRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemovePodSandboxRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemovePodSandboxResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemovePodSandboxResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemovePodSandboxResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxStatusRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxStatusRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxStatusRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Verbose", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Verbose = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxNetworkStatus) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxNetworkStatus: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxNetworkStatus: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Ip", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Ip = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Namespace) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Namespace: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Namespace: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Options", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Options == nil {
- m.Options = &NamespaceOption{}
- }
- if err := m.Options.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxPodSandboxStatus) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxPodSandboxStatus: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxPodSandboxStatus: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Namespaces", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Namespaces == nil {
- m.Namespaces = &Namespace{}
- }
- if err := m.Namespaces.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxStatus) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxStatus: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxStatus: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &PodSandboxMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (PodSandboxState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CreatedAt", wireType)
- }
- m.CreatedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CreatedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Network", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Network == nil {
- m.Network = &PodSandboxNetworkStatus{}
- }
- if err := m.Network.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Linux", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Linux == nil {
- m.Linux = &LinuxPodSandboxStatus{}
- }
- if err := m.Linux.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 7:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 8:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxStatusResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxStatusResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxStatusResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Status == nil {
- m.Status = &PodSandboxStatus{}
- }
- if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Info", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Info == nil {
- m.Info = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Info[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Info[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxStateValue) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxStateValue: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxStateValue: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (PodSandboxState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandboxFilter) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandboxFilter: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandboxFilter: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.State == nil {
- m.State = &PodSandboxStateValue{}
- }
- if err := m.State.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LabelSelector", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.LabelSelector == nil {
- m.LabelSelector = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.LabelSelector[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.LabelSelector[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListPodSandboxRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListPodSandboxRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListPodSandboxRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Filter", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Filter == nil {
- m.Filter = &PodSandboxFilter{}
- }
- if err := m.Filter.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PodSandbox) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PodSandbox: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PodSandbox: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &PodSandboxMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (PodSandboxState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CreatedAt", wireType)
- }
- m.CreatedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CreatedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListPodSandboxResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListPodSandboxResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListPodSandboxResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Items = append(m.Items, &PodSandbox{})
- if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageSpec) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageSpec: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageSpec: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Image = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *KeyValue) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: KeyValue: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: KeyValue: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Key", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Key = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Value", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Value = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxContainerResources) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxContainerResources: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxContainerResources: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuPeriod", wireType)
- }
- m.CpuPeriod = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuPeriod |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuQuota", wireType)
- }
- m.CpuQuota = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuQuota |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuShares", wireType)
- }
- m.CpuShares = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuShares |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field MemoryLimitInBytes", wireType)
- }
- m.MemoryLimitInBytes = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.MemoryLimitInBytes |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 5:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field OomScoreAdj", wireType)
- }
- m.OomScoreAdj = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.OomScoreAdj |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpusetCpus", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.CpusetCpus = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 7:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpusetMems", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.CpusetMems = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *SELinuxOption) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: SELinuxOption: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: SELinuxOption: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field User", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.User = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Role", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Role = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Type = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Level", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Level = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Capability) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Capability: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Capability: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field AddCapabilities", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.AddCapabilities = append(m.AddCapabilities, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field DropCapabilities", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.DropCapabilities = append(m.DropCapabilities, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxContainerSecurityContext) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxContainerSecurityContext: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxContainerSecurityContext: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Capabilities", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Capabilities == nil {
- m.Capabilities = &Capability{}
- }
- if err := m.Capabilities.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Privileged", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Privileged = bool(v != 0)
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field NamespaceOptions", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.NamespaceOptions == nil {
- m.NamespaceOptions = &NamespaceOption{}
- }
- if err := m.NamespaceOptions.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SelinuxOptions", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SelinuxOptions == nil {
- m.SelinuxOptions = &SELinuxOption{}
- }
- if err := m.SelinuxOptions.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RunAsUser", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.RunAsUser == nil {
- m.RunAsUser = &Int64Value{}
- }
- if err := m.RunAsUser.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RunAsUsername", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RunAsUsername = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 7:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field ReadonlyRootfs", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.ReadonlyRootfs = bool(v != 0)
- case 8:
- if wireType == 0 {
- var v int64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.SupplementalGroups = append(m.SupplementalGroups, v)
- } else if wireType == 2 {
- var packedLen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- packedLen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if packedLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + packedLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- for iNdEx < postIndex {
- var v int64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.SupplementalGroups = append(m.SupplementalGroups, v)
- }
- } else {
- return fmt.Errorf("proto: wrong wireType = %d for field SupplementalGroups", wireType)
- }
- case 9:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ApparmorProfile", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ApparmorProfile = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 10:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SeccompProfilePath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.SeccompProfilePath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 11:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field NoNewPrivs", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.NoNewPrivs = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *LinuxContainerConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: LinuxContainerConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: LinuxContainerConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Resources", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Resources == nil {
- m.Resources = &LinuxContainerResources{}
- }
- if err := m.Resources.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SecurityContext", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SecurityContext == nil {
- m.SecurityContext = &LinuxContainerSecurityContext{}
- }
- if err := m.SecurityContext.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *WindowsContainerConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: WindowsContainerConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: WindowsContainerConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Resources", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Resources == nil {
- m.Resources = &WindowsContainerResources{}
- }
- if err := m.Resources.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *WindowsContainerResources) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: WindowsContainerResources: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: WindowsContainerResources: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuShares", wireType)
- }
- m.CpuShares = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuShares |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuCount", wireType)
- }
- m.CpuCount = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuCount |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CpuMaximum", wireType)
- }
- m.CpuMaximum = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CpuMaximum |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field MemoryLimitInBytes", wireType)
- }
- m.MemoryLimitInBytes = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.MemoryLimitInBytes |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerMetadata) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerMetadata: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerMetadata: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Name = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Attempt", wireType)
- }
- m.Attempt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Attempt |= (uint32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Device) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Device: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Device: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field HostPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.HostPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Permissions", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Permissions = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &ContainerMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Command", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Command = append(m.Command, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Args", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Args = append(m.Args, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field WorkingDir", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.WorkingDir = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Envs", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Envs = append(m.Envs, &KeyValue{})
- if err := m.Envs[len(m.Envs)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 7:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Mounts", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Mounts = append(m.Mounts, &Mount{})
- if err := m.Mounts[len(m.Mounts)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 8:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Devices", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Devices = append(m.Devices, &Device{})
- if err := m.Devices[len(m.Devices)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 9:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 10:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 11:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LogPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.LogPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 12:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdin", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stdin = bool(v != 0)
- case 13:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field StdinOnce", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.StdinOnce = bool(v != 0)
- case 14:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Tty", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Tty = bool(v != 0)
- case 15:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Linux", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Linux == nil {
- m.Linux = &LinuxContainerConfig{}
- }
- if err := m.Linux.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 16:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Windows", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Windows == nil {
- m.Windows = &WindowsContainerConfig{}
- }
- if err := m.Windows.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *CreateContainerRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: CreateContainerRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: CreateContainerRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Config", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Config == nil {
- m.Config = &ContainerConfig{}
- }
- if err := m.Config.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SandboxConfig", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SandboxConfig == nil {
- m.SandboxConfig = &PodSandboxConfig{}
- }
- if err := m.SandboxConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *CreateContainerResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: CreateContainerResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: CreateContainerResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StartContainerRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StartContainerRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StartContainerRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StartContainerResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StartContainerResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StartContainerResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StopContainerRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StopContainerRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StopContainerRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Timeout", wireType)
- }
- m.Timeout = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Timeout |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StopContainerResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StopContainerResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StopContainerResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemoveContainerRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemoveContainerRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemoveContainerRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemoveContainerResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemoveContainerResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemoveContainerResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStateValue) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStateValue: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStateValue: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (ContainerState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerFilter) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerFilter: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerFilter: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.State == nil {
- m.State = &ContainerStateValue{}
- }
- if err := m.State.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LabelSelector", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.LabelSelector == nil {
- m.LabelSelector = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.LabelSelector[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.LabelSelector[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListContainersRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListContainersRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListContainersRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Filter", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Filter == nil {
- m.Filter = &ContainerFilter{}
- }
- if err := m.Filter.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Container) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Container: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Container: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &ContainerMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ImageRef", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ImageRef = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 6:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (ContainerState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 7:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CreatedAt", wireType)
- }
- m.CreatedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CreatedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 8:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 9:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListContainersResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListContainersResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListContainersResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Containers", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Containers = append(m.Containers, &Container{})
- if err := m.Containers[len(m.Containers)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatusRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatusRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatusRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Verbose", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Verbose = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatus) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatus: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatus: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &ContainerMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field State", wireType)
- }
- m.State = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.State |= (ContainerState(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field CreatedAt", wireType)
- }
- m.CreatedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.CreatedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 5:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field StartedAt", wireType)
- }
- m.StartedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.StartedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 6:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field FinishedAt", wireType)
- }
- m.FinishedAt = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.FinishedAt |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 7:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field ExitCode", wireType)
- }
- m.ExitCode = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.ExitCode |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 8:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 9:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ImageRef", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ImageRef = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 10:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Reason = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 11:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Message = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 12:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 13:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 14:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Mounts", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Mounts = append(m.Mounts, &Mount{})
- if err := m.Mounts[len(m.Mounts)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 15:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LogPath", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.LogPath = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatusResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatusResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatusResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Status == nil {
- m.Status = &ContainerStatus{}
- }
- if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Info", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Info == nil {
- m.Info = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Info[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Info[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *UpdateContainerResourcesRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: UpdateContainerResourcesRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: UpdateContainerResourcesRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Linux", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Linux == nil {
- m.Linux = &LinuxContainerResources{}
- }
- if err := m.Linux.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *UpdateContainerResourcesResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: UpdateContainerResourcesResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: UpdateContainerResourcesResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ExecSyncRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ExecSyncRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ExecSyncRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Cmd", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Cmd = append(m.Cmd, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Timeout", wireType)
- }
- m.Timeout = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Timeout |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ExecSyncResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ExecSyncResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ExecSyncResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdout", wireType)
- }
- var byteLen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- byteLen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if byteLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + byteLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Stdout = append(m.Stdout[:0], dAtA[iNdEx:postIndex]...)
- if m.Stdout == nil {
- m.Stdout = []byte{}
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stderr", wireType)
- }
- var byteLen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- byteLen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if byteLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + byteLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Stderr = append(m.Stderr[:0], dAtA[iNdEx:postIndex]...)
- if m.Stderr == nil {
- m.Stderr = []byte{}
- }
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field ExitCode", wireType)
- }
- m.ExitCode = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.ExitCode |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ExecRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ExecRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ExecRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Cmd", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Cmd = append(m.Cmd, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Tty", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Tty = bool(v != 0)
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdin", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stdin = bool(v != 0)
- case 5:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdout", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stdout = bool(v != 0)
- case 6:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stderr", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stderr = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ExecResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ExecResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ExecResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Url", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Url = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *AttachRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: AttachRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: AttachRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdin", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stdin = bool(v != 0)
- case 3:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Tty", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Tty = bool(v != 0)
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stdout", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stdout = bool(v != 0)
- case 5:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stderr", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Stderr = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *AttachResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: AttachResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: AttachResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Url", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Url = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PortForwardRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PortForwardRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PortForwardRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType == 0 {
- var v int32
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Port = append(m.Port, v)
- } else if wireType == 2 {
- var packedLen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- packedLen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if packedLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + packedLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- for iNdEx < postIndex {
- var v int32
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int32(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Port = append(m.Port, v)
- }
- } else {
- return fmt.Errorf("proto: wrong wireType = %d for field Port", wireType)
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PortForwardResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PortForwardResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PortForwardResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Url", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Url = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageFilter) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageFilter: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageFilter: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListImagesRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListImagesRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListImagesRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Filter", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Filter == nil {
- m.Filter = &ImageFilter{}
- }
- if err := m.Filter.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *Image) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: Image: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: Image: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RepoTags", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RepoTags = append(m.RepoTags, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RepoDigests", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RepoDigests = append(m.RepoDigests, string(dAtA[iNdEx:postIndex]))
- iNdEx = postIndex
- case 4:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Size_", wireType)
- }
- m.Size_ = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Size_ |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Uid", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Uid == nil {
- m.Uid = &Int64Value{}
- }
- if err := m.Uid.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Username", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Username = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListImagesResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListImagesResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListImagesResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Images", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Images = append(m.Images, &Image{})
- if err := m.Images[len(m.Images)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageStatusRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageStatusRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageStatusRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Verbose", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Verbose = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageStatusResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageStatusResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageStatusResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &Image{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Info", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Info == nil {
- m.Info = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Info[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Info[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *AuthConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: AuthConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: AuthConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Username", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Username = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Password", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Password = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Auth", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Auth = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ServerAddress", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ServerAddress = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 5:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field IdentityToken", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.IdentityToken = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 6:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RegistryToken", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.RegistryToken = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PullImageRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PullImageRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PullImageRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Auth", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Auth == nil {
- m.Auth = &AuthConfig{}
- }
- if err := m.Auth.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field SandboxConfig", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.SandboxConfig == nil {
- m.SandboxConfig = &PodSandboxConfig{}
- }
- if err := m.SandboxConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *PullImageResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: PullImageResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: PullImageResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ImageRef", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ImageRef = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemoveImageRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemoveImageRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemoveImageRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Image", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Image == nil {
- m.Image = &ImageSpec{}
- }
- if err := m.Image.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RemoveImageResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RemoveImageResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RemoveImageResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *NetworkConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: NetworkConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: NetworkConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodCidr", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodCidr = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RuntimeConfig) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RuntimeConfig: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RuntimeConfig: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field NetworkConfig", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.NetworkConfig == nil {
- m.NetworkConfig = &NetworkConfig{}
- }
- if err := m.NetworkConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *UpdateRuntimeConfigRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: UpdateRuntimeConfigRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: UpdateRuntimeConfigRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field RuntimeConfig", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.RuntimeConfig == nil {
- m.RuntimeConfig = &RuntimeConfig{}
- }
- if err := m.RuntimeConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *UpdateRuntimeConfigResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: UpdateRuntimeConfigResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: UpdateRuntimeConfigResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RuntimeCondition) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RuntimeCondition: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RuntimeCondition: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Type = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Status = bool(v != 0)
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Reason = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Message = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *RuntimeStatus) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: RuntimeStatus: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: RuntimeStatus: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Conditions", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Conditions = append(m.Conditions, &RuntimeCondition{})
- if err := m.Conditions[len(m.Conditions)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StatusRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StatusRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StatusRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Verbose", wireType)
- }
- var v int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- v |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- m.Verbose = bool(v != 0)
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *StatusResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: StatusResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: StatusResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Status == nil {
- m.Status = &RuntimeStatus{}
- }
- if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Info", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Info == nil {
- m.Info = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Info[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Info[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageFsInfoRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageFsInfoRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageFsInfoRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *UInt64Value) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: UInt64Value: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: UInt64Value: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Value", wireType)
- }
- m.Value = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Value |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *FilesystemIdentifier) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: FilesystemIdentifier: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: FilesystemIdentifier: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Mountpoint", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Mountpoint = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *FilesystemUsage) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: FilesystemUsage: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: FilesystemUsage: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Timestamp", wireType)
- }
- m.Timestamp = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Timestamp |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field FsId", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.FsId == nil {
- m.FsId = &FilesystemIdentifier{}
- }
- if err := m.FsId.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field UsedBytes", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.UsedBytes == nil {
- m.UsedBytes = &UInt64Value{}
- }
- if err := m.UsedBytes.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field InodesUsed", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.InodesUsed == nil {
- m.InodesUsed = &UInt64Value{}
- }
- if err := m.InodesUsed.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ImageFsInfoResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ImageFsInfoResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ImageFsInfoResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ImageFilesystems", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ImageFilesystems = append(m.ImageFilesystems, &FilesystemUsage{})
- if err := m.ImageFilesystems[len(m.ImageFilesystems)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatsRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatsRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatsRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatsResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatsResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatsResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stats", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Stats == nil {
- m.Stats = &ContainerStats{}
- }
- if err := m.Stats.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListContainerStatsRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListContainerStatsRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListContainerStatsRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Filter", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Filter == nil {
- m.Filter = &ContainerStatsFilter{}
- }
- if err := m.Filter.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStatsFilter) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStatsFilter: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStatsFilter: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field PodSandboxId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.PodSandboxId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field LabelSelector", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.LabelSelector == nil {
- m.LabelSelector = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.LabelSelector[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.LabelSelector[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ListContainerStatsResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ListContainerStatsResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ListContainerStatsResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Stats", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Stats = append(m.Stats, &ContainerStats{})
- if err := m.Stats[len(m.Stats)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerAttributes) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerAttributes: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerAttributes: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Id", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.Id = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Metadata", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Metadata == nil {
- m.Metadata = &ContainerMetadata{}
- }
- if err := m.Metadata.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Labels == nil {
- m.Labels = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Labels[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Labels[mapkey] = mapvalue
- }
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Annotations", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- var keykey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- keykey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapkey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapkey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapkey := int(stringLenmapkey)
- if intStringLenmapkey < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapkey := iNdEx + intStringLenmapkey
- if postStringIndexmapkey > l {
- return io.ErrUnexpectedEOF
- }
- mapkey := string(dAtA[iNdEx:postStringIndexmapkey])
- iNdEx = postStringIndexmapkey
- if m.Annotations == nil {
- m.Annotations = make(map[string]string)
- }
- if iNdEx < postIndex {
- var valuekey uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- valuekey |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- var stringLenmapvalue uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLenmapvalue |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLenmapvalue := int(stringLenmapvalue)
- if intStringLenmapvalue < 0 {
- return ErrInvalidLengthApi
- }
- postStringIndexmapvalue := iNdEx + intStringLenmapvalue
- if postStringIndexmapvalue > l {
- return io.ErrUnexpectedEOF
- }
- mapvalue := string(dAtA[iNdEx:postStringIndexmapvalue])
- iNdEx = postStringIndexmapvalue
- m.Annotations[mapkey] = mapvalue
- } else {
- var mapvalue string
- m.Annotations[mapkey] = mapvalue
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ContainerStats) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ContainerStats: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ContainerStats: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Attributes", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Attributes == nil {
- m.Attributes = &ContainerAttributes{}
- }
- if err := m.Attributes.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Cpu", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Cpu == nil {
- m.Cpu = &CpuUsage{}
- }
- if err := m.Cpu.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 3:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field Memory", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.Memory == nil {
- m.Memory = &MemoryUsage{}
- }
- if err := m.Memory.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- case 4:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field WritableLayer", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.WritableLayer == nil {
- m.WritableLayer = &FilesystemUsage{}
- }
- if err := m.WritableLayer.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *CpuUsage) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: CpuUsage: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: CpuUsage: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Timestamp", wireType)
- }
- m.Timestamp = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Timestamp |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field UsageCoreNanoSeconds", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.UsageCoreNanoSeconds == nil {
- m.UsageCoreNanoSeconds = &UInt64Value{}
- }
- if err := m.UsageCoreNanoSeconds.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *MemoryUsage) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: MemoryUsage: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: MemoryUsage: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 0 {
- return fmt.Errorf("proto: wrong wireType = %d for field Timestamp", wireType)
- }
- m.Timestamp = 0
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- m.Timestamp |= (int64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- case 2:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field WorkingSetBytes", wireType)
- }
- var msglen int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- msglen |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- if msglen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + msglen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- if m.WorkingSetBytes == nil {
- m.WorkingSetBytes = &UInt64Value{}
- }
- if err := m.WorkingSetBytes.Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
- return err
- }
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ReopenContainerLogRequest) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ReopenContainerLogRequest: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ReopenContainerLogRequest: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- case 1:
- if wireType != 2 {
- return fmt.Errorf("proto: wrong wireType = %d for field ContainerId", wireType)
- }
- var stringLen uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- stringLen |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- intStringLen := int(stringLen)
- if intStringLen < 0 {
- return ErrInvalidLengthApi
- }
- postIndex := iNdEx + intStringLen
- if postIndex > l {
- return io.ErrUnexpectedEOF
- }
- m.ContainerId = string(dAtA[iNdEx:postIndex])
- iNdEx = postIndex
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func (m *ReopenContainerLogResponse) Unmarshal(dAtA []byte) error {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- preIndex := iNdEx
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return ErrIntOverflowApi
- }
- if iNdEx >= l {
- return io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- fieldNum := int32(wire >> 3)
- wireType := int(wire & 0x7)
- if wireType == 4 {
- return fmt.Errorf("proto: ReopenContainerLogResponse: wiretype end group for non-group")
- }
- if fieldNum <= 0 {
- return fmt.Errorf("proto: ReopenContainerLogResponse: illegal tag %d (wire type %d)", fieldNum, wire)
- }
- switch fieldNum {
- default:
- iNdEx = preIndex
- skippy, err := skipApi(dAtA[iNdEx:])
- if err != nil {
- return err
- }
- if skippy < 0 {
- return ErrInvalidLengthApi
- }
- if (iNdEx + skippy) > l {
- return io.ErrUnexpectedEOF
- }
- iNdEx += skippy
- }
- }
-
- if iNdEx > l {
- return io.ErrUnexpectedEOF
- }
- return nil
-}
-func skipApi(dAtA []byte) (n int, err error) {
- l := len(dAtA)
- iNdEx := 0
- for iNdEx < l {
- var wire uint64
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return 0, ErrIntOverflowApi
- }
- if iNdEx >= l {
- return 0, io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- wire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- wireType := int(wire & 0x7)
- switch wireType {
- case 0:
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return 0, ErrIntOverflowApi
- }
- if iNdEx >= l {
- return 0, io.ErrUnexpectedEOF
- }
- iNdEx++
- if dAtA[iNdEx-1] < 0x80 {
- break
- }
- }
- return iNdEx, nil
- case 1:
- iNdEx += 8
- return iNdEx, nil
- case 2:
- var length int
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return 0, ErrIntOverflowApi
- }
- if iNdEx >= l {
- return 0, io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- length |= (int(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- iNdEx += length
- if length < 0 {
- return 0, ErrInvalidLengthApi
- }
- return iNdEx, nil
- case 3:
- for {
- var innerWire uint64
- var start int = iNdEx
- for shift := uint(0); ; shift += 7 {
- if shift >= 64 {
- return 0, ErrIntOverflowApi
- }
- if iNdEx >= l {
- return 0, io.ErrUnexpectedEOF
- }
- b := dAtA[iNdEx]
- iNdEx++
- innerWire |= (uint64(b) & 0x7F) << shift
- if b < 0x80 {
- break
- }
- }
- innerWireType := int(innerWire & 0x7)
- if innerWireType == 4 {
- break
- }
- next, err := skipApi(dAtA[start:])
- if err != nil {
- return 0, err
- }
- iNdEx = start + next
- }
- return iNdEx, nil
- case 4:
- return iNdEx, nil
- case 5:
- iNdEx += 4
- return iNdEx, nil
- default:
- return 0, fmt.Errorf("proto: illegal wireType %d", wireType)
- }
- }
- panic("unreachable")
-}
-
-var (
- ErrInvalidLengthApi = fmt.Errorf("proto: negative length found during unmarshaling")
- ErrIntOverflowApi = fmt.Errorf("proto: integer overflow")
-)
-
-func init() { proto.RegisterFile("api.proto", fileDescriptorApi) }
-
-var fileDescriptorApi = []byte{
- // 4602 bytes of a gzipped FileDescriptorProto
- 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x5c, 0xcd, 0x6f, 0x1b, 0x49,
- 0x76, 0x17, 0x49, 0x7d, 0x90, 0x8f, 0xa2, 0x44, 0x95, 0x65, 0x8b, 0xa6, 0x6d, 0x8d, 0xd5, 0x1e,
- 0x7f, 0xce, 0x58, 0x1e, 0x6b, 0x76, 0x3d, 0xb1, 0x3d, 0x6b, 0x9b, 0x96, 0x64, 0x9b, 0x59, 0x9b,
- 0x52, 0x9a, 0xd2, 0x7c, 0xec, 0x2c, 0xd0, 0xdb, 0x62, 0x97, 0xa8, 0x5e, 0x93, 0x5d, 0x3d, 0xdd,
- 0x4d, 0xdb, 0xca, 0x21, 0x58, 0x20, 0xc8, 0x1e, 0x02, 0x04, 0xc8, 0x39, 0xb7, 0x6c, 0x0e, 0x39,
- 0xe4, 0x16, 0x20, 0xc8, 0x21, 0xa7, 0x04, 0x7b, 0xd8, 0x4b, 0x80, 0x9c, 0x16, 0xf9, 0xb8, 0x64,
- 0x27, 0xc8, 0x25, 0x87, 0x20, 0x7f, 0x40, 0x0e, 0x41, 0x7d, 0xf5, 0x77, 0xf3, 0x43, 0xe3, 0x9d,
- 0x99, 0x9c, 0xd4, 0xf5, 0xfa, 0xbd, 0x57, 0xaf, 0x5e, 0xbd, 0x7e, 0xf5, 0xea, 0x57, 0x45, 0x41,
- 0x49, 0xb7, 0xcd, 0x75, 0xdb, 0x21, 0x1e, 0x41, 0x55, 0x67, 0x60, 0x79, 0x66, 0x1f, 0xaf, 0xbf,
- 0xba, 0xad, 0xf7, 0xec, 0x23, 0x7d, 0xa3, 0x7e, 0xb3, 0x6b, 0x7a, 0x47, 0x83, 0x83, 0xf5, 0x0e,
- 0xe9, 0xdf, 0xea, 0x92, 0x2e, 0xb9, 0xc5, 0x18, 0x0f, 0x06, 0x87, 0xac, 0xc5, 0x1a, 0xec, 0x89,
- 0x2b, 0x50, 0x6e, 0xc0, 0xc2, 0x27, 0xd8, 0x71, 0x4d, 0x62, 0xa9, 0xf8, 0xcb, 0x01, 0x76, 0x3d,
- 0x54, 0x83, 0xb9, 0x57, 0x9c, 0x52, 0xcb, 0x5d, 0xcc, 0x5d, 0x2b, 0xa9, 0xb2, 0xa9, 0xfc, 0x65,
- 0x0e, 0x16, 0x7d, 0x66, 0xd7, 0x26, 0x96, 0x8b, 0xb3, 0xb9, 0xd1, 0x1a, 0xcc, 0x0b, 0xe3, 0x34,
- 0x4b, 0xef, 0xe3, 0x5a, 0x9e, 0xbd, 0x2e, 0x0b, 0x5a, 0x4b, 0xef, 0x63, 0x74, 0x15, 0x16, 0x25,
- 0x8b, 0x54, 0x52, 0x60, 0x5c, 0x0b, 0x82, 0x2c, 0x7a, 0x43, 0xeb, 0x70, 0x4a, 0x32, 0xea, 0xb6,
- 0xe9, 0x33, 0x4f, 0x33, 0xe6, 0x25, 0xf1, 0xaa, 0x61, 0x9b, 0x82, 0x5f, 0xf9, 0x02, 0x4a, 0x5b,
- 0xad, 0xf6, 0x26, 0xb1, 0x0e, 0xcd, 0x2e, 0x35, 0xd1, 0xc5, 0x0e, 0x95, 0xa9, 0xe5, 0x2e, 0x16,
- 0xa8, 0x89, 0xa2, 0x89, 0xea, 0x50, 0x74, 0xb1, 0xee, 0x74, 0x8e, 0xb0, 0x5b, 0xcb, 0xb3, 0x57,
- 0x7e, 0x9b, 0x4a, 0x11, 0xdb, 0x33, 0x89, 0xe5, 0xd6, 0x0a, 0x5c, 0x4a, 0x34, 0x95, 0x5f, 0xe4,
- 0xa0, 0xbc, 0x4b, 0x1c, 0xef, 0x85, 0x6e, 0xdb, 0xa6, 0xd5, 0x45, 0x77, 0xa0, 0xc8, 0x7c, 0xd9,
- 0x21, 0x3d, 0xe6, 0x83, 0x85, 0x8d, 0xfa, 0x7a, 0x7c, 0x5a, 0xd6, 0x77, 0x05, 0x87, 0xea, 0xf3,
- 0xa2, 0xcb, 0xb0, 0xd0, 0x21, 0x96, 0xa7, 0x9b, 0x16, 0x76, 0x34, 0x9b, 0x38, 0x1e, 0x73, 0xd1,
- 0x8c, 0x5a, 0xf1, 0xa9, 0xb4, 0x17, 0x74, 0x0e, 0x4a, 0x47, 0xc4, 0xf5, 0x38, 0x47, 0x81, 0x71,
- 0x14, 0x29, 0x81, 0xbd, 0x5c, 0x81, 0x39, 0xf6, 0xd2, 0xb4, 0x85, 0x33, 0x66, 0x69, 0xb3, 0x69,
- 0x2b, 0xbf, 0xce, 0xc1, 0xcc, 0x0b, 0x32, 0xb0, 0xbc, 0x58, 0x37, 0xba, 0x77, 0x24, 0x26, 0x2a,
- 0xd4, 0x8d, 0xee, 0x1d, 0x05, 0xdd, 0x50, 0x0e, 0x3e, 0x57, 0xbc, 0x1b, 0xfa, 0xb2, 0x0e, 0x45,
- 0x07, 0xeb, 0x06, 0xb1, 0x7a, 0xc7, 0xcc, 0x84, 0xa2, 0xea, 0xb7, 0xe9, 0x24, 0xba, 0xb8, 0x67,
- 0x5a, 0x83, 0x37, 0x9a, 0x83, 0x7b, 0xfa, 0x01, 0xee, 0x31, 0x53, 0x8a, 0xea, 0x82, 0x20, 0xab,
- 0x9c, 0x8a, 0xb6, 0xa0, 0x6c, 0x3b, 0xc4, 0xd6, 0xbb, 0x3a, 0xf5, 0x63, 0x6d, 0x86, 0xb9, 0x4a,
- 0x49, 0xba, 0x8a, 0x99, 0xbd, 0x1b, 0x70, 0xaa, 0x61, 0x31, 0xe5, 0xaf, 0x73, 0xb0, 0x48, 0x83,
- 0xc7, 0xb5, 0xf5, 0x0e, 0xde, 0x61, 0x53, 0x82, 0xee, 0xc2, 0x9c, 0x85, 0xbd, 0xd7, 0xc4, 0x79,
- 0x29, 0x26, 0xe0, 0x9d, 0xa4, 0x56, 0x5f, 0xe6, 0x05, 0x31, 0xb0, 0x2a, 0xf9, 0xd1, 0x6d, 0x28,
- 0xd8, 0xa6, 0xc1, 0x06, 0x3c, 0x86, 0x18, 0xe5, 0xa5, 0x22, 0xa6, 0xdd, 0x61, 0x7e, 0x18, 0x47,
- 0xc4, 0xb4, 0x3b, 0x8a, 0x02, 0xd0, 0xb4, 0xbc, 0x3b, 0xdf, 0xfb, 0x44, 0xef, 0x0d, 0x30, 0x5a,
- 0x86, 0x99, 0x57, 0xf4, 0x81, 0x19, 0x5b, 0x50, 0x79, 0x43, 0xf9, 0xf3, 0x02, 0x9c, 0x7b, 0x4e,
- 0xfd, 0xd5, 0xd6, 0x2d, 0xe3, 0x80, 0xbc, 0x69, 0xe3, 0xce, 0xc0, 0x31, 0xbd, 0xe3, 0x4d, 0x62,
- 0x79, 0xf8, 0x8d, 0x87, 0x5a, 0xb0, 0x64, 0x49, 0xcd, 0x9a, 0x0c, 0x4d, 0xaa, 0xa1, 0xbc, 0xb1,
- 0x36, 0xc4, 0x08, 0xee, 0x22, 0xb5, 0x6a, 0x45, 0x09, 0x2e, 0x7a, 0x16, 0xcc, 0x9b, 0xd4, 0x96,
- 0x67, 0xda, 0x52, 0x86, 0xd4, 0xde, 0x66, 0x96, 0x09, 0x5d, 0x72, 0x62, 0xa5, 0xa6, 0x8f, 0x81,
- 0x7e, 0xd5, 0x9a, 0xee, 0x6a, 0x03, 0x17, 0x3b, 0xcc, 0x31, 0xe5, 0x8d, 0xf3, 0x49, 0x2d, 0x81,
- 0x0b, 0xd4, 0x92, 0x33, 0xb0, 0x1a, 0xee, 0xbe, 0x8b, 0x1d, 0x96, 0x04, 0x44, 0x2c, 0x69, 0x0e,
- 0x21, 0xde, 0xa1, 0x2b, 0xe3, 0x47, 0x92, 0x55, 0x46, 0x45, 0xb7, 0xe0, 0x94, 0x3b, 0xb0, 0xed,
- 0x1e, 0xee, 0x63, 0xcb, 0xd3, 0x7b, 0x5a, 0xd7, 0x21, 0x03, 0xdb, 0xad, 0xcd, 0x5c, 0x2c, 0x5c,
- 0x2b, 0xa8, 0x28, 0xfc, 0xea, 0x29, 0x7b, 0x83, 0x56, 0x01, 0x6c, 0xc7, 0x7c, 0x65, 0xf6, 0x70,
- 0x17, 0x1b, 0xb5, 0x59, 0xa6, 0x34, 0x44, 0x41, 0x1f, 0xc0, 0xb2, 0x8b, 0x3b, 0x1d, 0xd2, 0xb7,
- 0x35, 0xdb, 0x21, 0x87, 0x66, 0x0f, 0xf3, 0xe8, 0x9f, 0x63, 0xd1, 0x8f, 0xc4, 0xbb, 0x5d, 0xfe,
- 0x8a, 0x7e, 0x07, 0xca, 0x2f, 0xf2, 0x70, 0x9a, 0x79, 0x62, 0x97, 0x18, 0x62, 0x9a, 0x44, 0x92,
- 0xb9, 0x04, 0x95, 0x0e, 0x33, 0x48, 0xb3, 0x75, 0x07, 0x5b, 0x9e, 0xf8, 0xc8, 0xe6, 0x39, 0x71,
- 0x97, 0xd1, 0xd0, 0x67, 0x50, 0x75, 0xc5, 0xac, 0x6a, 0x1d, 0x3e, 0xad, 0xc2, 0xe7, 0x37, 0x93,
- 0xde, 0x1a, 0x12, 0x0b, 0xea, 0xa2, 0x9b, 0x08, 0x8e, 0x39, 0xf7, 0xd8, 0xed, 0x78, 0x3d, 0x9e,
- 0xad, 0xca, 0x1b, 0xdf, 0xcb, 0x50, 0x18, 0x37, 0x7c, 0xbd, 0xcd, 0xc5, 0xb6, 0x2d, 0xcf, 0x39,
- 0x56, 0xa5, 0x92, 0xfa, 0x3d, 0x98, 0x0f, 0xbf, 0x40, 0x55, 0x28, 0xbc, 0xc4, 0xc7, 0x62, 0x50,
- 0xf4, 0x31, 0x08, 0x62, 0x9e, 0x2b, 0x78, 0xe3, 0x5e, 0xfe, 0x77, 0x72, 0x8a, 0x03, 0x28, 0xe8,
- 0xe5, 0x05, 0xf6, 0x74, 0x43, 0xf7, 0x74, 0x84, 0x60, 0x9a, 0x2d, 0x03, 0x5c, 0x05, 0x7b, 0xa6,
- 0x5a, 0x07, 0xe2, 0xe3, 0x2b, 0xa9, 0xf4, 0x11, 0x9d, 0x87, 0x92, 0x1f, 0xa8, 0x62, 0x2d, 0x08,
- 0x08, 0x34, 0x27, 0xeb, 0x9e, 0x87, 0xfb, 0xb6, 0xc7, 0x42, 0xa4, 0xa2, 0xca, 0xa6, 0xf2, 0xdf,
- 0xd3, 0x50, 0x4d, 0xcc, 0xc9, 0x23, 0x28, 0xf6, 0x45, 0xf7, 0xe2, 0x43, 0x79, 0x37, 0x25, 0x31,
- 0x27, 0x4c, 0x55, 0x7d, 0x29, 0x9a, 0xf7, 0x68, 0x0e, 0x0c, 0xad, 0x5f, 0x7e, 0x9b, 0xce, 0x78,
- 0x8f, 0x74, 0x35, 0xc3, 0x74, 0x70, 0xc7, 0x23, 0xce, 0xb1, 0x30, 0x77, 0xbe, 0x47, 0xba, 0x5b,
- 0x92, 0x86, 0xee, 0x01, 0x18, 0x96, 0x4b, 0x27, 0xfb, 0xd0, 0xec, 0x32, 0xa3, 0xcb, 0x1b, 0xe7,
- 0x92, 0x46, 0xf8, 0x8b, 0x95, 0x5a, 0x32, 0x2c, 0x57, 0x98, 0xff, 0x18, 0x2a, 0x34, 0xe7, 0x6b,
- 0x7d, 0xbe, 0xce, 0xf0, 0x48, 0x2f, 0x6f, 0x5c, 0x48, 0x1b, 0x83, 0xbf, 0x1a, 0xa9, 0xf3, 0x76,
- 0xd0, 0x70, 0xd1, 0x13, 0x98, 0x65, 0xc9, 0xd7, 0xad, 0xcd, 0x32, 0xe1, 0xf5, 0x61, 0x0e, 0x10,
- 0x11, 0xf1, 0x9c, 0x09, 0xf0, 0x80, 0x10, 0xd2, 0x68, 0x1f, 0xca, 0xba, 0x65, 0x11, 0x4f, 0xe7,
- 0x89, 0x62, 0x8e, 0x29, 0xfb, 0x70, 0x0c, 0x65, 0x8d, 0x40, 0x8a, 0x6b, 0x0c, 0xeb, 0x41, 0x3f,
- 0x80, 0x19, 0x96, 0x49, 0x6a, 0x45, 0xe6, 0x99, 0xab, 0x63, 0x06, 0xad, 0xca, 0xa5, 0xea, 0x77,
- 0xa1, 0x1c, 0x32, 0x76, 0x92, 0x20, 0xad, 0x3f, 0x80, 0x6a, 0xdc, 0xb4, 0x89, 0x82, 0x5c, 0x85,
- 0x65, 0x75, 0x60, 0x05, 0x86, 0xc9, 0xea, 0xe9, 0x1e, 0xcc, 0x8a, 0xc9, 0xe6, 0x11, 0xa7, 0x8c,
- 0xf6, 0x91, 0x2a, 0x24, 0x94, 0x1f, 0xc0, 0xe9, 0x98, 0x4e, 0x51, 0x64, 0xbd, 0x0b, 0x0b, 0x36,
- 0x31, 0x34, 0x97, 0x93, 0x35, 0xd3, 0x90, 0xd9, 0xc5, 0xf6, 0x79, 0x9b, 0x06, 0x15, 0x6f, 0x7b,
- 0xc4, 0x4e, 0xda, 0x34, 0x9e, 0x78, 0x0d, 0xce, 0xc4, 0xc5, 0x79, 0xf7, 0xca, 0x43, 0x58, 0x51,
- 0x71, 0x9f, 0xbc, 0xc2, 0x27, 0x55, 0x5d, 0x87, 0x5a, 0x52, 0x81, 0x50, 0xfe, 0x39, 0xac, 0x04,
- 0xd4, 0xb6, 0xa7, 0x7b, 0x03, 0x77, 0x22, 0xe5, 0xa2, 0x02, 0x3d, 0x20, 0x2e, 0x9f, 0xa5, 0xa2,
- 0x2a, 0x9b, 0xca, 0xf5, 0xb0, 0xea, 0x16, 0x5f, 0xf0, 0x79, 0x0f, 0x68, 0x01, 0xf2, 0xa6, 0x2d,
- 0xd4, 0xe5, 0x4d, 0x5b, 0x79, 0x06, 0x25, 0x7f, 0xc5, 0x44, 0xf7, 0x83, 0xd2, 0x2f, 0x3f, 0xee,
- 0xfa, 0xea, 0x57, 0x87, 0x7b, 0x89, 0x15, 0x42, 0x74, 0x79, 0x1f, 0xc0, 0xcf, 0x64, 0x72, 0xe1,
- 0x3e, 0x37, 0x44, 0xb1, 0x1a, 0x62, 0x57, 0xfe, 0x35, 0x92, 0xdf, 0x42, 0x83, 0x30, 0xfc, 0x41,
- 0x18, 0x91, 0x7c, 0x97, 0x3f, 0x51, 0xbe, 0xfb, 0x08, 0x66, 0x5c, 0x4f, 0xf7, 0xb0, 0x28, 0x6e,
- 0xd6, 0x86, 0x89, 0x53, 0x23, 0xb0, 0xca, 0xf9, 0xd1, 0x05, 0x80, 0x8e, 0x83, 0x75, 0x0f, 0x1b,
- 0x9a, 0xce, 0x93, 0x73, 0x41, 0x2d, 0x09, 0x4a, 0xc3, 0x43, 0x9b, 0x41, 0x81, 0x36, 0xc3, 0x0c,
- 0xbb, 0x3e, 0x4c, 0x73, 0x64, 0xaa, 0x82, 0x52, 0xcd, 0x4f, 0x16, 0xb3, 0x63, 0x26, 0x0b, 0xa1,
- 0x80, 0x4b, 0x85, 0x52, 0xe1, 0xdc, 0xe8, 0x54, 0xc8, 0x45, 0xc7, 0x49, 0x85, 0xc5, 0xd1, 0xa9,
- 0x50, 0x28, 0x1b, 0x9a, 0x0a, 0xbf, 0xcd, 0x5c, 0xf6, 0x2f, 0x39, 0xa8, 0x25, 0xbf, 0x41, 0x91,
- 0x7b, 0xee, 0xc1, 0xac, 0xcb, 0x28, 0xe3, 0x24, 0x34, 0x21, 0x2b, 0x24, 0xd0, 0x33, 0x98, 0x36,
- 0xad, 0x43, 0xc2, 0xf6, 0x56, 0xa9, 0x25, 0x49, 0x56, 0xaf, 0xeb, 0x4d, 0xeb, 0x90, 0x70, 0x27,
- 0x31, 0x0d, 0xf5, 0x8f, 0xa0, 0xe4, 0x93, 0x26, 0x1a, 0xdb, 0x0e, 0x2c, 0xc7, 0x42, 0x96, 0xd7,
- 0xe0, 0x7e, 0xa4, 0xe7, 0x26, 0x8b, 0x74, 0xe5, 0x67, 0xf9, 0xf0, 0x97, 0xf8, 0xc4, 0xec, 0x79,
- 0xd8, 0x49, 0x7c, 0x89, 0x1f, 0x4b, 0xed, 0xfc, 0x33, 0xbc, 0x32, 0x52, 0x3b, 0xaf, 0x8a, 0xc5,
- 0xc7, 0xf4, 0x63, 0x58, 0x60, 0xb1, 0xa6, 0xb9, 0xb8, 0xc7, 0xea, 0x08, 0x51, 0xd3, 0x7d, 0x7f,
- 0x98, 0x1a, 0x6e, 0x09, 0x8f, 0xd8, 0xb6, 0x90, 0xe3, 0x1e, 0xac, 0xf4, 0xc2, 0xb4, 0xfa, 0x23,
- 0x40, 0x49, 0xa6, 0x89, 0x7c, 0xda, 0xa6, 0x29, 0x8e, 0x6e, 0x40, 0x53, 0x16, 0xbf, 0x43, 0x66,
- 0xc6, 0x38, 0xb1, 0xc2, 0x0d, 0x56, 0x85, 0x84, 0xf2, 0xcb, 0x02, 0x40, 0xf0, 0xf2, 0xff, 0x51,
- 0x6e, 0x7b, 0xe4, 0xe7, 0x15, 0x5e, 0x9f, 0x5d, 0x1b, 0xa6, 0x38, 0x35, 0xa3, 0xec, 0x44, 0x33,
- 0x0a, 0xaf, 0xd4, 0x6e, 0x0e, 0x55, 0xf3, 0x9d, 0xcd, 0x25, 0xcf, 0xe1, 0x4c, 0x3c, 0x36, 0x44,
- 0x22, 0xd9, 0x80, 0x19, 0xd3, 0xc3, 0x7d, 0x0e, 0xc2, 0xa4, 0xee, 0x0f, 0x43, 0x42, 0x9c, 0x55,
- 0x59, 0x83, 0x52, 0xb3, 0xaf, 0x77, 0x71, 0xdb, 0xc6, 0x1d, 0xda, 0xa9, 0x49, 0x1b, 0xc2, 0x10,
- 0xde, 0x50, 0x36, 0xa0, 0xf8, 0x43, 0x7c, 0xcc, 0x3f, 0xea, 0x31, 0x0d, 0x55, 0xfe, 0x24, 0x0f,
- 0x2b, 0x6c, 0xad, 0xd8, 0x94, 0x10, 0x88, 0x8a, 0x5d, 0x32, 0x70, 0x3a, 0xd8, 0x65, 0xb3, 0x6d,
- 0x0f, 0x34, 0x1b, 0x3b, 0x26, 0x31, 0xc4, 0x0e, 0xbd, 0xd4, 0xb1, 0x07, 0xbb, 0x8c, 0x80, 0xce,
- 0x01, 0x6d, 0x68, 0x5f, 0x0e, 0x88, 0x08, 0xc4, 0x82, 0x5a, 0xec, 0xd8, 0x83, 0xdf, 0xa3, 0x6d,
- 0x29, 0xeb, 0x1e, 0xe9, 0x0e, 0x76, 0x59, 0x9c, 0x71, 0xd9, 0x36, 0x23, 0xa0, 0xdb, 0x70, 0xba,
- 0x8f, 0xfb, 0xc4, 0x39, 0xd6, 0x7a, 0x66, 0xdf, 0xf4, 0x34, 0xd3, 0xd2, 0x0e, 0x8e, 0x3d, 0xec,
- 0x8a, 0x98, 0x42, 0xfc, 0xe5, 0x73, 0xfa, 0xae, 0x69, 0x3d, 0xa6, 0x6f, 0x90, 0x02, 0x15, 0x42,
- 0xfa, 0x9a, 0xdb, 0x21, 0x0e, 0xd6, 0x74, 0xe3, 0xa7, 0x6c, 0xf9, 0x2c, 0xa8, 0x65, 0x42, 0xfa,
- 0x6d, 0x4a, 0x6b, 0x18, 0x3f, 0x45, 0xef, 0x40, 0xb9, 0x63, 0x0f, 0x5c, 0xec, 0x69, 0xf4, 0x0f,
- 0x5b, 0x1d, 0x4b, 0x2a, 0x70, 0xd2, 0xa6, 0x3d, 0x70, 0x43, 0x0c, 0x7d, 0xea, 0xff, 0xb9, 0x30,
- 0xc3, 0x0b, 0xea, 0x66, 0x1d, 0x2a, 0x91, 0x1d, 0x3e, 0xdd, 0xac, 0xb1, 0xad, 0xbc, 0xd8, 0xac,
- 0xd1, 0x67, 0x4a, 0x73, 0x48, 0x4f, 0x7a, 0x92, 0x3d, 0x53, 0x9a, 0x77, 0x6c, 0xcb, 0x9d, 0x1a,
- 0x7b, 0xa6, 0x2e, 0xef, 0xe1, 0x57, 0x02, 0x05, 0x2a, 0xa9, 0xbc, 0xa1, 0x18, 0x00, 0x9b, 0xba,
- 0xad, 0x1f, 0x98, 0x3d, 0xd3, 0x3b, 0x46, 0xd7, 0xa1, 0xaa, 0x1b, 0x86, 0xd6, 0x91, 0x14, 0x13,
- 0x4b, 0x6c, 0x6e, 0x51, 0x37, 0x8c, 0xcd, 0x10, 0x19, 0xbd, 0x07, 0x4b, 0x86, 0x43, 0xec, 0x28,
- 0x2f, 0x07, 0xeb, 0xaa, 0xf4, 0x45, 0x98, 0x99, 0x96, 0x49, 0x17, 0xa2, 0x13, 0x1b, 0x47, 0x51,
- 0x1e, 0xc1, 0x7c, 0xac, 0xd7, 0x0c, 0xb0, 0x22, 0xb0, 0x56, 0x8d, 0x48, 0xc4, 0x50, 0x85, 0x7c,
- 0x02, 0x55, 0x48, 0xc5, 0x69, 0x0a, 0x6f, 0x15, 0xa7, 0x99, 0x7e, 0x2b, 0x38, 0xcd, 0xcc, 0x64,
- 0x38, 0xcd, 0x15, 0x06, 0xd6, 0x4a, 0x69, 0xb6, 0x25, 0xe6, 0xa1, 0x56, 0xf1, 0x79, 0x2c, 0x09,
- 0xea, 0xc6, 0xf0, 0x9c, 0xb9, 0x49, 0xf0, 0x9c, 0x62, 0x26, 0x9e, 0x43, 0xa3, 0xc6, 0xb6, 0x75,
- 0xa7, 0x4f, 0x1c, 0x09, 0xd8, 0xd4, 0x4a, 0xcc, 0x84, 0x45, 0x49, 0x17, 0x60, 0x4d, 0x26, 0xb4,
- 0x03, 0x59, 0xd0, 0x0e, 0xba, 0x08, 0xf3, 0x16, 0xd1, 0x2c, 0xfc, 0x5a, 0xa3, 0x73, 0xe9, 0xd6,
- 0xca, 0x7c, 0x62, 0x2d, 0xd2, 0xc2, 0xaf, 0x77, 0x29, 0x45, 0xf9, 0xbb, 0x1c, 0x2c, 0x47, 0x83,
- 0x4b, 0x6c, 0xd4, 0x9f, 0x42, 0xc9, 0x91, 0xf9, 0x43, 0x04, 0xd4, 0xf5, 0x8c, 0xe2, 0x34, 0x99,
- 0x70, 0xd4, 0x40, 0x16, 0xfd, 0x28, 0x13, 0x1f, 0xba, 0x35, 0x4a, 0xdf, 0x28, 0x84, 0x48, 0xe9,
- 0xc0, 0x99, 0x4f, 0x4d, 0xcb, 0x20, 0xaf, 0xdd, 0xb8, 0xf9, 0xcd, 0xa4, 0xf9, 0xef, 0x25, 0xbb,
- 0x8b, 0x0b, 0xa7, 0x0d, 0x40, 0xf9, 0xab, 0x1c, 0x9c, 0xcd, 0x64, 0x8c, 0xa5, 0xc7, 0x5c, 0x3c,
- 0x3d, 0x8a, 0xd4, 0xda, 0x21, 0x03, 0xcb, 0x0b, 0xa5, 0xd6, 0x4d, 0x86, 0x62, 0xf3, 0x1c, 0xa6,
- 0xf5, 0xf5, 0x37, 0x66, 0x7f, 0xd0, 0x17, 0xb9, 0x95, 0xaa, 0x7b, 0xc1, 0x29, 0x27, 0x48, 0xae,
- 0x4a, 0x03, 0x96, 0x7c, 0x2b, 0x87, 0xe2, 0x54, 0x21, 0xdc, 0x29, 0x1f, 0xc5, 0x9d, 0x2c, 0x98,
- 0xdd, 0xc2, 0xaf, 0xcc, 0x0e, 0x7e, 0x2b, 0x30, 0xfb, 0x45, 0x28, 0xdb, 0xd8, 0xe9, 0x9b, 0xae,
- 0xeb, 0x27, 0x8d, 0x92, 0x1a, 0x26, 0x29, 0xff, 0x39, 0x0b, 0x8b, 0xf1, 0xf9, 0x7b, 0x98, 0x80,
- 0xb9, 0x2e, 0xa5, 0xa4, 0xb3, 0xf8, 0x40, 0x43, 0x95, 0xd1, 0x6d, 0xb9, 0xb0, 0xe6, 0xb3, 0x36,
- 0xa5, 0xfe, 0x22, 0x2c, 0x56, 0x5d, 0xea, 0x91, 0x0e, 0xe9, 0xf7, 0x75, 0xcb, 0x90, 0xa7, 0x23,
- 0xa2, 0x49, 0xfd, 0xa7, 0x3b, 0x5d, 0xea, 0x76, 0x4a, 0x66, 0xcf, 0x74, 0xf2, 0xe8, 0x0e, 0xce,
- 0xb4, 0x18, 0x5c, 0xc6, 0x12, 0x4f, 0x49, 0x05, 0x41, 0xda, 0x32, 0x1d, 0xb4, 0x0e, 0xd3, 0xd8,
- 0x7a, 0x25, 0x4b, 0x9f, 0x94, 0xe3, 0x13, 0xb9, 0xc4, 0xab, 0x8c, 0x0f, 0xdd, 0x82, 0xd9, 0x3e,
- 0x0d, 0x0b, 0xb9, 0x97, 0x5b, 0xc9, 0x38, 0x45, 0x50, 0x05, 0x1b, 0xda, 0x80, 0x39, 0x83, 0xcd,
- 0x93, 0xdc, 0xb0, 0xd5, 0x52, 0x40, 0x38, 0xc6, 0xa0, 0x4a, 0x46, 0xb4, 0xed, 0x17, 0x76, 0xa5,
- 0xac, 0x8a, 0x2c, 0x36, 0x15, 0xa9, 0xd5, 0xdd, 0x5e, 0xb4, 0xba, 0x03, 0xa6, 0x6b, 0x63, 0xb4,
- 0xae, 0xe1, 0xc8, 0xd9, 0x59, 0x28, 0xf6, 0x48, 0x97, 0x87, 0x51, 0x99, 0x1f, 0xbc, 0xf5, 0x48,
- 0x97, 0x45, 0xd1, 0x32, 0x2d, 0x74, 0x0d, 0xd3, 0xaa, 0xcd, 0xb3, 0x14, 0xc6, 0x1b, 0xf4, 0xe3,
- 0x63, 0x0f, 0x1a, 0xb1, 0x3a, 0xb8, 0x56, 0x61, 0xaf, 0x4a, 0x8c, 0xb2, 0x63, 0x75, 0x58, 0xe9,
- 0xe4, 0x79, 0xc7, 0xb5, 0x05, 0x46, 0xa7, 0x8f, 0x74, 0x0f, 0xc3, 0xb7, 0xdb, 0x8b, 0x59, 0x7b,
- 0x98, 0xb4, 0x64, 0x28, 0x77, 0xdb, 0x8f, 0x61, 0xee, 0x35, 0x4f, 0x04, 0xb5, 0x2a, 0x93, 0xbf,
- 0x36, 0x3a, 0xa5, 0x08, 0x0d, 0x52, 0xf0, 0xdb, 0x2c, 0x63, 0x7f, 0x99, 0x83, 0x33, 0x9b, 0xac,
- 0xc4, 0x0f, 0xe5, 0xb1, 0x49, 0x50, 0xa9, 0xbb, 0x3e, 0x0e, 0x98, 0x09, 0x21, 0xc5, 0xc7, 0x2d,
- 0x04, 0x50, 0x13, 0x16, 0xa4, 0x72, 0xa1, 0xa2, 0x30, 0x36, 0x94, 0x58, 0x71, 0xc3, 0x4d, 0xe5,
- 0x63, 0x58, 0x49, 0x8c, 0x42, 0x94, 0xe3, 0x6b, 0x30, 0x1f, 0xe4, 0x2b, 0x7f, 0x10, 0x65, 0x9f,
- 0xd6, 0x34, 0x94, 0x7b, 0x70, 0xba, 0xed, 0xe9, 0x8e, 0x97, 0x70, 0xc1, 0x18, 0xb2, 0x0c, 0x4d,
- 0x8c, 0xca, 0x0a, 0xc0, 0xaf, 0x0d, 0xcb, 0x6d, 0x8f, 0xd8, 0x27, 0x50, 0x4a, 0xb3, 0x0e, 0x1d,
- 0x3f, 0x19, 0xc8, 0xf5, 0x41, 0x36, 0x95, 0x15, 0x8e, 0x7d, 0x26, 0x7b, 0xbb, 0x0f, 0x67, 0x38,
- 0xf4, 0x78, 0x92, 0x41, 0x9c, 0x95, 0xc0, 0x67, 0x52, 0xef, 0x0b, 0x38, 0x15, 0xac, 0xbd, 0x01,
- 0xac, 0x70, 0x27, 0x0a, 0x2b, 0x5c, 0x1c, 0x32, 0xeb, 0x11, 0x54, 0xe1, 0x2f, 0xf2, 0xa1, 0xbc,
- 0x9e, 0x01, 0x2a, 0xdc, 0x8f, 0x82, 0x0a, 0x97, 0x47, 0xe9, 0x8e, 0x60, 0x0a, 0xc9, 0xa8, 0x2d,
- 0xa4, 0x44, 0xed, 0x17, 0x09, 0xe4, 0x61, 0x3a, 0x0b, 0xba, 0x89, 0x59, 0xfb, 0x8d, 0x00, 0x0f,
- 0x2a, 0x07, 0x1e, 0xfc, 0xae, 0x7d, 0xa4, 0xf8, 0x6e, 0x0c, 0x78, 0x58, 0x1b, 0x69, 0xaf, 0x8f,
- 0x3b, 0xfc, 0xcd, 0x34, 0x94, 0xfc, 0x77, 0x09, 0x9f, 0x27, 0xdd, 0x96, 0x4f, 0x71, 0x5b, 0x78,
- 0x05, 0x2e, 0x7c, 0xad, 0x15, 0x78, 0x7a, 0xec, 0x15, 0xf8, 0x1c, 0x94, 0xd8, 0x83, 0xe6, 0xe0,
- 0x43, 0xb1, 0xa2, 0x16, 0x19, 0x41, 0xc5, 0x87, 0x41, 0x18, 0xce, 0x4e, 0x14, 0x86, 0x31, 0xa8,
- 0x63, 0x2e, 0x0e, 0x75, 0x3c, 0xf4, 0x57, 0x44, 0xbe, 0x88, 0x5e, 0x1d, 0xa2, 0x37, 0x75, 0x2d,
- 0x6c, 0x45, 0xd7, 0x42, 0xbe, 0xae, 0xbe, 0x3f, 0x4c, 0xcb, 0x77, 0x16, 0xe8, 0xd8, 0xe7, 0x40,
- 0x47, 0x38, 0x16, 0x45, 0x66, 0xbd, 0x0f, 0xe0, 0x27, 0x11, 0x89, 0x76, 0x9c, 0x1b, 0x32, 0x46,
- 0x35, 0xc4, 0x4e, 0xd5, 0x46, 0xa6, 0x26, 0x38, 0x0d, 0x19, 0x2f, 0x3f, 0x66, 0x1c, 0x85, 0xfc,
- 0xef, 0x4c, 0x28, 0xbf, 0x64, 0x1c, 0x1f, 0x3c, 0x4c, 0x40, 0x6c, 0x13, 0x46, 0xf1, 0x9d, 0x28,
- 0xc2, 0x76, 0xc2, 0xa8, 0x4b, 0x00, 0x6c, 0xac, 0x72, 0xd1, 0x1d, 0xf1, 0x9a, 0x03, 0x20, 0x25,
- 0x41, 0x69, 0xb0, 0x9d, 0xc1, 0xa1, 0x69, 0x99, 0xee, 0x11, 0x7f, 0x3f, 0xcb, 0x77, 0x06, 0x92,
- 0xd4, 0x60, 0x17, 0x68, 0xf0, 0x1b, 0xd3, 0xd3, 0x3a, 0xc4, 0xc0, 0x2c, 0xa6, 0x67, 0xd4, 0x22,
- 0x25, 0x6c, 0x12, 0x03, 0x07, 0x5f, 0x5e, 0xf1, 0x64, 0x5f, 0x5e, 0x29, 0xf6, 0xe5, 0x9d, 0x81,
- 0x59, 0x07, 0xeb, 0x2e, 0xb1, 0xc4, 0x56, 0x53, 0xb4, 0xe8, 0xd4, 0xf4, 0xb1, 0xeb, 0xd2, 0x9e,
- 0x44, 0xb9, 0x26, 0x9a, 0xa1, 0x32, 0x73, 0x7e, 0x64, 0x99, 0x39, 0xe4, 0x58, 0x22, 0x56, 0x66,
- 0x56, 0x46, 0x96, 0x99, 0xe3, 0x9c, 0x4a, 0x84, 0x0a, 0xed, 0x85, 0xf1, 0x0a, 0xed, 0x70, 0x5d,
- 0xba, 0x18, 0xa9, 0x4b, 0xbf, 0xcd, 0x8f, 0xf5, 0xd7, 0x39, 0x58, 0x49, 0x7c, 0x56, 0xe2, 0x73,
- 0xbd, 0x1b, 0x3b, 0xe0, 0x58, 0x1b, 0xe9, 0x33, 0xff, 0x7c, 0xe3, 0x69, 0xe4, 0x7c, 0xe3, 0xc3,
- 0xd1, 0x82, 0x6f, 0xfd, 0x78, 0xe3, 0x8f, 0x72, 0xf0, 0xce, 0xbe, 0x6d, 0xc4, 0x2a, 0x3c, 0xb1,
- 0x31, 0x1f, 0x3f, 0x71, 0x3c, 0x94, 0xb5, 0x7e, 0x7e, 0x52, 0xf4, 0x82, 0xcb, 0x29, 0x0a, 0x5c,
- 0xcc, 0x36, 0x43, 0x94, 0x4c, 0x3f, 0x81, 0xc5, 0xed, 0x37, 0xb8, 0xd3, 0x3e, 0xb6, 0x3a, 0x13,
- 0x98, 0x56, 0x85, 0x42, 0xa7, 0x6f, 0x08, 0xc4, 0x8f, 0x3e, 0x86, 0xab, 0xc0, 0x42, 0xb4, 0x0a,
- 0xd4, 0xa0, 0x1a, 0xf4, 0x20, 0xa6, 0xf7, 0x0c, 0x9d, 0x5e, 0x83, 0x32, 0x53, 0xe5, 0xf3, 0xaa,
- 0x68, 0x09, 0x3a, 0x76, 0x1c, 0x36, 0x66, 0x4e, 0xc7, 0x8e, 0x13, 0xcd, 0x16, 0x85, 0x68, 0xb6,
- 0x50, 0xfe, 0x2c, 0x07, 0x65, 0xda, 0xc3, 0xd7, 0xb2, 0x5f, 0x6c, 0xb5, 0x0a, 0xc1, 0x56, 0xcb,
- 0xdf, 0xb1, 0x4d, 0x87, 0x77, 0x6c, 0x81, 0xe5, 0x33, 0x8c, 0x9c, 0xb4, 0x7c, 0xd6, 0xa7, 0x63,
- 0xc7, 0x51, 0x2e, 0xc2, 0x3c, 0xb7, 0x4d, 0x8c, 0xbc, 0x0a, 0x85, 0x81, 0xd3, 0x93, 0x71, 0x34,
- 0x70, 0x7a, 0xca, 0x1f, 0xe7, 0xa0, 0xd2, 0xf0, 0x3c, 0xbd, 0x73, 0x34, 0xc1, 0x00, 0x7c, 0xe3,
- 0xf2, 0x61, 0xe3, 0x92, 0x83, 0x08, 0xcc, 0x9d, 0xce, 0x30, 0x77, 0x26, 0x62, 0xae, 0x02, 0x0b,
- 0xd2, 0x96, 0x4c, 0x83, 0x5b, 0x80, 0x76, 0x89, 0xe3, 0x3d, 0x21, 0xce, 0x6b, 0xdd, 0x31, 0x26,
- 0xdb, 0x81, 0x21, 0x98, 0x16, 0x97, 0x2a, 0x0b, 0xd7, 0x66, 0x54, 0xf6, 0xac, 0x5c, 0x85, 0x53,
- 0x11, 0x7d, 0x99, 0x1d, 0x3f, 0x82, 0x32, 0xcb, 0xfb, 0xa2, 0x14, 0xbf, 0x1d, 0x3e, 0x7a, 0x18,
- 0x6b, 0x95, 0x50, 0x7e, 0x17, 0x96, 0x68, 0x7d, 0xc0, 0xe8, 0xfe, 0xa7, 0xf8, 0xfd, 0x58, 0x9d,
- 0x7a, 0x21, 0x43, 0x51, 0xac, 0x46, 0xfd, 0xdb, 0x1c, 0xcc, 0x30, 0x7a, 0x62, 0xcd, 0x3e, 0x07,
- 0x25, 0x07, 0xdb, 0x44, 0xf3, 0xf4, 0xae, 0x7f, 0x85, 0x95, 0x12, 0xf6, 0xf4, 0xae, 0xcb, 0x6e,
- 0xe0, 0xd2, 0x97, 0x86, 0xd9, 0xc5, 0xae, 0x27, 0xef, 0xb1, 0x96, 0x29, 0x6d, 0x8b, 0x93, 0xa8,
- 0x93, 0x5c, 0xf3, 0xf7, 0x79, 0xdd, 0x39, 0xad, 0xb2, 0x67, 0xb4, 0xce, 0x6f, 0x65, 0x8d, 0x03,
- 0x0f, 0xb3, 0x3b, 0x5b, 0x75, 0x28, 0xc6, 0x10, 0x61, 0xbf, 0xad, 0x6c, 0x03, 0x0a, 0x7b, 0x41,
- 0xf8, 0xfb, 0x16, 0xcc, 0x32, 0x27, 0xc9, 0xea, 0x68, 0x25, 0xc3, 0x0d, 0xaa, 0x60, 0x53, 0x74,
- 0x40, 0xdc, 0xc1, 0x91, 0x8a, 0x68, 0xf2, 0x59, 0x19, 0x52, 0x21, 0xfd, 0x43, 0x0e, 0x4e, 0x45,
- 0xfa, 0x10, 0xb6, 0xde, 0x8c, 0x76, 0x92, 0x69, 0xaa, 0xe8, 0x60, 0x33, 0xb2, 0x24, 0xdc, 0xca,
- 0x32, 0xe9, 0xb7, 0xb4, 0x1c, 0xfc, 0x63, 0x0e, 0xa0, 0x31, 0xf0, 0x8e, 0x04, 0x32, 0x18, 0x9e,
- 0x99, 0x5c, 0x74, 0x66, 0xe8, 0x3b, 0x5b, 0x77, 0xdd, 0xd7, 0xc4, 0x91, 0x7b, 0x1a, 0xbf, 0xcd,
- 0x30, 0xbc, 0x81, 0x77, 0x24, 0x8f, 0x75, 0xe8, 0x33, 0xba, 0x0c, 0x0b, 0xfc, 0xda, 0xb4, 0xa6,
- 0x1b, 0x86, 0x83, 0x5d, 0x57, 0x9c, 0xef, 0x54, 0x38, 0xb5, 0xc1, 0x89, 0x94, 0xcd, 0x34, 0xb0,
- 0xe5, 0x99, 0xde, 0xb1, 0xe6, 0x91, 0x97, 0xd8, 0x12, 0x7b, 0x93, 0x8a, 0xa4, 0xee, 0x51, 0x22,
- 0x65, 0x73, 0x70, 0xd7, 0x74, 0x3d, 0x47, 0xb2, 0xc9, 0xb3, 0x04, 0x41, 0x65, 0x6c, 0x74, 0x52,
- 0xaa, 0xbb, 0x83, 0x5e, 0x8f, 0xbb, 0xf8, 0xe4, 0xd3, 0xfe, 0x81, 0x18, 0x50, 0x3e, 0x2b, 0xa6,
- 0x03, 0xa7, 0x89, 0xe1, 0xbe, 0x45, 0x10, 0xe6, 0x03, 0x58, 0x0a, 0x8d, 0x41, 0x84, 0x55, 0xa4,
- 0x88, 0xcc, 0x45, 0x8b, 0x48, 0xe5, 0x29, 0x20, 0x8e, 0x3b, 0x7c, 0xcd, 0x71, 0x2b, 0xa7, 0xe1,
- 0x54, 0x44, 0x91, 0x58, 0x89, 0x6f, 0x40, 0x45, 0xdc, 0xb1, 0x11, 0x81, 0x72, 0x16, 0x8a, 0x34,
- 0xa3, 0x76, 0x4c, 0x43, 0x9e, 0xf9, 0xcd, 0xd9, 0xc4, 0xd8, 0x34, 0x0d, 0x47, 0xf9, 0x14, 0x2a,
- 0x2a, 0xef, 0x47, 0xf0, 0x3e, 0x81, 0x05, 0x71, 0x23, 0x47, 0x8b, 0xdc, 0x74, 0x4b, 0xbb, 0x09,
- 0x1d, 0xee, 0x44, 0xad, 0x58, 0xe1, 0xa6, 0x62, 0x40, 0x9d, 0x97, 0x0c, 0x11, 0xf5, 0x72, 0xb0,
- 0x4f, 0x40, 0xfe, 0x06, 0x60, 0x64, 0x2f, 0x51, 0xf9, 0x8a, 0x13, 0x6e, 0x2a, 0x17, 0xe0, 0x5c,
- 0x6a, 0x2f, 0xc2, 0x13, 0x36, 0x54, 0x83, 0x17, 0x86, 0x29, 0x0f, 0x3f, 0xd9, 0xa1, 0x66, 0x2e,
- 0x74, 0xa8, 0x79, 0xc6, 0x2f, 0x12, 0xf3, 0x72, 0x11, 0x63, 0x15, 0x60, 0x50, 0xee, 0x17, 0xb2,
- 0xca, 0xfd, 0xe9, 0x48, 0xb9, 0xaf, 0xb4, 0x7d, 0x7f, 0x8a, 0x6d, 0xd8, 0x63, 0xb6, 0x5d, 0xe4,
- 0x7d, 0xcb, 0x84, 0xa8, 0x0c, 0x1b, 0x25, 0x67, 0x55, 0x43, 0x52, 0xca, 0x75, 0xa8, 0x44, 0x53,
- 0x63, 0x28, 0xcf, 0xe5, 0x12, 0x79, 0x6e, 0x21, 0x96, 0xe2, 0x3e, 0x8a, 0x55, 0xc0, 0xd9, 0x3e,
- 0x8e, 0xd5, 0xbf, 0x0f, 0x22, 0xc9, 0xee, 0x46, 0xca, 0x79, 0xe4, 0x6f, 0x29, 0xcf, 0x2d, 0x8b,
- 0xf5, 0xe0, 0x89, 0x4b, 0xe5, 0xc5, 0xa0, 0x95, 0x4b, 0x50, 0xde, 0xcf, 0xba, 0x66, 0x3f, 0x2d,
- 0xcf, 0xfe, 0xef, 0xc0, 0xf2, 0x13, 0xb3, 0x87, 0xdd, 0x63, 0xd7, 0xc3, 0xfd, 0x26, 0x4b, 0x4a,
- 0x87, 0x26, 0x76, 0xd0, 0x2a, 0x00, 0xdb, 0xc2, 0xd8, 0xc4, 0xf4, 0x6f, 0x6f, 0x87, 0x28, 0xca,
- 0x7f, 0xe5, 0x60, 0x31, 0x10, 0xdc, 0x67, 0x5b, 0xb7, 0xf3, 0x50, 0xa2, 0xe3, 0x75, 0x3d, 0xbd,
- 0x6f, 0xcb, 0xf3, 0x2c, 0x9f, 0x80, 0xee, 0xc3, 0xcc, 0xa1, 0x2b, 0x21, 0xa3, 0x54, 0x00, 0x3d,
- 0xcd, 0x10, 0x75, 0xfa, 0xd0, 0x6d, 0x1a, 0xe8, 0x63, 0x80, 0x81, 0x8b, 0x0d, 0x71, 0x86, 0x55,
- 0xc8, 0xaa, 0x16, 0xf6, 0xc3, 0x67, 0xb5, 0x54, 0x80, 0x5f, 0x1b, 0x78, 0x00, 0x65, 0xd3, 0x22,
- 0x06, 0x66, 0x67, 0xb5, 0x86, 0x40, 0x95, 0x46, 0x88, 0x03, 0x97, 0xd8, 0x77, 0xb1, 0xa1, 0x60,
- 0xb1, 0x16, 0x4a, 0xff, 0x8a, 0x40, 0x69, 0xc1, 0x12, 0x4f, 0x5a, 0x87, 0xbe, 0xe1, 0x32, 0x62,
- 0xd7, 0x86, 0x8d, 0x8e, 0x79, 0x4b, 0xad, 0x9a, 0xa2, 0xb4, 0x91, 0xa2, 0xca, 0x3d, 0x38, 0x1d,
- 0xd9, 0x21, 0x4d, 0xb0, 0x65, 0x51, 0x76, 0x63, 0x40, 0x49, 0x10, 0xce, 0x02, 0x86, 0x90, 0xd1,
- 0x3c, 0x0a, 0x86, 0x70, 0x39, 0x0c, 0xe1, 0x2a, 0x5f, 0xc0, 0xd9, 0x08, 0xa2, 0x13, 0xb1, 0xe8,
- 0x41, 0xac, 0x72, 0xbb, 0x32, 0x4a, 0x6b, 0xac, 0x84, 0xfb, 0x9f, 0x1c, 0x2c, 0xa7, 0x31, 0x9c,
- 0x10, 0x71, 0xfc, 0x49, 0xc6, 0x15, 0xb1, 0xbb, 0xe3, 0x99, 0xf5, 0x8d, 0xa0, 0xb5, 0x7b, 0x50,
- 0x4f, 0xf3, 0x67, 0x72, 0x96, 0x0a, 0x93, 0xcc, 0xd2, 0xcf, 0x0b, 0x21, 0xe4, 0xbd, 0xe1, 0x79,
- 0x8e, 0x79, 0x30, 0xa0, 0x21, 0xff, 0xd6, 0xd1, 0xac, 0xa6, 0x8f, 0xcb, 0x70, 0xd7, 0xde, 0x1e,
- 0x22, 0x1e, 0xd8, 0x91, 0x8a, 0xcd, 0x7c, 0x16, 0xc5, 0x66, 0x38, 0xa6, 0x7e, 0x67, 0x3c, 0x7d,
- 0xdf, 0x59, 0x00, 0xf4, 0xe7, 0x79, 0x58, 0x88, 0x4e, 0x11, 0xda, 0x06, 0xd0, 0x7d, 0xcb, 0xc5,
- 0x87, 0x72, 0x79, 0xac, 0x61, 0xaa, 0x21, 0x41, 0xf4, 0x3e, 0x14, 0x3a, 0xf6, 0x40, 0xcc, 0x5a,
- 0xca, 0x61, 0xf0, 0xa6, 0x3d, 0xe0, 0x19, 0x85, 0xb2, 0xd1, 0x3d, 0x15, 0x3f, 0xdb, 0xcf, 0xce,
- 0x92, 0x2f, 0xd8, 0x7b, 0x2e, 0x23, 0x98, 0xd1, 0x33, 0x58, 0x78, 0xed, 0x98, 0x9e, 0x7e, 0xd0,
- 0xc3, 0x5a, 0x4f, 0x3f, 0xc6, 0x8e, 0xc8, 0x92, 0x63, 0x24, 0xb2, 0x8a, 0x14, 0x7c, 0x4e, 0xe5,
- 0x94, 0x3f, 0x80, 0xa2, 0xb4, 0x68, 0xc4, 0x8a, 0xb0, 0x07, 0x2b, 0x03, 0xca, 0xa6, 0xb1, 0xeb,
- 0x5c, 0x96, 0x6e, 0x11, 0xcd, 0xc5, 0x74, 0x19, 0x97, 0x17, 0xcd, 0x47, 0xa4, 0xe8, 0x65, 0x26,
- 0xbd, 0x49, 0x1c, 0xdc, 0xd2, 0x2d, 0xd2, 0xe6, 0xa2, 0xca, 0x2b, 0x28, 0x87, 0x06, 0x38, 0xc2,
- 0x84, 0x26, 0x2c, 0xc9, 0xa3, 0x78, 0x17, 0x7b, 0x62, 0x79, 0x19, 0xab, 0xf3, 0x45, 0x21, 0xd7,
- 0xc6, 0x1e, 0xbf, 0x3e, 0xf1, 0x00, 0xce, 0xaa, 0x98, 0xd8, 0xd8, 0xf2, 0xe7, 0xf3, 0x39, 0xe9,
- 0x4e, 0x90, 0xc1, 0xcf, 0x43, 0x3d, 0x4d, 0x9e, 0xe7, 0x87, 0x1b, 0xe7, 0xa1, 0x28, 0x7f, 0x33,
- 0x89, 0xe6, 0xa0, 0xb0, 0xb7, 0xb9, 0x5b, 0x9d, 0xa2, 0x0f, 0xfb, 0x5b, 0xbb, 0xd5, 0xdc, 0x8d,
- 0x3e, 0x54, 0xe3, 0x3f, 0x13, 0x44, 0x2b, 0x70, 0x6a, 0x57, 0xdd, 0xd9, 0x6d, 0x3c, 0x6d, 0xec,
- 0x35, 0x77, 0x5a, 0xda, 0xae, 0xda, 0xfc, 0xa4, 0xb1, 0xb7, 0x5d, 0x9d, 0x42, 0x6b, 0x70, 0x21,
- 0xfc, 0xe2, 0xd9, 0x4e, 0x7b, 0x4f, 0xdb, 0xdb, 0xd1, 0x36, 0x77, 0x5a, 0x7b, 0x8d, 0x66, 0x6b,
- 0x5b, 0xad, 0xe6, 0xd0, 0x05, 0x38, 0x1b, 0x66, 0x79, 0xdc, 0xdc, 0x6a, 0xaa, 0xdb, 0x9b, 0xf4,
- 0xb9, 0xf1, 0xbc, 0x9a, 0xbf, 0x71, 0x1b, 0x2a, 0x91, 0x5f, 0xf5, 0x51, 0x43, 0x76, 0x77, 0xb6,
- 0xaa, 0x53, 0xa8, 0x02, 0xa5, 0xb0, 0x9e, 0x22, 0x4c, 0xb7, 0x76, 0xb6, 0xb6, 0xab, 0xf9, 0x1b,
- 0xf7, 0x60, 0x31, 0x76, 0x9f, 0x14, 0x2d, 0x41, 0xa5, 0xdd, 0x68, 0x6d, 0x3d, 0xde, 0xf9, 0x4c,
- 0x53, 0xb7, 0x1b, 0x5b, 0x9f, 0x57, 0xa7, 0xd0, 0x32, 0x54, 0x25, 0xa9, 0xb5, 0xb3, 0xc7, 0xa9,
- 0xb9, 0x1b, 0x2f, 0x63, 0x5f, 0x16, 0x46, 0xa7, 0x61, 0xc9, 0xef, 0x46, 0xdb, 0x54, 0xb7, 0x1b,
- 0x7b, 0xdb, 0xb4, 0xf7, 0x08, 0x59, 0xdd, 0x6f, 0xb5, 0x9a, 0xad, 0xa7, 0xd5, 0x1c, 0xd5, 0x1a,
- 0x90, 0xb7, 0x3f, 0x6b, 0x52, 0xe6, 0x7c, 0x94, 0x79, 0xbf, 0xf5, 0xc3, 0xd6, 0xce, 0xa7, 0xad,
- 0x6a, 0x61, 0xe3, 0xef, 0x97, 0x60, 0x41, 0x96, 0x77, 0xd8, 0x61, 0x77, 0x59, 0x76, 0x61, 0x4e,
- 0xfe, 0xf2, 0x36, 0x25, 0x2f, 0x47, 0x7f, 0x2f, 0x5c, 0x5f, 0x1b, 0xc2, 0x21, 0xaa, 0xec, 0x29,
- 0x74, 0xc0, 0xaa, 0xde, 0xd0, 0xfd, 0xde, 0x2b, 0xa9, 0x35, 0x66, 0xe2, 0x4a, 0x71, 0xfd, 0xea,
- 0x48, 0x3e, 0xbf, 0x0f, 0x4c, 0x0b, 0xdb, 0xf0, 0x0f, 0x58, 0xd0, 0xd5, 0xb4, 0x8a, 0x34, 0xe5,
- 0x17, 0x32, 0xf5, 0x6b, 0xa3, 0x19, 0xfd, 0x6e, 0x5e, 0x42, 0x35, 0xfe, 0x63, 0x16, 0x94, 0x02,
- 0x98, 0x66, 0xfc, 0x62, 0xa6, 0x7e, 0x63, 0x1c, 0xd6, 0x70, 0x67, 0x89, 0x9f, 0x7d, 0x5c, 0x1f,
- 0xe7, 0x1e, 0x7d, 0x66, 0x67, 0x59, 0x57, 0xee, 0xb9, 0x03, 0xa3, 0x77, 0x77, 0x51, 0xea, 0x6f,
- 0x2c, 0x52, 0x6e, 0x7e, 0xa7, 0x39, 0x30, 0xfd, 0x1a, 0xb0, 0x32, 0x85, 0x8e, 0x60, 0x31, 0x76,
- 0x29, 0x01, 0xa5, 0x88, 0xa7, 0xdf, 0xbe, 0xa8, 0x5f, 0x1f, 0x83, 0x33, 0x1a, 0x11, 0xe1, 0x4b,
- 0x08, 0xe9, 0x11, 0x91, 0x72, 0xc5, 0x21, 0x3d, 0x22, 0x52, 0xef, 0x33, 0xb0, 0xe0, 0x8e, 0x5c,
- 0x3e, 0x48, 0x0b, 0xee, 0xb4, 0x2b, 0x0f, 0xf5, 0xab, 0x23, 0xf9, 0xc2, 0x4e, 0x8b, 0x5d, 0x45,
- 0x48, 0x73, 0x5a, 0xfa, 0x55, 0x87, 0xfa, 0xf5, 0x31, 0x38, 0xe3, 0x51, 0x10, 0x1c, 0x6c, 0x66,
- 0x45, 0x41, 0xe2, 0x18, 0x3e, 0x2b, 0x0a, 0x92, 0x67, 0xa4, 0x22, 0x0a, 0x62, 0x07, 0x92, 0xd7,
- 0xc6, 0x38, 0x40, 0xc9, 0x8e, 0x82, 0xf4, 0xa3, 0x16, 0x65, 0x0a, 0xfd, 0x61, 0x0e, 0x6a, 0x59,
- 0x87, 0x13, 0x28, 0xa5, 0xaa, 0x1b, 0x71, 0x9e, 0x52, 0xdf, 0x98, 0x44, 0xc4, 0xb7, 0xe2, 0x4b,
- 0x40, 0xc9, 0xd5, 0x0e, 0xbd, 0x97, 0x36, 0x33, 0x19, 0x6b, 0x6a, 0xfd, 0xfd, 0xf1, 0x98, 0xfd,
- 0x2e, 0xdb, 0x50, 0x94, 0xc7, 0x21, 0x28, 0x25, 0x4b, 0xc7, 0x0e, 0x63, 0xea, 0xca, 0x30, 0x16,
- 0x5f, 0xe9, 0x53, 0x98, 0xa6, 0x54, 0x74, 0x21, 0x9d, 0x5b, 0x2a, 0x5b, 0xcd, 0x7a, 0xed, 0x2b,
- 0x7a, 0x01, 0xb3, 0x1c, 0xff, 0x47, 0x29, 0x78, 0x43, 0xe4, 0x94, 0xa2, 0x7e, 0x31, 0x9b, 0xc1,
- 0x57, 0xf7, 0x63, 0xfe, 0x4f, 0x19, 0x04, 0xb4, 0x8f, 0xde, 0x4d, 0xff, 0x95, 0x6c, 0xf4, 0x24,
- 0xa1, 0x7e, 0x79, 0x04, 0x57, 0xf8, 0xa3, 0x88, 0xd5, 0xba, 0x57, 0x47, 0x6e, 0x58, 0xb2, 0x3f,
- 0x8a, 0xf4, 0x2d, 0x11, 0x0f, 0x92, 0xe4, 0x96, 0x29, 0x2d, 0x48, 0x32, 0x37, 0xaa, 0x69, 0x41,
- 0x92, 0xbd, 0x0b, 0x53, 0xa6, 0x90, 0x07, 0xa7, 0x52, 0x00, 0x32, 0xf4, 0x7e, 0x56, 0x90, 0xa7,
- 0xa1, 0x75, 0xf5, 0x9b, 0x63, 0x72, 0x87, 0x27, 0x5f, 0x7c, 0xf4, 0xef, 0x64, 0xa3, 0x46, 0x99,
- 0x93, 0x1f, 0xff, 0xc4, 0x37, 0xfe, 0xad, 0x00, 0xf3, 0x1c, 0xfc, 0x14, 0x15, 0xcc, 0xe7, 0x00,
- 0xc1, 0xb9, 0x03, 0xba, 0x94, 0xee, 0x93, 0xc8, 0xd9, 0x4c, 0xfd, 0xdd, 0xe1, 0x4c, 0xe1, 0x40,
- 0x0b, 0x61, 0xf8, 0x69, 0x81, 0x96, 0x3c, 0xaa, 0x48, 0x0b, 0xb4, 0x94, 0x83, 0x00, 0x65, 0x0a,
- 0x7d, 0x02, 0x25, 0x1f, 0x2c, 0x46, 0x69, 0x60, 0x73, 0x0c, 0x0d, 0xaf, 0x5f, 0x1a, 0xca, 0x13,
- 0xb6, 0x3a, 0x84, 0x04, 0xa7, 0x59, 0x9d, 0x44, 0x9c, 0xd3, 0xac, 0x4e, 0x83, 0x93, 0x03, 0x9f,
- 0x70, 0xbc, 0x28, 0xd3, 0x27, 0x11, 0xb8, 0x2e, 0xd3, 0x27, 0x51, 0xd0, 0x49, 0x99, 0x7a, 0x7c,
- 0xe5, 0x57, 0xbf, 0x59, 0xcd, 0xfd, 0xf3, 0x6f, 0x56, 0xa7, 0x7e, 0xf6, 0xd5, 0x6a, 0xee, 0x57,
- 0x5f, 0xad, 0xe6, 0xfe, 0xe9, 0xab, 0xd5, 0xdc, 0xbf, 0x7f, 0xb5, 0x9a, 0xfb, 0xd3, 0xff, 0x58,
- 0x9d, 0xfa, 0x51, 0x51, 0x4a, 0x1f, 0xcc, 0xb2, 0x7f, 0xad, 0xf2, 0xe1, 0xff, 0x05, 0x00, 0x00,
- 0xff, 0xff, 0x36, 0xf8, 0x35, 0x67, 0x20, 0x47, 0x00, 0x00,
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.proto b/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.proto
deleted file mode 100644
index 257cfbc2e..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/api.proto
+++ /dev/null
@@ -1,1217 +0,0 @@
-// To regenerate api.pb.go run hack/update-generated-runtime.sh
-syntax = 'proto3';
-
-package runtime.v1alpha2;
-option go_package = "v1alpha2";
-
-import "github.com/gogo/protobuf/gogoproto/gogo.proto";
-
-option (gogoproto.goproto_stringer_all) = false;
-option (gogoproto.stringer_all) = true;
-option (gogoproto.goproto_getters_all) = true;
-option (gogoproto.marshaler_all) = true;
-option (gogoproto.sizer_all) = true;
-option (gogoproto.unmarshaler_all) = true;
-option (gogoproto.goproto_unrecognized_all) = false;
-
-// Runtime service defines the public APIs for remote container runtimes
-service RuntimeService {
- // Version returns the runtime name, runtime version, and runtime API version.
- rpc Version(VersionRequest) returns (VersionResponse) {}
-
- // RunPodSandbox creates and starts a pod-level sandbox. Runtimes must ensure
- // the sandbox is in the ready state on success.
- rpc RunPodSandbox(RunPodSandboxRequest) returns (RunPodSandboxResponse) {}
- // StopPodSandbox stops any running process that is part of the sandbox and
- // reclaims network resources (e.g., IP addresses) allocated to the sandbox.
- // If there are any running containers in the sandbox, they must be forcibly
- // terminated.
- // This call is idempotent, and must not return an error if all relevant
- // resources have already been reclaimed. kubelet will call StopPodSandbox
- // at least once before calling RemovePodSandbox. It will also attempt to
- // reclaim resources eagerly, as soon as a sandbox is not needed. Hence,
- // multiple StopPodSandbox calls are expected.
- rpc StopPodSandbox(StopPodSandboxRequest) returns (StopPodSandboxResponse) {}
- // RemovePodSandbox removes the sandbox. If there are any running containers
- // in the sandbox, they must be forcibly terminated and removed.
- // This call is idempotent, and must not return an error if the sandbox has
- // already been removed.
- rpc RemovePodSandbox(RemovePodSandboxRequest) returns (RemovePodSandboxResponse) {}
- // PodSandboxStatus returns the status of the PodSandbox. If the PodSandbox is not
- // present, returns an error.
- rpc PodSandboxStatus(PodSandboxStatusRequest) returns (PodSandboxStatusResponse) {}
- // ListPodSandbox returns a list of PodSandboxes.
- rpc ListPodSandbox(ListPodSandboxRequest) returns (ListPodSandboxResponse) {}
-
- // CreateContainer creates a new container in specified PodSandbox
- rpc CreateContainer(CreateContainerRequest) returns (CreateContainerResponse) {}
- // StartContainer starts the container.
- rpc StartContainer(StartContainerRequest) returns (StartContainerResponse) {}
- // StopContainer stops a running container with a grace period (i.e., timeout).
- // This call is idempotent, and must not return an error if the container has
- // already been stopped.
- // TODO: what must the runtime do after the grace period is reached?
- rpc StopContainer(StopContainerRequest) returns (StopContainerResponse) {}
- // RemoveContainer removes the container. If the container is running, the
- // container must be forcibly removed.
- // This call is idempotent, and must not return an error if the container has
- // already been removed.
- rpc RemoveContainer(RemoveContainerRequest) returns (RemoveContainerResponse) {}
- // ListContainers lists all containers by filters.
- rpc ListContainers(ListContainersRequest) returns (ListContainersResponse) {}
- // ContainerStatus returns status of the container. If the container is not
- // present, returns an error.
- rpc ContainerStatus(ContainerStatusRequest) returns (ContainerStatusResponse) {}
- // UpdateContainerResources updates ContainerConfig of the container.
- rpc UpdateContainerResources(UpdateContainerResourcesRequest) returns (UpdateContainerResourcesResponse) {}
- // ReopenContainerLog asks runtime to reopen the stdout/stderr log file
- // for the container. This is often called after the log file has been
- // rotated. If the container is not running, container runtime can choose
- // to either create a new log file and return nil, or return an error.
- // Once it returns error, new container log file MUST NOT be created.
- rpc ReopenContainerLog(ReopenContainerLogRequest) returns (ReopenContainerLogResponse) {}
-
- // ExecSync runs a command in a container synchronously.
- rpc ExecSync(ExecSyncRequest) returns (ExecSyncResponse) {}
- // Exec prepares a streaming endpoint to execute a command in the container.
- rpc Exec(ExecRequest) returns (ExecResponse) {}
- // Attach prepares a streaming endpoint to attach to a running container.
- rpc Attach(AttachRequest) returns (AttachResponse) {}
- // PortForward prepares a streaming endpoint to forward ports from a PodSandbox.
- rpc PortForward(PortForwardRequest) returns (PortForwardResponse) {}
-
- // ContainerStats returns stats of the container. If the container does not
- // exist, the call returns an error.
- rpc ContainerStats(ContainerStatsRequest) returns (ContainerStatsResponse) {}
- // ListContainerStats returns stats of all running containers.
- rpc ListContainerStats(ListContainerStatsRequest) returns (ListContainerStatsResponse) {}
-
- // UpdateRuntimeConfig updates the runtime configuration based on the given request.
- rpc UpdateRuntimeConfig(UpdateRuntimeConfigRequest) returns (UpdateRuntimeConfigResponse) {}
-
- // Status returns the status of the runtime.
- rpc Status(StatusRequest) returns (StatusResponse) {}
-}
-
-// ImageService defines the public APIs for managing images.
-service ImageService {
- // ListImages lists existing images.
- rpc ListImages(ListImagesRequest) returns (ListImagesResponse) {}
- // ImageStatus returns the status of the image. If the image is not
- // present, returns a response with ImageStatusResponse.Image set to
- // nil.
- rpc ImageStatus(ImageStatusRequest) returns (ImageStatusResponse) {}
- // PullImage pulls an image with authentication config.
- rpc PullImage(PullImageRequest) returns (PullImageResponse) {}
- // RemoveImage removes the image.
- // This call is idempotent, and must not return an error if the image has
- // already been removed.
- rpc RemoveImage(RemoveImageRequest) returns (RemoveImageResponse) {}
- // ImageFSInfo returns information of the filesystem that is used to store images.
- rpc ImageFsInfo(ImageFsInfoRequest) returns (ImageFsInfoResponse) {}
-}
-
-message VersionRequest {
- // Version of the kubelet runtime API.
- string version = 1;
-}
-
-message VersionResponse {
- // Version of the kubelet runtime API.
- string version = 1;
- // Name of the container runtime.
- string runtime_name = 2;
- // Version of the container runtime. The string must be
- // semver-compatible.
- string runtime_version = 3;
- // API version of the container runtime. The string must be
- // semver-compatible.
- string runtime_api_version = 4;
-}
-
-// DNSConfig specifies the DNS servers and search domains of a sandbox.
-message DNSConfig {
- // List of DNS servers of the cluster.
- repeated string servers = 1;
- // List of DNS search domains of the cluster.
- repeated string searches = 2;
- // List of DNS options. See https://linux.die.net/man/5/resolv.conf
- // for all available options.
- repeated string options = 3;
-}
-
-enum Protocol {
- TCP = 0;
- UDP = 1;
-}
-
-// PortMapping specifies the port mapping configurations of a sandbox.
-message PortMapping {
- // Protocol of the port mapping.
- Protocol protocol = 1;
- // Port number within the container. Default: 0 (not specified).
- int32 container_port = 2;
- // Port number on the host. Default: 0 (not specified).
- int32 host_port = 3;
- // Host IP.
- string host_ip = 4;
-}
-
-enum MountPropagation {
- // No mount propagation ("private" in Linux terminology).
- PROPAGATION_PRIVATE = 0;
- // Mounts get propagated from the host to the container ("rslave" in Linux).
- PROPAGATION_HOST_TO_CONTAINER = 1;
- // Mounts get propagated from the host to the container and from the
- // container to the host ("rshared" in Linux).
- PROPAGATION_BIDIRECTIONAL = 2;
-}
-
-// Mount specifies a host volume to mount into a container.
-message Mount {
- // Path of the mount within the container.
- string container_path = 1;
- // Path of the mount on the host.
- string host_path = 2;
- // If set, the mount is read-only.
- bool readonly = 3;
- // If set, the mount needs SELinux relabeling.
- bool selinux_relabel = 4;
- // Requested propagation mode.
- MountPropagation propagation = 5;
-}
-
-// A NamespaceMode describes the intended namespace configuration for each
-// of the namespaces (Network, PID, IPC) in NamespaceOption. Runtimes should
-// map these modes as appropriate for the technology underlying the runtime.
-enum NamespaceMode {
- // A POD namespace is common to all containers in a pod.
- // For example, a container with a PID namespace of POD expects to view
- // all of the processes in all of the containers in the pod.
- POD = 0;
- // A CONTAINER namespace is restricted to a single container.
- // For example, a container with a PID namespace of CONTAINER expects to
- // view only the processes in that container.
- CONTAINER = 1;
- // A NODE namespace is the namespace of the Kubernetes node.
- // For example, a container with a PID namespace of NODE expects to view
- // all of the processes on the host running the kubelet.
- NODE = 2;
-}
-
-// NamespaceOption provides options for Linux namespaces.
-message NamespaceOption {
- // Network namespace for this container/sandbox.
- // Note: There is currently no way to set CONTAINER scoped network in the Kubernetes API.
- // Namespaces currently set by the kubelet: POD, NODE
- NamespaceMode network = 1;
- // PID namespace for this container/sandbox.
- // Note: The CRI default is POD, but the v1.PodSpec default is CONTAINER.
- // The kubelet's runtime manager will set this to CONTAINER explicitly for v1 pods.
- // Namespaces currently set by the kubelet: POD, CONTAINER, NODE
- NamespaceMode pid = 2;
- // IPC namespace for this container/sandbox.
- // Note: There is currently no way to set CONTAINER scoped IPC in the Kubernetes API.
- // Namespaces currently set by the kubelet: POD, NODE
- NamespaceMode ipc = 3;
-}
-
-// Int64Value is the wrapper of int64.
-message Int64Value {
- // The value.
- int64 value = 1;
-}
-
-// LinuxSandboxSecurityContext holds linux security configuration that will be
-// applied to a sandbox. Note that:
-// 1) It does not apply to containers in the pods.
-// 2) It may not be applicable to a PodSandbox which does not contain any running
-// process.
-message LinuxSandboxSecurityContext {
- // Configurations for the sandbox's namespaces.
- // This will be used only if the PodSandbox uses namespace for isolation.
- NamespaceOption namespace_options = 1;
- // Optional SELinux context to be applied.
- SELinuxOption selinux_options = 2;
- // UID to run sandbox processes as, when applicable.
- Int64Value run_as_user = 3;
- // If set, the root filesystem of the sandbox is read-only.
- bool readonly_rootfs = 4;
- // List of groups applied to the first process run in the sandbox, in
- // addition to the sandbox's primary GID.
- repeated int64 supplemental_groups = 5;
- // Indicates whether the sandbox will be asked to run a privileged
- // container. If a privileged container is to be executed within it, this
- // MUST be true.
- // This allows a sandbox to take additional security precautions if no
- // privileged containers are expected to be run.
- bool privileged = 6;
- // Seccomp profile for the sandbox, candidate values are:
- // * docker/default: the default profile for the docker container runtime
- // * unconfined: unconfined profile, ie, no seccomp sandboxing
- // * localhost/<full-path-to-profile>: the profile installed on the node.
- // <full-path-to-profile> is the full path of the profile.
- // Default: "", which is identical with unconfined.
- string seccomp_profile_path = 7;
-}
-
-// LinuxPodSandboxConfig holds platform-specific configurations for Linux
-// host platforms and Linux-based containers.
-message LinuxPodSandboxConfig {
- // Parent cgroup of the PodSandbox.
- // The cgroupfs style syntax will be used, but the container runtime can
- // convert it to systemd semantics if needed.
- string cgroup_parent = 1;
- // LinuxSandboxSecurityContext holds sandbox security attributes.
- LinuxSandboxSecurityContext security_context = 2;
- // Sysctls holds linux sysctls config for the sandbox.
- map<string, string> sysctls = 3;
-}
-
-// PodSandboxMetadata holds all necessary information for building the sandbox name.
-// The container runtime is encouraged to expose the metadata associated with the
-// PodSandbox in its user interface for better user experience. For example,
-// the runtime can construct a unique PodSandboxName based on the metadata.
-message PodSandboxMetadata {
- // Pod name of the sandbox. Same as the pod name in the PodSpec.
- string name = 1;
- // Pod UID of the sandbox. Same as the pod UID in the PodSpec.
- string uid = 2;
- // Pod namespace of the sandbox. Same as the pod namespace in the PodSpec.
- string namespace = 3;
- // Attempt number of creating the sandbox. Default: 0.
- uint32 attempt = 4;
-}
-
-// PodSandboxConfig holds all the required and optional fields for creating a
-// sandbox.
-message PodSandboxConfig {
- // Metadata of the sandbox. This information will uniquely identify the
- // sandbox, and the runtime should leverage this to ensure correct
- // operation. The runtime may also use this information to improve UX, such
- // as by constructing a readable name.
- PodSandboxMetadata metadata = 1;
- // Hostname of the sandbox.
- string hostname = 2;
- // Path to the directory on the host in which container log files are
- // stored.
- // By default the log of a container going into the LogDirectory will be
- // hooked up to STDOUT and STDERR. However, the LogDirectory may contain
- // binary log files with structured logging data from the individual
- // containers. For example, the files might be newline separated JSON
- // structured logs, systemd-journald journal files, gRPC trace files, etc.
- // E.g.,
- // PodSandboxConfig.LogDirectory = `/var/log/pods/<podUID>/`
- // ContainerConfig.LogPath = `containerName_Instance#.log`
- //
- // WARNING: Log management and how kubelet should interface with the
- // container logs are under active discussion in
- // https://issues.k8s.io/24677. There *may* be future change of direction
- // for logging as the discussion carries on.
- string log_directory = 3;
- // DNS config for the sandbox.
- DNSConfig dns_config = 4;
- // Port mappings for the sandbox.
- repeated PortMapping port_mappings = 5;
- // Key-value pairs that may be used to scope and select individual resources.
- map<string, string> labels = 6;
- // Unstructured key-value map that may be set by the kubelet to store and
- // retrieve arbitrary metadata. This will include any annotations set on a
- // pod through the Kubernetes API.
- //
- // Annotations MUST NOT be altered by the runtime; the annotations stored
- // here MUST be returned in the PodSandboxStatus associated with the pod
- // this PodSandboxConfig creates.
- //
- // In general, in order to preserve a well-defined interface between the
- // kubelet and the container runtime, annotations SHOULD NOT influence
- // runtime behaviour.
- //
- // Annotations can also be useful for runtime authors to experiment with
- // new features that are opaque to the Kubernetes APIs (both user-facing
- // and the CRI). Whenever possible, however, runtime authors SHOULD
- // consider proposing new typed fields for any new features instead.
- map<string, string> annotations = 7;
- // Optional configurations specific to Linux hosts.
- LinuxPodSandboxConfig linux = 8;
-}
-
-message RunPodSandboxRequest {
- // Configuration for creating a PodSandbox.
- PodSandboxConfig config = 1;
-}
-
-message RunPodSandboxResponse {
- // ID of the PodSandbox to run.
- string pod_sandbox_id = 1;
-}
-
-message StopPodSandboxRequest {
- // ID of the PodSandbox to stop.
- string pod_sandbox_id = 1;
-}
-
-message StopPodSandboxResponse {}
-
-message RemovePodSandboxRequest {
- // ID of the PodSandbox to remove.
- string pod_sandbox_id = 1;
-}
-
-message RemovePodSandboxResponse {}
-
-message PodSandboxStatusRequest {
- // ID of the PodSandbox for which to retrieve status.
- string pod_sandbox_id = 1;
- // Verbose indicates whether to return extra information about the pod sandbox.
- bool verbose = 2;
-}
-
-// PodSandboxNetworkStatus is the status of the network for a PodSandbox.
-message PodSandboxNetworkStatus {
- // IP address of the PodSandbox.
- string ip = 1;
-}
-
-// Namespace contains paths to the namespaces.
-message Namespace {
- // Namespace options for Linux namespaces.
- NamespaceOption options = 2;
-}
-
-// LinuxSandboxStatus contains status specific to Linux sandboxes.
-message LinuxPodSandboxStatus {
- // Paths to the sandbox's namespaces.
- Namespace namespaces = 1;
-}
-
-enum PodSandboxState {
- SANDBOX_READY = 0;
- SANDBOX_NOTREADY = 1;
-}
-
-// PodSandboxStatus contains the status of the PodSandbox.
-message PodSandboxStatus {
- // ID of the sandbox.
- string id = 1;
- // Metadata of the sandbox.
- PodSandboxMetadata metadata = 2;
- // State of the sandbox.
- PodSandboxState state = 3;
- // Creation timestamp of the sandbox in nanoseconds. Must be > 0.
- int64 created_at = 4;
- // Network contains network status if network is handled by the runtime.
- PodSandboxNetworkStatus network = 5;
- // Linux-specific status to a pod sandbox.
- LinuxPodSandboxStatus linux = 6;
- // Labels are key-value pairs that may be used to scope and select individual resources.
- map<string, string> labels = 7;
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding PodSandboxConfig used to
- // instantiate the pod sandbox this status represents.
- map<string, string> annotations = 8;
-}
-
-message PodSandboxStatusResponse {
- // Status of the PodSandbox.
- PodSandboxStatus status = 1;
- // Info is extra information of the PodSandbox. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. network namespace for linux container based container runtime.
- // It should only be returned non-empty when Verbose is true.
- map<string, string> info = 2;
-}
-
-// PodSandboxStateValue is the wrapper of PodSandboxState.
-message PodSandboxStateValue {
- // State of the sandbox.
- PodSandboxState state = 1;
-}
-
-// PodSandboxFilter is used to filter a list of PodSandboxes.
-// All those fields are combined with 'AND'
-message PodSandboxFilter {
- // ID of the sandbox.
- string id = 1;
- // State of the sandbox.
- PodSandboxStateValue state = 2;
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- map<string, string> label_selector = 3;
-}
-
-message ListPodSandboxRequest {
- // PodSandboxFilter to filter a list of PodSandboxes.
- PodSandboxFilter filter = 1;
-}
-
-
-// PodSandbox contains minimal information about a sandbox.
-message PodSandbox {
- // ID of the PodSandbox.
- string id = 1;
- // Metadata of the PodSandbox.
- PodSandboxMetadata metadata = 2;
- // State of the PodSandbox.
- PodSandboxState state = 3;
- // Creation timestamps of the PodSandbox in nanoseconds. Must be > 0.
- int64 created_at = 4;
- // Labels of the PodSandbox.
- map<string, string> labels = 5;
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding PodSandboxConfig used to
- // instantiate this PodSandbox.
- map<string, string> annotations = 6;
-}
-
-message ListPodSandboxResponse {
- // List of PodSandboxes.
- repeated PodSandbox items = 1;
-}
-
-// ImageSpec is an internal representation of an image. Currently, it wraps the
-// value of a Container's Image field (e.g. imageID or imageDigest), but in the
-// future it will include more detailed information about the different image types.
-message ImageSpec {
- string image = 1;
-}
-
-message KeyValue {
- string key = 1;
- string value = 2;
-}
-
-// LinuxContainerResources specifies Linux specific configuration for
-// resources.
-// TODO: Consider using Resources from opencontainers/runtime-spec/specs-go
-// directly.
-message LinuxContainerResources {
- // CPU CFS (Completely Fair Scheduler) period. Default: 0 (not specified).
- int64 cpu_period = 1;
- // CPU CFS (Completely Fair Scheduler) quota. Default: 0 (not specified).
- int64 cpu_quota = 2;
- // CPU shares (relative weight vs. other containers). Default: 0 (not specified).
- int64 cpu_shares = 3;
- // Memory limit in bytes. Default: 0 (not specified).
- int64 memory_limit_in_bytes = 4;
- // OOMScoreAdj adjusts the oom-killer score. Default: 0 (not specified).
- int64 oom_score_adj = 5;
- // CpusetCpus constrains the allowed set of logical CPUs. Default: "" (not specified).
- string cpuset_cpus = 6;
- // CpusetMems constrains the allowed set of memory nodes. Default: "" (not specified).
- string cpuset_mems = 7;
-}
-
-// SELinuxOption are the labels to be applied to the container.
-message SELinuxOption {
- string user = 1;
- string role = 2;
- string type = 3;
- string level = 4;
-}
-
-// Capability contains the container capabilities to add or drop
-message Capability {
- // List of capabilities to add.
- repeated string add_capabilities = 1;
- // List of capabilities to drop.
- repeated string drop_capabilities = 2;
-}
-
-// LinuxContainerSecurityContext holds linux security configuration that will be applied to a container.
-message LinuxContainerSecurityContext {
- // Capabilities to add or drop.
- Capability capabilities = 1;
- // If set, run container in privileged mode.
- // Privileged mode is incompatible with the following options. If
- // privileged is set, the following features MAY have no effect:
- // 1. capabilities
- // 2. selinux_options
- // 4. seccomp
- // 5. apparmor
- //
- // Privileged mode implies the following specific options are applied:
- // 1. All capabilities are added.
- // 2. Sensitive paths, such as kernel module paths within sysfs, are not masked.
- // 3. Any sysfs and procfs mounts are mounted RW.
- // 4. Apparmor confinement is not applied.
- // 5. Seccomp restrictions are not applied.
- // 6. The device cgroup does not restrict access to any devices.
- // 7. All devices from the host's /dev are available within the container.
- // 8. SELinux restrictions are not applied (e.g. label=disabled).
- bool privileged = 2;
- // Configurations for the container's namespaces.
- // Only used if the container uses namespace for isolation.
- NamespaceOption namespace_options = 3;
- // SELinux context to be optionally applied.
- SELinuxOption selinux_options = 4;
- // UID to run the container process as. Only one of run_as_user and
- // run_as_username can be specified at a time.
- Int64Value run_as_user = 5;
- // User name to run the container process as. If specified, the user MUST
- // exist in the container image (i.e. in the /etc/passwd inside the image),
- // and be resolved there by the runtime; otherwise, the runtime MUST error.
- string run_as_username = 6;
- // If set, the root filesystem of the container is read-only.
- bool readonly_rootfs = 7;
- // List of groups applied to the first process run in the container, in
- // addition to the container's primary GID.
- repeated int64 supplemental_groups = 8;
- // AppArmor profile for the container, candidate values are:
- // * runtime/default: equivalent to not specifying a profile.
- // * unconfined: no profiles are loaded
- // * localhost/<profile_name>: profile loaded on the node
- // (localhost) by name. The possible profile names are detailed at
- // http://wiki.apparmor.net/index.php/AppArmor_Core_Policy_Reference
- string apparmor_profile = 9;
- // Seccomp profile for the container, candidate values are:
- // * docker/default: the default profile for the docker container runtime
- // * unconfined: unconfined profile, ie, no seccomp sandboxing
- // * localhost/<full-path-to-profile>: the profile installed on the node.
- // <full-path-to-profile> is the full path of the profile.
- // Default: "", which is identical with unconfined.
- string seccomp_profile_path = 10;
- // no_new_privs defines if the flag for no_new_privs should be set on the
- // container.
- bool no_new_privs = 11;
-}
-
-// LinuxContainerConfig contains platform-specific configuration for
-// Linux-based containers.
-message LinuxContainerConfig {
- // Resources specification for the container.
- LinuxContainerResources resources = 1;
- // LinuxContainerSecurityContext configuration for the container.
- LinuxContainerSecurityContext security_context = 2;
-}
-
-// WindowsContainerConfig contains platform-specific configuration for
-// Windows-based containers.
-message WindowsContainerConfig {
- // Resources specification for the container.
- WindowsContainerResources resources = 1;
-}
-
-// WindowsContainerResources specifies Windows specific configuration for
-// resources.
-message WindowsContainerResources {
- // CPU shares (relative weight vs. other containers). Default: 0 (not specified).
- int64 cpu_shares = 1;
- // Number of CPUs available to the container. Default: 0 (not specified).
- int64 cpu_count = 2;
- // Specifies the portion of processor cycles that this container can use as a percentage times 100.
- int64 cpu_maximum = 3;
- // Memory limit in bytes. Default: 0 (not specified).
- int64 memory_limit_in_bytes = 4;
-}
-
-// ContainerMetadata holds all necessary information for building the container
-// name. The container runtime is encouraged to expose the metadata in its user
-// interface for better user experience. E.g., runtime can construct a unique
-// container name based on the metadata. Note that (name, attempt) is unique
-// within a sandbox for the entire lifetime of the sandbox.
-message ContainerMetadata {
- // Name of the container. Same as the container name in the PodSpec.
- string name = 1;
- // Attempt number of creating the container. Default: 0.
- uint32 attempt = 2;
-}
-
-// Device specifies a host device to mount into a container.
-message Device {
- // Path of the device within the container.
- string container_path = 1;
- // Path of the device on the host.
- string host_path = 2;
- // Cgroups permissions of the device, candidates are one or more of
- // * r - allows container to read from the specified device.
- // * w - allows container to write to the specified device.
- // * m - allows container to create device files that do not yet exist.
- string permissions = 3;
-}
-
-// ContainerConfig holds all the required and optional fields for creating a
-// container.
-message ContainerConfig {
- // Metadata of the container. This information will uniquely identify the
- // container, and the runtime should leverage this to ensure correct
- // operation. The runtime may also use this information to improve UX, such
- // as by constructing a readable name.
- ContainerMetadata metadata = 1 ;
- // Image to use.
- ImageSpec image = 2;
- // Command to execute (i.e., entrypoint for docker)
- repeated string command = 3;
- // Args for the Command (i.e., command for docker)
- repeated string args = 4;
- // Current working directory of the command.
- string working_dir = 5;
- // List of environment variable to set in the container.
- repeated KeyValue envs = 6;
- // Mounts for the container.
- repeated Mount mounts = 7;
- // Devices for the container.
- repeated Device devices = 8;
- // Key-value pairs that may be used to scope and select individual resources.
- // Label keys are of the form:
- // label-key ::= prefixed-name | name
- // prefixed-name ::= prefix '/' name
- // prefix ::= DNS_SUBDOMAIN
- // name ::= DNS_LABEL
- map<string, string> labels = 9;
- // Unstructured key-value map that may be used by the kubelet to store and
- // retrieve arbitrary metadata.
- //
- // Annotations MUST NOT be altered by the runtime; the annotations stored
- // here MUST be returned in the ContainerStatus associated with the container
- // this ContainerConfig creates.
- //
- // In general, in order to preserve a well-defined interface between the
- // kubelet and the container runtime, annotations SHOULD NOT influence
- // runtime behaviour.
- map<string, string> annotations = 10;
- // Path relative to PodSandboxConfig.LogDirectory for container to store
- // the log (STDOUT and STDERR) on the host.
- // E.g.,
- // PodSandboxConfig.LogDirectory = `/var/log/pods/<podUID>/`
- // ContainerConfig.LogPath = `containerName_Instance#.log`
- //
- // WARNING: Log management and how kubelet should interface with the
- // container logs are under active discussion in
- // https://issues.k8s.io/24677. There *may* be future change of direction
- // for logging as the discussion carries on.
- string log_path = 11;
-
- // Variables for interactive containers, these have very specialized
- // use-cases (e.g. debugging).
- // TODO: Determine if we need to continue supporting these fields that are
- // part of Kubernetes's Container Spec.
- bool stdin = 12;
- bool stdin_once = 13;
- bool tty = 14;
-
- // Configuration specific to Linux containers.
- LinuxContainerConfig linux = 15;
- // Configuration specific to Windows containers.
- WindowsContainerConfig windows = 16;
-}
-
-message CreateContainerRequest {
- // ID of the PodSandbox in which the container should be created.
- string pod_sandbox_id = 1;
- // Config of the container.
- ContainerConfig config = 2;
- // Config of the PodSandbox. This is the same config that was passed
- // to RunPodSandboxRequest to create the PodSandbox. It is passed again
- // here just for easy reference. The PodSandboxConfig is immutable and
- // remains the same throughout the lifetime of the pod.
- PodSandboxConfig sandbox_config = 3;
-}
-
-message CreateContainerResponse {
- // ID of the created container.
- string container_id = 1;
-}
-
-message StartContainerRequest {
- // ID of the container to start.
- string container_id = 1;
-}
-
-message StartContainerResponse {}
-
-message StopContainerRequest {
- // ID of the container to stop.
- string container_id = 1;
- // Timeout in seconds to wait for the container to stop before forcibly
- // terminating it. Default: 0 (forcibly terminate the container immediately)
- int64 timeout = 2;
-}
-
-message StopContainerResponse {}
-
-message RemoveContainerRequest {
- // ID of the container to remove.
- string container_id = 1;
-}
-
-message RemoveContainerResponse {}
-
-enum ContainerState {
- CONTAINER_CREATED = 0;
- CONTAINER_RUNNING = 1;
- CONTAINER_EXITED = 2;
- CONTAINER_UNKNOWN = 3;
-}
-
-// ContainerStateValue is the wrapper of ContainerState.
-message ContainerStateValue {
- // State of the container.
- ContainerState state = 1;
-}
-
-// ContainerFilter is used to filter containers.
-// All those fields are combined with 'AND'
-message ContainerFilter {
- // ID of the container.
- string id = 1;
- // State of the container.
- ContainerStateValue state = 2;
- // ID of the PodSandbox.
- string pod_sandbox_id = 3;
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- map<string, string> label_selector = 4;
-}
-
-message ListContainersRequest {
- ContainerFilter filter = 1;
-}
-
-// Container provides the runtime information for a container, such as ID, hash,
-// state of the container.
-message Container {
- // ID of the container, used by the container runtime to identify
- // a container.
- string id = 1;
- // ID of the sandbox to which this container belongs.
- string pod_sandbox_id = 2;
- // Metadata of the container.
- ContainerMetadata metadata = 3;
- // Spec of the image.
- ImageSpec image = 4;
- // Reference to the image in use. For most runtimes, this should be an
- // image ID.
- string image_ref = 5;
- // State of the container.
- ContainerState state = 6;
- // Creation time of the container in nanoseconds.
- int64 created_at = 7;
- // Key-value pairs that may be used to scope and select individual resources.
- map<string, string> labels = 8;
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate this Container.
- map<string, string> annotations = 9;
-}
-
-message ListContainersResponse {
- // List of containers.
- repeated Container containers = 1;
-}
-
-message ContainerStatusRequest {
- // ID of the container for which to retrieve status.
- string container_id = 1;
- // Verbose indicates whether to return extra information about the container.
- bool verbose = 2;
-}
-
-// ContainerStatus represents the status of a container.
-message ContainerStatus {
- // ID of the container.
- string id = 1;
- // Metadata of the container.
- ContainerMetadata metadata = 2;
- // Status of the container.
- ContainerState state = 3;
- // Creation time of the container in nanoseconds.
- int64 created_at = 4;
- // Start time of the container in nanoseconds. Default: 0 (not specified).
- int64 started_at = 5;
- // Finish time of the container in nanoseconds. Default: 0 (not specified).
- int64 finished_at = 6;
- // Exit code of the container. Only required when finished_at != 0. Default: 0.
- int32 exit_code = 7;
- // Spec of the image.
- ImageSpec image = 8;
- // Reference to the image in use. For most runtimes, this should be an
- // image ID
- string image_ref = 9;
- // Brief CamelCase string explaining why container is in its current state.
- string reason = 10;
- // Human-readable message indicating details about why container is in its
- // current state.
- string message = 11;
- // Key-value pairs that may be used to scope and select individual resources.
- map<string,string> labels = 12;
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate the Container this status represents.
- map<string,string> annotations = 13;
- // Mounts for the container.
- repeated Mount mounts = 14;
- // Log path of container.
- string log_path = 15;
-}
-
-message ContainerStatusResponse {
- // Status of the container.
- ContainerStatus status = 1;
- // Info is extra information of the Container. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. pid for linux container based container runtime.
- // It should only be returned non-empty when Verbose is true.
- map<string, string> info = 2;
-}
-
-message UpdateContainerResourcesRequest {
- // ID of the container to update.
- string container_id = 1;
- // Resource configuration specific to Linux containers.
- LinuxContainerResources linux = 2;
-}
-
-message UpdateContainerResourcesResponse {}
-
-message ExecSyncRequest {
- // ID of the container.
- string container_id = 1;
- // Command to execute.
- repeated string cmd = 2;
- // Timeout in seconds to stop the command. Default: 0 (run forever).
- int64 timeout = 3;
-}
-
-message ExecSyncResponse {
- // Captured command stdout output.
- bytes stdout = 1;
- // Captured command stderr output.
- bytes stderr = 2;
- // Exit code the command finished with. Default: 0 (success).
- int32 exit_code = 3;
-}
-
-message ExecRequest {
- // ID of the container in which to execute the command.
- string container_id = 1;
- // Command to execute.
- repeated string cmd = 2;
- // Whether to exec the command in a TTY.
- bool tty = 3;
- // Whether to stream stdin.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- bool stdin = 4;
- // Whether to stream stdout.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- bool stdout = 5;
- // Whether to stream stderr.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- // If `tty` is true, `stderr` MUST be false. Multiplexing is not supported
- // in this case. The output of stdout and stderr will be combined to a
- // single stream.
- bool stderr = 6;
-}
-
-message ExecResponse {
- // Fully qualified URL of the exec streaming server.
- string url = 1;
-}
-
-message AttachRequest {
- // ID of the container to which to attach.
- string container_id = 1;
- // Whether to stream stdin.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- bool stdin = 2;
- // Whether the process being attached is running in a TTY.
- // This must match the TTY setting in the ContainerConfig.
- bool tty = 3;
- // Whether to stream stdout.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- bool stdout = 4;
- // Whether to stream stderr.
- // One of `stdin`, `stdout`, and `stderr` MUST be true.
- // If `tty` is true, `stderr` MUST be false. Multiplexing is not supported
- // in this case. The output of stdout and stderr will be combined to a
- // single stream.
- bool stderr = 5;
-}
-
-message AttachResponse {
- // Fully qualified URL of the attach streaming server.
- string url = 1;
-}
-
-message PortForwardRequest {
- // ID of the container to which to forward the port.
- string pod_sandbox_id = 1;
- // Port to forward.
- repeated int32 port = 2;
-}
-
-message PortForwardResponse {
- // Fully qualified URL of the port-forward streaming server.
- string url = 1;
-}
-
-message ImageFilter {
- // Spec of the image.
- ImageSpec image = 1;
-}
-
-message ListImagesRequest {
- // Filter to list images.
- ImageFilter filter = 1;
-}
-
-// Basic information about a container image.
-message Image {
- // ID of the image.
- string id = 1;
- // Other names by which this image is known.
- repeated string repo_tags = 2;
- // Digests by which this image is known.
- repeated string repo_digests = 3;
- // Size of the image in bytes. Must be > 0.
- uint64 size = 4;
- // UID that will run the command(s). This is used as a default if no user is
- // specified when creating the container. UID and the following user name
- // are mutually exclusive.
- Int64Value uid = 5;
- // User name that will run the command(s). This is used if UID is not set
- // and no user is specified when creating container.
- string username = 6;
-}
-
-message ListImagesResponse {
- // List of images.
- repeated Image images = 1;
-}
-
-message ImageStatusRequest {
- // Spec of the image.
- ImageSpec image = 1;
- // Verbose indicates whether to return extra information about the image.
- bool verbose = 2;
-}
-
-message ImageStatusResponse {
- // Status of the image.
- Image image = 1;
- // Info is extra information of the Image. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful
- // for debug, e.g. image config for oci image based container runtime.
- // It should only be returned non-empty when Verbose is true.
- map<string, string> info = 2;
-}
-
-// AuthConfig contains authorization information for connecting to a registry.
-message AuthConfig {
- string username = 1;
- string password = 2;
- string auth = 3;
- string server_address = 4;
- // IdentityToken is used to authenticate the user and get
- // an access token for the registry.
- string identity_token = 5;
- // RegistryToken is a bearer token to be sent to a registry
- string registry_token = 6;
-}
-
-message PullImageRequest {
- // Spec of the image.
- ImageSpec image = 1;
- // Authentication configuration for pulling the image.
- AuthConfig auth = 2;
- // Config of the PodSandbox, which is used to pull image in PodSandbox context.
- PodSandboxConfig sandbox_config = 3;
-}
-
-message PullImageResponse {
- // Reference to the image in use. For most runtimes, this should be an
- // image ID or digest.
- string image_ref = 1;
-}
-
-message RemoveImageRequest {
- // Spec of the image to remove.
- ImageSpec image = 1;
-}
-
-message RemoveImageResponse {}
-
-message NetworkConfig {
- // CIDR to use for pod IP addresses.
- string pod_cidr = 1;
-}
-
-message RuntimeConfig {
- NetworkConfig network_config = 1;
-}
-
-message UpdateRuntimeConfigRequest {
- RuntimeConfig runtime_config = 1;
-}
-
-message UpdateRuntimeConfigResponse {}
-
-// RuntimeCondition contains condition information for the runtime.
-// There are 2 kinds of runtime conditions:
-// 1. Required conditions: Conditions are required for kubelet to work
-// properly. If any required condition is unmet, the node will be not ready.
-// The required conditions include:
-// * RuntimeReady: RuntimeReady means the runtime is up and ready to accept
-// basic containers e.g. container only needs host network.
-// * NetworkReady: NetworkReady means the runtime network is up and ready to
-// accept containers which require container network.
-// 2. Optional conditions: Conditions are informative to the user, but kubelet
-// will not rely on. Since condition type is an arbitrary string, all conditions
-// not required are optional. These conditions will be exposed to users to help
-// them understand the status of the system.
-message RuntimeCondition {
- // Type of runtime condition.
- string type = 1;
- // Status of the condition, one of true/false. Default: false.
- bool status = 2;
- // Brief CamelCase string containing reason for the condition's last transition.
- string reason = 3;
- // Human-readable message indicating details about last transition.
- string message = 4;
-}
-
-// RuntimeStatus is information about the current status of the runtime.
-message RuntimeStatus {
- // List of current observed runtime conditions.
- repeated RuntimeCondition conditions = 1;
-}
-
-message StatusRequest {
- // Verbose indicates whether to return extra information about the runtime.
- bool verbose = 1;
-}
-
-message StatusResponse {
- // Status of the Runtime.
- RuntimeStatus status = 1;
- // Info is extra information of the Runtime. The key could be arbitrary string, and
- // value should be in json format. The information could include anything useful for
- // debug, e.g. plugins used by the container runtime.
- // It should only be returned non-empty when Verbose is true.
- map<string, string> info = 2;
-}
-
-message ImageFsInfoRequest {}
-
-// UInt64Value is the wrapper of uint64.
-message UInt64Value {
- // The value.
- uint64 value = 1;
-}
-
-// FilesystemIdentifier uniquely identify the filesystem.
-message FilesystemIdentifier{
- // Mountpoint of a filesystem.
- string mountpoint = 1;
-}
-
-// FilesystemUsage provides the filesystem usage information.
-message FilesystemUsage {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- int64 timestamp = 1;
- // The unique identifier of the filesystem.
- FilesystemIdentifier fs_id = 2;
- // UsedBytes represents the bytes used for images on the filesystem.
- // This may differ from the total bytes used on the filesystem and may not
- // equal CapacityBytes - AvailableBytes.
- UInt64Value used_bytes = 3;
- // InodesUsed represents the inodes used by the images.
- // This may not equal InodesCapacity - InodesAvailable because the underlying
- // filesystem may also be used for purposes other than storing images.
- UInt64Value inodes_used = 4;
-}
-
-message ImageFsInfoResponse {
- // Information of image filesystem(s).
- repeated FilesystemUsage image_filesystems = 1;
-}
-
-message ContainerStatsRequest{
- // ID of the container for which to retrieve stats.
- string container_id = 1;
-}
-
-message ContainerStatsResponse {
- // Stats of the container.
- ContainerStats stats = 1;
-}
-
-message ListContainerStatsRequest{
- // Filter for the list request.
- ContainerStatsFilter filter = 1;
-}
-
-// ContainerStatsFilter is used to filter containers.
-// All those fields are combined with 'AND'
-message ContainerStatsFilter {
- // ID of the container.
- string id = 1;
- // ID of the PodSandbox.
- string pod_sandbox_id = 2;
- // LabelSelector to select matches.
- // Only api.MatchLabels is supported for now and the requirements
- // are ANDed. MatchExpressions is not supported yet.
- map<string, string> label_selector = 3;
-}
-
-message ListContainerStatsResponse {
- // Stats of the container.
- repeated ContainerStats stats = 1;
-}
-
-// ContainerAttributes provides basic information of the container.
-message ContainerAttributes {
- // ID of the container.
- string id = 1;
- // Metadata of the container.
- ContainerMetadata metadata = 2;
- // Key-value pairs that may be used to scope and select individual resources.
- map<string,string> labels = 3;
- // Unstructured key-value map holding arbitrary metadata.
- // Annotations MUST NOT be altered by the runtime; the value of this field
- // MUST be identical to that of the corresponding ContainerConfig used to
- // instantiate the Container this status represents.
- map<string,string> annotations = 4;
-}
-
-// ContainerStats provides the resource usage statistics for a container.
-message ContainerStats {
- // Information of the container.
- ContainerAttributes attributes = 1;
- // CPU usage gathered from the container.
- CpuUsage cpu = 2;
- // Memory usage gathered from the container.
- MemoryUsage memory = 3;
- // Usage of the writeable layer.
- FilesystemUsage writable_layer = 4;
-}
-
-// CpuUsage provides the CPU usage information.
-message CpuUsage {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- int64 timestamp = 1;
- // Cumulative CPU usage (sum across all cores) since object creation.
- UInt64Value usage_core_nano_seconds = 2;
-}
-
-// MemoryUsage provides the memory usage information.
-message MemoryUsage {
- // Timestamp in nanoseconds at which the information were collected. Must be > 0.
- int64 timestamp = 1;
- // The amount of working set memory in bytes.
- UInt64Value working_set_bytes = 2;
-}
-
-message ReopenContainerLogRequest {
- // ID of the container for which to reopen the log.
- string container_id = 1;
-}
-
-message ReopenContainerLogResponse{
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/constants.go b/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/constants.go
deleted file mode 100644
index 0e141b7d7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2/constants.go
+++ /dev/null
@@ -1,55 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package v1alpha2
-
-// This file contains all constants defined in CRI.
-
-// Required runtime condition type.
-const (
- // RuntimeReady means the runtime is up and ready to accept basic containers.
- RuntimeReady = "RuntimeReady"
- // NetworkReady means the runtime network is up and ready to accept containers which require network.
- NetworkReady = "NetworkReady"
-)
-
-// LogStreamType is the type of the stream in CRI container log.
-type LogStreamType string
-
-const (
- // Stdout is the stream type for stdout.
- Stdout LogStreamType = "stdout"
- // Stderr is the stream type for stderr.
- Stderr LogStreamType = "stderr"
-)
-
-// LogTag is the tag of a log line in CRI container log.
-// Currently defined log tags:
-// * First tag: Partial/Full - P/F.
-// The field in the container log format can be extended to include multiple
-// tags by using a delimiter, but changes should be rare. If it becomes clear
-// that better extensibility is desired, a more extensible format (e.g., json)
-// should be adopted as a replacement and/or addition.
-type LogTag string
-
-const (
- // LogTagPartial means the line is part of multiple lines.
- LogTagPartial LogTag = "P"
- // LogTagFull means the line is a single full line or the end of multiple lines.
- LogTagFull LogTag = "F"
- // LogTagDelimiter is the delimiter for different log tags.
- LogTagDelimiter = ":"
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/cache.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/cache.go
deleted file mode 100644
index 82852a9d9..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/cache.go
+++ /dev/null
@@ -1,199 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "sync"
- "time"
-
- "k8s.io/apimachinery/pkg/types"
-)
-
-// Cache stores the PodStatus for the pods. It represents *all* the visible
-// pods/containers in the container runtime. All cache entries are at least as
-// new or newer than the global timestamp (set by UpdateTime()), while
-// individual entries may be slightly newer than the global timestamp. If a pod
-// has no states known by the runtime, Cache returns an empty PodStatus object
-// with ID populated.
-//
-// Cache provides two methods to retrive the PodStatus: the non-blocking Get()
-// and the blocking GetNewerThan() method. The component responsible for
-// populating the cache is expected to call Delete() to explicitly free the
-// cache entries.
-type Cache interface {
- Get(types.UID) (*PodStatus, error)
- Set(types.UID, *PodStatus, error, time.Time)
- // GetNewerThan is a blocking call that only returns the status
- // when it is newer than the given time.
- GetNewerThan(types.UID, time.Time) (*PodStatus, error)
- Delete(types.UID)
- UpdateTime(time.Time)
-}
-
-type data struct {
- // Status of the pod.
- status *PodStatus
- // Error got when trying to inspect the pod.
- err error
- // Time when the data was last modified.
- modified time.Time
-}
-
-type subRecord struct {
- time time.Time
- ch chan *data
-}
-
-// cache implements Cache.
-type cache struct {
- // Lock which guards all internal data structures.
- lock sync.RWMutex
- // Map that stores the pod statuses.
- pods map[types.UID]*data
- // A global timestamp represents how fresh the cached data is. All
- // cache content is at the least newer than this timestamp. Note that the
- // timestamp is nil after initialization, and will only become non-nil when
- // it is ready to serve the cached statuses.
- timestamp *time.Time
- // Map that stores the subscriber records.
- subscribers map[types.UID][]*subRecord
-}
-
-// NewCache creates a pod cache.
-func NewCache() Cache {
- return &cache{pods: map[types.UID]*data{}, subscribers: map[types.UID][]*subRecord{}}
-}
-
-// Get returns the PodStatus for the pod; callers are expected not to
-// modify the objects returned.
-func (c *cache) Get(id types.UID) (*PodStatus, error) {
- c.lock.RLock()
- defer c.lock.RUnlock()
- d := c.get(id)
- return d.status, d.err
-}
-
-func (c *cache) GetNewerThan(id types.UID, minTime time.Time) (*PodStatus, error) {
- ch := c.subscribe(id, minTime)
- d := <-ch
- return d.status, d.err
-}
-
-// Set sets the PodStatus for the pod.
-func (c *cache) Set(id types.UID, status *PodStatus, err error, timestamp time.Time) {
- c.lock.Lock()
- defer c.lock.Unlock()
- defer c.notify(id, timestamp)
- c.pods[id] = &data{status: status, err: err, modified: timestamp}
-}
-
-// Delete removes the entry of the pod.
-func (c *cache) Delete(id types.UID) {
- c.lock.Lock()
- defer c.lock.Unlock()
- delete(c.pods, id)
-}
-
-// UpdateTime modifies the global timestamp of the cache and notify
-// subscribers if needed.
-func (c *cache) UpdateTime(timestamp time.Time) {
- c.lock.Lock()
- defer c.lock.Unlock()
- c.timestamp = &timestamp
- // Notify all the subscribers if the condition is met.
- for id := range c.subscribers {
- c.notify(id, *c.timestamp)
- }
-}
-
-func makeDefaultData(id types.UID) *data {
- return &data{status: &PodStatus{ID: id}, err: nil}
-}
-
-func (c *cache) get(id types.UID) *data {
- d, ok := c.pods[id]
- if !ok {
- // Cache should store *all* pod/container information known by the
- // container runtime. A cache miss indicates that there are no states
- // regarding the pod last time we queried the container runtime.
- // What this *really* means is that there are no visible pod/containers
- // associated with this pod. Simply return an default (mostly empty)
- // PodStatus to reflect this.
- return makeDefaultData(id)
- }
- return d
-}
-
-// getIfNewerThan returns the data it is newer than the given time.
-// Otherwise, it returns nil. The caller should acquire the lock.
-func (c *cache) getIfNewerThan(id types.UID, minTime time.Time) *data {
- d, ok := c.pods[id]
- globalTimestampIsNewer := (c.timestamp != nil && c.timestamp.After(minTime))
- if !ok && globalTimestampIsNewer {
- // Status is not cached, but the global timestamp is newer than
- // minTime, return the default status.
- return makeDefaultData(id)
- }
- if ok && (d.modified.After(minTime) || globalTimestampIsNewer) {
- // Status is cached, return status if either of the following is true.
- // * status was modified after minTime
- // * the global timestamp of the cache is newer than minTime.
- return d
- }
- // The pod status is not ready.
- return nil
-}
-
-// notify sends notifications for pod with the given id, if the requirements
-// are met. Note that the caller should acquire the lock.
-func (c *cache) notify(id types.UID, timestamp time.Time) {
- list, ok := c.subscribers[id]
- if !ok {
- // No one to notify.
- return
- }
- newList := []*subRecord{}
- for i, r := range list {
- if timestamp.Before(r.time) {
- // Doesn't meet the time requirement; keep the record.
- newList = append(newList, list[i])
- continue
- }
- r.ch <- c.get(id)
- close(r.ch)
- }
- if len(newList) == 0 {
- delete(c.subscribers, id)
- } else {
- c.subscribers[id] = newList
- }
-}
-
-func (c *cache) subscribe(id types.UID, timestamp time.Time) chan *data {
- ch := make(chan *data, 1)
- c.lock.Lock()
- defer c.lock.Unlock()
- d := c.getIfNewerThan(id, timestamp)
- if d != nil {
- // If the cache entry is ready, send the data and return immediately.
- ch <- d
- return ch
- }
- // Add the subscription record.
- c.subscribers[id] = append(c.subscribers[id], &subRecord{time: timestamp, ch: ch})
- return ch
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_gc.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_gc.go
deleted file mode 100644
index 72fa4bd72..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_gc.go
+++ /dev/null
@@ -1,87 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "fmt"
- "time"
-
- "github.com/golang/glog"
-)
-
-// Specified a policy for garbage collecting containers.
-type ContainerGCPolicy struct {
- // Minimum age at which a container can be garbage collected, zero for no limit.
- MinAge time.Duration
-
- // Max number of dead containers any single pod (UID, container name) pair is
- // allowed to have, less than zero for no limit.
- MaxPerPodContainer int
-
- // Max number of total dead containers, less than zero for no limit.
- MaxContainers int
-}
-
-// Manages garbage collection of dead containers.
-//
-// Implementation is thread-compatible.
-type ContainerGC interface {
- // Garbage collect containers.
- GarbageCollect() error
- // Deletes all unused containers, including containers belonging to pods that are terminated but not deleted
- DeleteAllUnusedContainers() error
-}
-
-// SourcesReadyProvider knows how to determine if configuration sources are ready
-type SourcesReadyProvider interface {
- // AllReady returns true if the currently configured sources have all been seen.
- AllReady() bool
-}
-
-// TODO(vmarmol): Preferentially remove pod infra containers.
-type realContainerGC struct {
- // Container runtime
- runtime Runtime
-
- // Policy for garbage collection.
- policy ContainerGCPolicy
-
- // sourcesReadyProvider provides the readiness of kubelet configuration sources.
- sourcesReadyProvider SourcesReadyProvider
-}
-
-// New ContainerGC instance with the specified policy.
-func NewContainerGC(runtime Runtime, policy ContainerGCPolicy, sourcesReadyProvider SourcesReadyProvider) (ContainerGC, error) {
- if policy.MinAge < 0 {
- return nil, fmt.Errorf("invalid minimum garbage collection age: %v", policy.MinAge)
- }
-
- return &realContainerGC{
- runtime: runtime,
- policy: policy,
- sourcesReadyProvider: sourcesReadyProvider,
- }, nil
-}
-
-func (cgc *realContainerGC) GarbageCollect() error {
- return cgc.runtime.GarbageCollect(cgc.policy, cgc.sourcesReadyProvider.AllReady(), false)
-}
-
-func (cgc *realContainerGC) DeleteAllUnusedContainers() error {
- glog.Infof("attempting to delete unused containers")
- return cgc.runtime.GarbageCollect(cgc.policy, cgc.sourcesReadyProvider.AllReady(), true)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_reference_manager.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_reference_manager.go
deleted file mode 100644
index d41d05a93..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/container_reference_manager.go
+++ /dev/null
@@ -1,60 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "sync"
-
- "k8s.io/api/core/v1"
-)
-
-// RefManager manages the references for the containers.
-// The references are used for reporting events such as creation,
-// failure, etc. This manager is thread-safe, no locks are necessary
-// for the caller.
-type RefManager struct {
- sync.RWMutex
- containerIDToRef map[ContainerID]*v1.ObjectReference
-}
-
-// NewRefManager creates and returns a container reference manager
-// with empty contents.
-func NewRefManager() *RefManager {
- return &RefManager{containerIDToRef: make(map[ContainerID]*v1.ObjectReference)}
-}
-
-// SetRef stores a reference to a pod's container, associating it with the given container ID.
-func (c *RefManager) SetRef(id ContainerID, ref *v1.ObjectReference) {
- c.Lock()
- defer c.Unlock()
- c.containerIDToRef[id] = ref
-}
-
-// ClearRef forgets the given container id and its associated container reference.
-func (c *RefManager) ClearRef(id ContainerID) {
- c.Lock()
- defer c.Unlock()
- delete(c.containerIDToRef, id)
-}
-
-// GetRef returns the container reference of the given ID, or (nil, false) if none is stored.
-func (c *RefManager) GetRef(id ContainerID) (ref *v1.ObjectReference, ok bool) {
- c.RLock()
- defer c.RUnlock()
- ref, ok = c.containerIDToRef[id]
- return ref, ok
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/helpers.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/helpers.go
deleted file mode 100644
index 2320a192b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/helpers.go
+++ /dev/null
@@ -1,359 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "bytes"
- "fmt"
- "hash/adler32"
- "hash/fnv"
- "strings"
- "time"
-
- "github.com/golang/glog"
-
- "k8s.io/api/core/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/types"
- "k8s.io/client-go/tools/record"
- runtimeapi "k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2"
- "k8s.io/kubernetes/pkg/kubelet/util/format"
- "k8s.io/kubernetes/pkg/kubelet/util/ioutils"
- hashutil "k8s.io/kubernetes/pkg/util/hash"
- "k8s.io/kubernetes/third_party/forked/golang/expansion"
-)
-
-// HandlerRunner runs a lifecycle handler for a container.
-type HandlerRunner interface {
- Run(containerID ContainerID, pod *v1.Pod, container *v1.Container, handler *v1.Handler) (string, error)
-}
-
-// RuntimeHelper wraps kubelet to make container runtime
-// able to get necessary informations like the RunContainerOptions, DNS settings, Host IP.
-type RuntimeHelper interface {
- GenerateRunContainerOptions(pod *v1.Pod, container *v1.Container, podIP string) (contOpts *RunContainerOptions, err error)
- GetPodDNS(pod *v1.Pod) (dnsConfig *runtimeapi.DNSConfig, err error)
- // GetPodCgroupParent returns the CgroupName identifier, and its literal cgroupfs form on the host
- // of a pod.
- GetPodCgroupParent(pod *v1.Pod) string
- GetPodDir(podUID types.UID) string
- GeneratePodHostNameAndDomain(pod *v1.Pod) (hostname string, hostDomain string, err error)
- // GetExtraSupplementalGroupsForPod returns a list of the extra
- // supplemental groups for the Pod. These extra supplemental groups come
- // from annotations on persistent volumes that the pod depends on.
- GetExtraSupplementalGroupsForPod(pod *v1.Pod) []int64
-}
-
-// ShouldContainerBeRestarted checks whether a container needs to be restarted.
-// TODO(yifan): Think about how to refactor this.
-func ShouldContainerBeRestarted(container *v1.Container, pod *v1.Pod, podStatus *PodStatus) bool {
- // Get latest container status.
- status := podStatus.FindContainerStatusByName(container.Name)
- // If the container was never started before, we should start it.
- // NOTE(random-liu): If all historical containers were GC'd, we'll also return true here.
- if status == nil {
- return true
- }
- // Check whether container is running
- if status.State == ContainerStateRunning {
- return false
- }
- // Always restart container in the unknown, or in the created state.
- if status.State == ContainerStateUnknown || status.State == ContainerStateCreated {
- return true
- }
- // Check RestartPolicy for dead container
- if pod.Spec.RestartPolicy == v1.RestartPolicyNever {
- glog.V(4).Infof("Already ran container %q of pod %q, do nothing", container.Name, format.Pod(pod))
- return false
- }
- if pod.Spec.RestartPolicy == v1.RestartPolicyOnFailure {
- // Check the exit code.
- if status.ExitCode == 0 {
- glog.V(4).Infof("Already successfully ran container %q of pod %q, do nothing", container.Name, format.Pod(pod))
- return false
- }
- }
- return true
-}
-
-// HashContainer returns the hash of the container. It is used to compare
-// the running container with its desired spec.
-func HashContainer(container *v1.Container) uint64 {
- hash := fnv.New32a()
- hashutil.DeepHashObject(hash, *container)
- return uint64(hash.Sum32())
-}
-
-// HashContainerLegacy returns the hash of the container. It is used to compare
-// the running container with its desired spec.
-// This is used by rktnetes and dockershim (for handling <=1.5 containers).
-// TODO: Remove this function when kubernetes version is >=1.8 AND rktnetes
-// update its hash function.
-func HashContainerLegacy(container *v1.Container) uint64 {
- hash := adler32.New()
- hashutil.DeepHashObject(hash, *container)
- return uint64(hash.Sum32())
-}
-
-// EnvVarsToMap constructs a map of environment name to value from a slice
-// of env vars.
-func EnvVarsToMap(envs []EnvVar) map[string]string {
- result := map[string]string{}
- for _, env := range envs {
- result[env.Name] = env.Value
- }
- return result
-}
-
-// V1EnvVarsToMap constructs a map of environment name to value from a slice
-// of env vars.
-func V1EnvVarsToMap(envs []v1.EnvVar) map[string]string {
- result := map[string]string{}
- for _, env := range envs {
- result[env.Name] = env.Value
- }
-
- return result
-}
-
-// ExpandContainerCommandOnlyStatic substitutes only static environment variable values from the
-// container environment definitions. This does *not* include valueFrom substitutions.
-// TODO: callers should use ExpandContainerCommandAndArgs with a fully resolved list of environment.
-func ExpandContainerCommandOnlyStatic(containerCommand []string, envs []v1.EnvVar) (command []string) {
- mapping := expansion.MappingFuncFor(V1EnvVarsToMap(envs))
- if len(containerCommand) != 0 {
- for _, cmd := range containerCommand {
- command = append(command, expansion.Expand(cmd, mapping))
- }
- }
- return command
-}
-
-func ExpandContainerCommandAndArgs(container *v1.Container, envs []EnvVar) (command []string, args []string) {
- mapping := expansion.MappingFuncFor(EnvVarsToMap(envs))
-
- if len(container.Command) != 0 {
- for _, cmd := range container.Command {
- command = append(command, expansion.Expand(cmd, mapping))
- }
- }
-
- if len(container.Args) != 0 {
- for _, arg := range container.Args {
- args = append(args, expansion.Expand(arg, mapping))
- }
- }
-
- return command, args
-}
-
-// Create an event recorder to record object's event except implicitly required container's, like infra container.
-func FilterEventRecorder(recorder record.EventRecorder) record.EventRecorder {
- return &innerEventRecorder{
- recorder: recorder,
- }
-}
-
-type innerEventRecorder struct {
- recorder record.EventRecorder
-}
-
-func (irecorder *innerEventRecorder) shouldRecordEvent(object runtime.Object) (*v1.ObjectReference, bool) {
- if object == nil {
- return nil, false
- }
- if ref, ok := object.(*v1.ObjectReference); ok {
- if !strings.HasPrefix(ref.FieldPath, ImplicitContainerPrefix) {
- return ref, true
- }
- }
- return nil, false
-}
-
-func (irecorder *innerEventRecorder) Event(object runtime.Object, eventtype, reason, message string) {
- if ref, ok := irecorder.shouldRecordEvent(object); ok {
- irecorder.recorder.Event(ref, eventtype, reason, message)
- }
-}
-
-func (irecorder *innerEventRecorder) Eventf(object runtime.Object, eventtype, reason, messageFmt string, args ...interface{}) {
- if ref, ok := irecorder.shouldRecordEvent(object); ok {
- irecorder.recorder.Eventf(ref, eventtype, reason, messageFmt, args...)
- }
-
-}
-
-func (irecorder *innerEventRecorder) PastEventf(object runtime.Object, timestamp metav1.Time, eventtype, reason, messageFmt string, args ...interface{}) {
- if ref, ok := irecorder.shouldRecordEvent(object); ok {
- irecorder.recorder.PastEventf(ref, timestamp, eventtype, reason, messageFmt, args...)
- }
-}
-
-// Pod must not be nil.
-func IsHostNetworkPod(pod *v1.Pod) bool {
- return pod.Spec.HostNetwork
-}
-
-// TODO(random-liu): Convert PodStatus to running Pod, should be deprecated soon
-func ConvertPodStatusToRunningPod(runtimeName string, podStatus *PodStatus) Pod {
- runningPod := Pod{
- ID: podStatus.ID,
- Name: podStatus.Name,
- Namespace: podStatus.Namespace,
- }
- for _, containerStatus := range podStatus.ContainerStatuses {
- if containerStatus.State != ContainerStateRunning {
- continue
- }
- container := &Container{
- ID: containerStatus.ID,
- Name: containerStatus.Name,
- Image: containerStatus.Image,
- ImageID: containerStatus.ImageID,
- Hash: containerStatus.Hash,
- State: containerStatus.State,
- }
- runningPod.Containers = append(runningPod.Containers, container)
- }
-
- // Populate sandboxes in kubecontainer.Pod
- for _, sandbox := range podStatus.SandboxStatuses {
- runningPod.Sandboxes = append(runningPod.Sandboxes, &Container{
- ID: ContainerID{Type: runtimeName, ID: sandbox.Id},
- State: SandboxToContainerState(sandbox.State),
- })
- }
- return runningPod
-}
-
-// SandboxToContainerState converts runtimeapi.PodSandboxState to
-// kubecontainer.ContainerState.
-// This is only needed because we need to return sandboxes as if they were
-// kubecontainer.Containers to avoid substantial changes to PLEG.
-// TODO: Remove this once it becomes obsolete.
-func SandboxToContainerState(state runtimeapi.PodSandboxState) ContainerState {
- switch state {
- case runtimeapi.PodSandboxState_SANDBOX_READY:
- return ContainerStateRunning
- case runtimeapi.PodSandboxState_SANDBOX_NOTREADY:
- return ContainerStateExited
- }
- return ContainerStateUnknown
-}
-
-// FormatPod returns a string representing a pod in a human readable format,
-// with pod UID as part of the string.
-func FormatPod(pod *Pod) string {
- // Use underscore as the delimiter because it is not allowed in pod name
- // (DNS subdomain format), while allowed in the container name format.
- return fmt.Sprintf("%s_%s(%s)", pod.Name, pod.Namespace, pod.ID)
-}
-
-type containerCommandRunnerWrapper struct {
- DirectStreamingRuntime
-}
-
-var _ ContainerCommandRunner = &containerCommandRunnerWrapper{}
-
-func DirectStreamingRunner(runtime DirectStreamingRuntime) ContainerCommandRunner {
- return &containerCommandRunnerWrapper{runtime}
-}
-
-func (r *containerCommandRunnerWrapper) RunInContainer(id ContainerID, cmd []string, timeout time.Duration) ([]byte, error) {
- var buffer bytes.Buffer
- output := ioutils.WriteCloserWrapper(&buffer)
- err := r.ExecInContainer(id, cmd, nil, output, output, false, nil, timeout)
- // Even if err is non-nil, there still may be output (e.g. the exec wrote to stdout or stderr but
- // the command returned a nonzero exit code). Therefore, always return the output along with the
- // error.
- return buffer.Bytes(), err
-}
-
-// GetContainerSpec gets the container spec by containerName.
-func GetContainerSpec(pod *v1.Pod, containerName string) *v1.Container {
- for i, c := range pod.Spec.Containers {
- if containerName == c.Name {
- return &pod.Spec.Containers[i]
- }
- }
- for i, c := range pod.Spec.InitContainers {
- if containerName == c.Name {
- return &pod.Spec.InitContainers[i]
- }
- }
- return nil
-}
-
-// HasPrivilegedContainer returns true if any of the containers in the pod are privileged.
-func HasPrivilegedContainer(pod *v1.Pod) bool {
- for _, c := range append(pod.Spec.Containers, pod.Spec.InitContainers...) {
- if c.SecurityContext != nil &&
- c.SecurityContext.Privileged != nil &&
- *c.SecurityContext.Privileged {
- return true
- }
- }
- return false
-}
-
-// MakeCapabilities creates string slices from Capability slices
-func MakeCapabilities(capAdd []v1.Capability, capDrop []v1.Capability) ([]string, []string) {
- var (
- addCaps []string
- dropCaps []string
- )
- for _, cap := range capAdd {
- addCaps = append(addCaps, string(cap))
- }
- for _, cap := range capDrop {
- dropCaps = append(dropCaps, string(cap))
- }
- return addCaps, dropCaps
-}
-
-// MakePortMappings creates internal port mapping from api port mapping.
-func MakePortMappings(container *v1.Container) (ports []PortMapping) {
- names := make(map[string]struct{})
- for _, p := range container.Ports {
- pm := PortMapping{
- HostPort: int(p.HostPort),
- ContainerPort: int(p.ContainerPort),
- Protocol: p.Protocol,
- HostIP: p.HostIP,
- }
-
- // We need to create some default port name if it's not specified, since
- // this is necessary for rkt.
- // http://issue.k8s.io/7710
- if p.Name == "" {
- pm.Name = fmt.Sprintf("%s-%s:%d", container.Name, p.Protocol, p.ContainerPort)
- } else {
- pm.Name = fmt.Sprintf("%s-%s", container.Name, p.Name)
- }
-
- // Protect against exposing the same protocol-port more than once in a container.
- if _, ok := names[pm.Name]; ok {
- glog.Warningf("Port name conflicted, %q is defined more than once", pm.Name)
- continue
- }
- ports = append(ports, pm)
- names[pm.Name] = struct{}{}
- }
- return
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/os.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/os.go
deleted file mode 100644
index 6126063b3..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/os.go
+++ /dev/null
@@ -1,107 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "io/ioutil"
- "os"
- "path/filepath"
- "time"
-)
-
-// OSInterface collects system level operations that need to be mocked out
-// during tests.
-type OSInterface interface {
- MkdirAll(path string, perm os.FileMode) error
- Symlink(oldname string, newname string) error
- Stat(path string) (os.FileInfo, error)
- Remove(path string) error
- RemoveAll(path string) error
- Create(path string) (*os.File, error)
- Chmod(path string, perm os.FileMode) error
- Hostname() (name string, err error)
- Chtimes(path string, atime time.Time, mtime time.Time) error
- Pipe() (r *os.File, w *os.File, err error)
- ReadDir(dirname string) ([]os.FileInfo, error)
- Glob(pattern string) ([]string, error)
-}
-
-// RealOS is used to dispatch the real system level operations.
-type RealOS struct{}
-
-// MkDir will will call os.Mkdir to create a directory.
-func (RealOS) MkdirAll(path string, perm os.FileMode) error {
- return os.MkdirAll(path, perm)
-}
-
-// Symlink will call os.Symlink to create a symbolic link.
-func (RealOS) Symlink(oldname string, newname string) error {
- return os.Symlink(oldname, newname)
-}
-
-// Stat will call os.Stat to get the FileInfo for a given path
-func (RealOS) Stat(path string) (os.FileInfo, error) {
- return os.Stat(path)
-}
-
-// Remove will call os.Remove to remove the path.
-func (RealOS) Remove(path string) error {
- return os.Remove(path)
-}
-
-// RemoveAll will call os.RemoveAll to remove the path and its children.
-func (RealOS) RemoveAll(path string) error {
- return os.RemoveAll(path)
-}
-
-// Create will call os.Create to create and return a file
-// at path.
-func (RealOS) Create(path string) (*os.File, error) {
- return os.Create(path)
-}
-
-// Chmod will change the permissions on the specified path or return
-// an error.
-func (RealOS) Chmod(path string, perm os.FileMode) error {
- return os.Chmod(path, perm)
-}
-
-// Hostname will call os.Hostname to return the hostname.
-func (RealOS) Hostname() (name string, err error) {
- return os.Hostname()
-}
-
-// Chtimes will call os.Chtimes to change the atime and mtime of the path
-func (RealOS) Chtimes(path string, atime time.Time, mtime time.Time) error {
- return os.Chtimes(path, atime, mtime)
-}
-
-// Pipe will call os.Pipe to return a connected pair of pipe.
-func (RealOS) Pipe() (r *os.File, w *os.File, err error) {
- return os.Pipe()
-}
-
-// ReadDir will call ioutil.ReadDir to return the files under the directory.
-func (RealOS) ReadDir(dirname string) ([]os.FileInfo, error) {
- return ioutil.ReadDir(dirname)
-}
-
-// Glob will call filepath.Glob to return the names of all files matching
-// pattern.
-func (RealOS) Glob(pattern string) ([]string, error) {
- return filepath.Glob(pattern)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_linux.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_linux.go
deleted file mode 100644
index 40906ce99..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_linux.go
+++ /dev/null
@@ -1,30 +0,0 @@
-// +build linux
-
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "os"
- "os/exec"
-
- "github.com/kr/pty"
-)
-
-func StartPty(c *exec.Cmd) (*os.File, error) {
- return pty.Start(c)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_unsupported.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_unsupported.go
deleted file mode 100644
index 24ea2f787..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/pty_unsupported.go
+++ /dev/null
@@ -1,28 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "os"
- "os/exec"
-)
-
-func StartPty(c *exec.Cmd) (pty *os.File, err error) {
- return nil, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/ref.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/ref.go
deleted file mode 100644
index f61c0fc4a..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/ref.go
+++ /dev/null
@@ -1,73 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "fmt"
-
- "k8s.io/api/core/v1"
- ref "k8s.io/client-go/tools/reference"
- "k8s.io/kubernetes/pkg/api/legacyscheme"
-)
-
-var ImplicitContainerPrefix string = "implicitly required container "
-
-// GenerateContainerRef returns an *v1.ObjectReference which references the given container
-// within the given pod. Returns an error if the reference can't be constructed or the
-// container doesn't actually belong to the pod.
-//
-// This function will return an error if the provided Pod does not have a selfLink,
-// but we expect selfLink to be populated at all call sites for the function.
-func GenerateContainerRef(pod *v1.Pod, container *v1.Container) (*v1.ObjectReference, error) {
- fieldPath, err := fieldPath(pod, container)
- if err != nil {
- // TODO: figure out intelligent way to refer to containers that we implicitly
- // start (like the pod infra container). This is not a good way, ugh.
- fieldPath = ImplicitContainerPrefix + container.Name
- }
- ref, err := ref.GetPartialReference(legacyscheme.Scheme, pod, fieldPath)
- if err != nil {
- return nil, err
- }
- return ref, nil
-}
-
-// fieldPath returns a fieldPath locating container within pod.
-// Returns an error if the container isn't part of the pod.
-func fieldPath(pod *v1.Pod, container *v1.Container) (string, error) {
- for i := range pod.Spec.Containers {
- here := &pod.Spec.Containers[i]
- if here.Name == container.Name {
- if here.Name == "" {
- return fmt.Sprintf("spec.containers[%d]", i), nil
- } else {
- return fmt.Sprintf("spec.containers{%s}", here.Name), nil
- }
- }
- }
- for i := range pod.Spec.InitContainers {
- here := &pod.Spec.InitContainers[i]
- if here.Name == container.Name {
- if here.Name == "" {
- return fmt.Sprintf("spec.initContainers[%d]", i), nil
- } else {
- return fmt.Sprintf("spec.initContainers{%s}", here.Name), nil
- }
- }
- }
- return "", fmt.Errorf("container %#v not found in pod %#v", container, pod)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/resize.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/resize.go
deleted file mode 100644
index d7b75eede..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/resize.go
+++ /dev/null
@@ -1,46 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "k8s.io/apimachinery/pkg/util/runtime"
- "k8s.io/client-go/tools/remotecommand"
-)
-
-// handleResizing spawns a goroutine that processes the resize channel, calling resizeFunc for each
-// remotecommand.TerminalSize received from the channel. The resize channel must be closed elsewhere to stop the
-// goroutine.
-func HandleResizing(resize <-chan remotecommand.TerminalSize, resizeFunc func(size remotecommand.TerminalSize)) {
- if resize == nil {
- return
- }
-
- go func() {
- defer runtime.HandleCrash()
-
- for {
- size, ok := <-resize
- if !ok {
- return
- }
- if size.Height < 1 || size.Width < 1 {
- continue
- }
- resizeFunc(size)
- }
- }()
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime.go
deleted file mode 100644
index 56aeaf813..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime.go
+++ /dev/null
@@ -1,664 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "fmt"
- "io"
- "net/url"
- "reflect"
- "strings"
- "time"
-
- "github.com/golang/glog"
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/types"
- "k8s.io/client-go/tools/remotecommand"
- "k8s.io/client-go/util/flowcontrol"
- runtimeapi "k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2"
- "k8s.io/kubernetes/pkg/volume"
-)
-
-type Version interface {
- // Compare compares two versions of the runtime. On success it returns -1
- // if the version is less than the other, 1 if it is greater than the other,
- // or 0 if they are equal.
- Compare(other string) (int, error)
- // String returns a string that represents the version.
- String() string
-}
-
-// ImageSpec is an internal representation of an image. Currently, it wraps the
-// value of a Container's Image field, but in the future it will include more detailed
-// information about the different image types.
-type ImageSpec struct {
- Image string
-}
-
-// ImageStats contains statistics about all the images currently available.
-type ImageStats struct {
- // Total amount of storage consumed by existing images.
- TotalStorageBytes uint64
-}
-
-// Runtime interface defines the interfaces that should be implemented
-// by a container runtime.
-// Thread safety is required from implementations of this interface.
-type Runtime interface {
- // Type returns the type of the container runtime.
- Type() string
-
- // Version returns the version information of the container runtime.
- Version() (Version, error)
-
- // APIVersion returns the cached API version information of the container
- // runtime. Implementation is expected to update this cache periodically.
- // This may be different from the runtime engine's version.
- // TODO(random-liu): We should fold this into Version()
- APIVersion() (Version, error)
- // Status returns the status of the runtime. An error is returned if the Status
- // function itself fails, nil otherwise.
- Status() (*RuntimeStatus, error)
- // GetPods returns a list of containers grouped by pods. The boolean parameter
- // specifies whether the runtime returns all containers including those already
- // exited and dead containers (used for garbage collection).
- GetPods(all bool) ([]*Pod, error)
- // GarbageCollect removes dead containers using the specified container gc policy
- // If allSourcesReady is not true, it means that kubelet doesn't have the
- // complete list of pods from all avialble sources (e.g., apiserver, http,
- // file). In this case, garbage collector should refrain itself from aggressive
- // behavior such as removing all containers of unrecognized pods (yet).
- // If evictNonDeletedPods is set to true, containers and sandboxes belonging to pods
- // that are terminated, but not deleted will be evicted. Otherwise, only deleted pods will be GC'd.
- // TODO: Revisit this method and make it cleaner.
- GarbageCollect(gcPolicy ContainerGCPolicy, allSourcesReady bool, evictNonDeletedPods bool) error
- // Syncs the running pod into the desired pod.
- SyncPod(pod *v1.Pod, apiPodStatus v1.PodStatus, podStatus *PodStatus, pullSecrets []v1.Secret, backOff *flowcontrol.Backoff) PodSyncResult
- // KillPod kills all the containers of a pod. Pod may be nil, running pod must not be.
- // TODO(random-liu): Return PodSyncResult in KillPod.
- // gracePeriodOverride if specified allows the caller to override the pod default grace period.
- // only hard kill paths are allowed to specify a gracePeriodOverride in the kubelet in order to not corrupt user data.
- // it is useful when doing SIGKILL for hard eviction scenarios, or max grace period during soft eviction scenarios.
- KillPod(pod *v1.Pod, runningPod Pod, gracePeriodOverride *int64) error
- // GetPodStatus retrieves the status of the pod, including the
- // information of all containers in the pod that are visble in Runtime.
- GetPodStatus(uid types.UID, name, namespace string) (*PodStatus, error)
- // Returns the filesystem path of the pod's network namespace; if the
- // runtime does not handle namespace creation itself, or cannot return
- // the network namespace path, it should return an error.
- // TODO: Change ContainerID to a Pod ID since the namespace is shared
- // by all containers in the pod.
- GetNetNS(containerID ContainerID) (string, error)
- // Returns the container ID that represents the Pod, as passed to network
- // plugins. For example, if the runtime uses an infra container, returns
- // the infra container's ContainerID.
- // TODO: Change ContainerID to a Pod ID, see GetNetNS()
- GetPodContainerID(*Pod) (ContainerID, error)
- // TODO(vmarmol): Unify pod and containerID args.
- // GetContainerLogs returns logs of a specific container. By
- // default, it returns a snapshot of the container log. Set 'follow' to true to
- // stream the log. Set 'follow' to false and specify the number of lines (e.g.
- // "100" or "all") to tail the log.
- GetContainerLogs(pod *v1.Pod, containerID ContainerID, logOptions *v1.PodLogOptions, stdout, stderr io.Writer) (err error)
- // Delete a container. If the container is still running, an error is returned.
- DeleteContainer(containerID ContainerID) error
- // ImageService provides methods to image-related methods.
- ImageService
- // UpdatePodCIDR sends a new podCIDR to the runtime.
- // This method just proxies a new runtimeConfig with the updated
- // CIDR value down to the runtime shim.
- UpdatePodCIDR(podCIDR string) error
-}
-
-// DirectStreamingRuntime is the interface implemented by runtimes for which the streaming calls
-// (exec/attach/port-forward) should be served directly by the Kubelet.
-type DirectStreamingRuntime interface {
- // Runs the command in the container of the specified pod. Attaches
- // the processes stdin, stdout, and stderr. Optionally uses a tty.
- ExecInContainer(containerID ContainerID, cmd []string, stdin io.Reader, stdout, stderr io.WriteCloser, tty bool, resize <-chan remotecommand.TerminalSize, timeout time.Duration) error
- // Forward the specified port from the specified pod to the stream.
- PortForward(pod *Pod, port int32, stream io.ReadWriteCloser) error
- // ContainerAttach encapsulates the attaching to containers for testability
- ContainerAttacher
-}
-
-// IndirectStreamingRuntime is the interface implemented by runtimes that handle the serving of the
-// streaming calls (exec/attach/port-forward) themselves. In this case, Kubelet should redirect to
-// the runtime server.
-type IndirectStreamingRuntime interface {
- GetExec(id ContainerID, cmd []string, stdin, stdout, stderr, tty bool) (*url.URL, error)
- GetAttach(id ContainerID, stdin, stdout, stderr, tty bool) (*url.URL, error)
- GetPortForward(podName, podNamespace string, podUID types.UID, ports []int32) (*url.URL, error)
-}
-
-type ImageService interface {
- // PullImage pulls an image from the network to local storage using the supplied
- // secrets if necessary. It returns a reference (digest or ID) to the pulled image.
- PullImage(image ImageSpec, pullSecrets []v1.Secret) (string, error)
- // GetImageRef gets the reference (digest or ID) of the image which has already been in
- // the local storage. It returns ("", nil) if the image isn't in the local storage.
- GetImageRef(image ImageSpec) (string, error)
- // Gets all images currently on the machine.
- ListImages() ([]Image, error)
- // Removes the specified image.
- RemoveImage(image ImageSpec) error
- // Returns Image statistics.
- ImageStats() (*ImageStats, error)
-}
-
-type ContainerAttacher interface {
- AttachContainer(id ContainerID, stdin io.Reader, stdout, stderr io.WriteCloser, tty bool, resize <-chan remotecommand.TerminalSize) (err error)
-}
-
-type ContainerCommandRunner interface {
- // RunInContainer synchronously executes the command in the container, and returns the output.
- // If the command completes with a non-0 exit code, a pkg/util/exec.ExitError will be returned.
- RunInContainer(id ContainerID, cmd []string, timeout time.Duration) ([]byte, error)
-}
-
-// Pod is a group of containers.
-type Pod struct {
- // The ID of the pod, which can be used to retrieve a particular pod
- // from the pod list returned by GetPods().
- ID types.UID
- // The name and namespace of the pod, which is readable by human.
- Name string
- Namespace string
- // List of containers that belongs to this pod. It may contain only
- // running containers, or mixed with dead ones (when GetPods(true)).
- Containers []*Container
- // List of sandboxes associated with this pod. The sandboxes are converted
- // to Container temporariliy to avoid substantial changes to other
- // components. This is only populated by kuberuntime.
- // TODO: use the runtimeApi.PodSandbox type directly.
- Sandboxes []*Container
-}
-
-// PodPair contains both runtime#Pod and api#Pod
-type PodPair struct {
- // APIPod is the v1.Pod
- APIPod *v1.Pod
- // RunningPod is the pod defined in pkg/kubelet/container/runtime#Pod
- RunningPod *Pod
-}
-
-// ContainerID is a type that identifies a container.
-type ContainerID struct {
- // The type of the container runtime. e.g. 'docker', 'rkt'.
- Type string
- // The identification of the container, this is comsumable by
- // the underlying container runtime. (Note that the container
- // runtime interface still takes the whole struct as input).
- ID string
-}
-
-func BuildContainerID(typ, ID string) ContainerID {
- return ContainerID{Type: typ, ID: ID}
-}
-
-// Convenience method for creating a ContainerID from an ID string.
-func ParseContainerID(containerID string) ContainerID {
- var id ContainerID
- if err := id.ParseString(containerID); err != nil {
- glog.Error(err)
- }
- return id
-}
-
-func (c *ContainerID) ParseString(data string) error {
- // Trim the quotes and split the type and ID.
- parts := strings.Split(strings.Trim(data, "\""), "://")
- if len(parts) != 2 {
- return fmt.Errorf("invalid container ID: %q", data)
- }
- c.Type, c.ID = parts[0], parts[1]
- return nil
-}
-
-func (c *ContainerID) String() string {
- return fmt.Sprintf("%s://%s", c.Type, c.ID)
-}
-
-func (c *ContainerID) IsEmpty() bool {
- return *c == ContainerID{}
-}
-
-func (c *ContainerID) MarshalJSON() ([]byte, error) {
- return []byte(fmt.Sprintf("%q", c.String())), nil
-}
-
-func (c *ContainerID) UnmarshalJSON(data []byte) error {
- return c.ParseString(string(data))
-}
-
-// DockerID is an ID of docker container. It is a type to make it clear when we're working with docker container Ids
-type DockerID string
-
-func (id DockerID) ContainerID() ContainerID {
- return ContainerID{
- Type: "docker",
- ID: string(id),
- }
-}
-
-type ContainerState string
-
-const (
- ContainerStateCreated ContainerState = "created"
- ContainerStateRunning ContainerState = "running"
- ContainerStateExited ContainerState = "exited"
- // This unknown encompasses all the states that we currently don't care.
- ContainerStateUnknown ContainerState = "unknown"
-)
-
-type ContainerType string
-
-const (
- ContainerTypeInit ContainerType = "INIT"
- ContainerTypeRegular ContainerType = "REGULAR"
-)
-
-// Container provides the runtime information for a container, such as ID, hash,
-// state of the container.
-type Container struct {
- // The ID of the container, used by the container runtime to identify
- // a container.
- ID ContainerID
- // The name of the container, which should be the same as specified by
- // v1.Container.
- Name string
- // The image name of the container, this also includes the tag of the image,
- // the expected form is "NAME:TAG".
- Image string
- // The id of the image used by the container.
- ImageID string
- // Hash of the container, used for comparison. Optional for containers
- // not managed by kubelet.
- Hash uint64
- // State is the state of the container.
- State ContainerState
-}
-
-// PodStatus represents the status of the pod and its containers.
-// v1.PodStatus can be derived from examining PodStatus and v1.Pod.
-type PodStatus struct {
- // ID of the pod.
- ID types.UID
- // Name of the pod.
- Name string
- // Namspace of the pod.
- Namespace string
- // IP of the pod.
- IP string
- // Status of containers in the pod.
- ContainerStatuses []*ContainerStatus
- // Status of the pod sandbox.
- // Only for kuberuntime now, other runtime may keep it nil.
- SandboxStatuses []*runtimeapi.PodSandboxStatus
-}
-
-// ContainerStatus represents the status of a container.
-type ContainerStatus struct {
- // ID of the container.
- ID ContainerID
- // Name of the container.
- Name string
- // Status of the container.
- State ContainerState
- // Creation time of the container.
- CreatedAt time.Time
- // Start time of the container.
- StartedAt time.Time
- // Finish time of the container.
- FinishedAt time.Time
- // Exit code of the container.
- ExitCode int
- // Name of the image, this also includes the tag of the image,
- // the expected form is "NAME:TAG".
- Image string
- // ID of the image.
- ImageID string
- // Hash of the container, used for comparison.
- Hash uint64
- // Number of times that the container has been restarted.
- RestartCount int
- // A string explains why container is in such a status.
- Reason string
- // Message written by the container before exiting (stored in
- // TerminationMessagePath).
- Message string
-}
-
-// FindContainerStatusByName returns container status in the pod status with the given name.
-// When there are multiple containers' statuses with the same name, the first match will be returned.
-func (podStatus *PodStatus) FindContainerStatusByName(containerName string) *ContainerStatus {
- for _, containerStatus := range podStatus.ContainerStatuses {
- if containerStatus.Name == containerName {
- return containerStatus
- }
- }
- return nil
-}
-
-// Get container status of all the running containers in a pod
-func (podStatus *PodStatus) GetRunningContainerStatuses() []*ContainerStatus {
- runningContainerStatuses := []*ContainerStatus{}
- for _, containerStatus := range podStatus.ContainerStatuses {
- if containerStatus.State == ContainerStateRunning {
- runningContainerStatuses = append(runningContainerStatuses, containerStatus)
- }
- }
- return runningContainerStatuses
-}
-
-// Basic information about a container image.
-type Image struct {
- // ID of the image.
- ID string
- // Other names by which this image is known.
- RepoTags []string
- // Digests by which this image is known.
- RepoDigests []string
- // The size of the image in bytes.
- Size int64
-}
-
-type EnvVar struct {
- Name string
- Value string
-}
-
-type Annotation struct {
- Name string
- Value string
-}
-
-type Mount struct {
- // Name of the volume mount.
- // TODO(yifan): Remove this field, as this is not representing the unique name of the mount,
- // but the volume name only.
- Name string
- // Path of the mount within the container.
- ContainerPath string
- // Path of the mount on the host.
- HostPath string
- // Whether the mount is read-only.
- ReadOnly bool
- // Whether the mount needs SELinux relabeling
- SELinuxRelabel bool
- // Requested propagation mode
- Propagation runtimeapi.MountPropagation
-}
-
-type PortMapping struct {
- // Name of the port mapping
- Name string
- // Protocol of the port mapping.
- Protocol v1.Protocol
- // The port number within the container.
- ContainerPort int
- // The port number on the host.
- HostPort int
- // The host IP.
- HostIP string
-}
-
-type DeviceInfo struct {
- // Path on host for mapping
- PathOnHost string
- // Path in Container to map
- PathInContainer string
- // Cgroup permissions
- Permissions string
-}
-
-// RunContainerOptions specify the options which are necessary for running containers
-type RunContainerOptions struct {
- // The environment variables list.
- Envs []EnvVar
- // The mounts for the containers.
- Mounts []Mount
- // The host devices mapped into the containers.
- Devices []DeviceInfo
- // The port mappings for the containers.
- PortMappings []PortMapping
- // The annotations for the container
- // These annotations are generated by other components (i.e.,
- // not users). Currently, only device plugins populate the annotations.
- Annotations []Annotation
- // If the container has specified the TerminationMessagePath, then
- // this directory will be used to create and mount the log file to
- // container.TerminationMessagePath
- PodContainerDir string
- // The parent cgroup to pass to Docker
- CgroupParent string
- // The type of container rootfs
- ReadOnly bool
- // hostname for pod containers
- Hostname string
- // EnableHostUserNamespace sets userns=host when users request host namespaces (pid, ipc, net),
- // are using non-namespaced capabilities (mknod, sys_time, sys_module), the pod contains a privileged container,
- // or using host path volumes.
- // This should only be enabled when the container runtime is performing user remapping AND if the
- // experimental behavior is desired.
- EnableHostUserNamespace bool
-}
-
-// VolumeInfo contains information about the volume.
-type VolumeInfo struct {
- // Mounter is the volume's mounter
- Mounter volume.Mounter
- // BlockVolumeMapper is the Block volume's mapper
- BlockVolumeMapper volume.BlockVolumeMapper
- // SELinuxLabeled indicates whether this volume has had the
- // pod's SELinux label applied to it or not
- SELinuxLabeled bool
- // Whether the volume permission is set to read-only or not
- // This value is passed from volume.spec
- ReadOnly bool
-}
-
-type VolumeMap map[string]VolumeInfo
-
-// RuntimeConditionType is the types of required runtime conditions.
-type RuntimeConditionType string
-
-const (
- // RuntimeReady means the runtime is up and ready to accept basic containers.
- RuntimeReady RuntimeConditionType = "RuntimeReady"
- // NetworkReady means the runtime network is up and ready to accept containers which require network.
- NetworkReady RuntimeConditionType = "NetworkReady"
-)
-
-// RuntimeStatus contains the status of the runtime.
-type RuntimeStatus struct {
- // Conditions is an array of current observed runtime conditions.
- Conditions []RuntimeCondition
-}
-
-// GetRuntimeCondition gets a specified runtime condition from the runtime status.
-func (r *RuntimeStatus) GetRuntimeCondition(t RuntimeConditionType) *RuntimeCondition {
- for i := range r.Conditions {
- c := &r.Conditions[i]
- if c.Type == t {
- return c
- }
- }
- return nil
-}
-
-// String formats the runtime status into human readable string.
-func (s *RuntimeStatus) String() string {
- var ss []string
- for _, c := range s.Conditions {
- ss = append(ss, c.String())
- }
- return fmt.Sprintf("Runtime Conditions: %s", strings.Join(ss, ", "))
-}
-
-// RuntimeCondition contains condition information for the runtime.
-type RuntimeCondition struct {
- // Type of runtime condition.
- Type RuntimeConditionType
- // Status of the condition, one of true/false.
- Status bool
- // Reason is brief reason for the condition's last transition.
- Reason string
- // Message is human readable message indicating details about last transition.
- Message string
-}
-
-// String formats the runtime condition into human readable string.
-func (c *RuntimeCondition) String() string {
- return fmt.Sprintf("%s=%t reason:%s message:%s", c.Type, c.Status, c.Reason, c.Message)
-}
-
-type Pods []*Pod
-
-// FindPodByID finds and returns a pod in the pod list by UID. It will return an empty pod
-// if not found.
-func (p Pods) FindPodByID(podUID types.UID) Pod {
- for i := range p {
- if p[i].ID == podUID {
- return *p[i]
- }
- }
- return Pod{}
-}
-
-// FindPodByFullName finds and returns a pod in the pod list by the full name.
-// It will return an empty pod if not found.
-func (p Pods) FindPodByFullName(podFullName string) Pod {
- for i := range p {
- if BuildPodFullName(p[i].Name, p[i].Namespace) == podFullName {
- return *p[i]
- }
- }
- return Pod{}
-}
-
-// FindPod combines FindPodByID and FindPodByFullName, it finds and returns a pod in the
-// pod list either by the full name or the pod ID. It will return an empty pod
-// if not found.
-func (p Pods) FindPod(podFullName string, podUID types.UID) Pod {
- if len(podFullName) > 0 {
- return p.FindPodByFullName(podFullName)
- }
- return p.FindPodByID(podUID)
-}
-
-// FindContainerByName returns a container in the pod with the given name.
-// When there are multiple containers with the same name, the first match will
-// be returned.
-func (p *Pod) FindContainerByName(containerName string) *Container {
- for _, c := range p.Containers {
- if c.Name == containerName {
- return c
- }
- }
- return nil
-}
-
-func (p *Pod) FindContainerByID(id ContainerID) *Container {
- for _, c := range p.Containers {
- if c.ID == id {
- return c
- }
- }
- return nil
-}
-
-func (p *Pod) FindSandboxByID(id ContainerID) *Container {
- for _, c := range p.Sandboxes {
- if c.ID == id {
- return c
- }
- }
- return nil
-}
-
-// ToAPIPod converts Pod to v1.Pod. Note that if a field in v1.Pod has no
-// corresponding field in Pod, the field would not be populated.
-func (p *Pod) ToAPIPod() *v1.Pod {
- var pod v1.Pod
- pod.UID = p.ID
- pod.Name = p.Name
- pod.Namespace = p.Namespace
-
- for _, c := range p.Containers {
- var container v1.Container
- container.Name = c.Name
- container.Image = c.Image
- pod.Spec.Containers = append(pod.Spec.Containers, container)
- }
- return &pod
-}
-
-// IsEmpty returns true if the pod is empty.
-func (p *Pod) IsEmpty() bool {
- return reflect.DeepEqual(p, &Pod{})
-}
-
-// GetPodFullName returns a name that uniquely identifies a pod.
-func GetPodFullName(pod *v1.Pod) string {
- // Use underscore as the delimiter because it is not allowed in pod name
- // (DNS subdomain format), while allowed in the container name format.
- return pod.Name + "_" + pod.Namespace
-}
-
-// Build the pod full name from pod name and namespace.
-func BuildPodFullName(name, namespace string) string {
- return name + "_" + namespace
-}
-
-// Parse the pod full name.
-func ParsePodFullName(podFullName string) (string, string, error) {
- parts := strings.Split(podFullName, "_")
- if len(parts) != 2 || parts[0] == "" || parts[1] == "" {
- return "", "", fmt.Errorf("failed to parse the pod full name %q", podFullName)
- }
- return parts[0], parts[1], nil
-}
-
-// Option is a functional option type for Runtime, useful for
-// completely optional settings.
-type Option func(Runtime)
-
-// Sort the container statuses by creation time.
-type SortContainerStatusesByCreationTime []*ContainerStatus
-
-func (s SortContainerStatusesByCreationTime) Len() int { return len(s) }
-func (s SortContainerStatusesByCreationTime) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
-func (s SortContainerStatusesByCreationTime) Less(i, j int) bool {
- return s[i].CreatedAt.Before(s[j].CreatedAt)
-}
-
-const (
- // MaxPodTerminationMessageLogLength is the maximum bytes any one pod may have written
- // as termination message output across all containers. Containers will be evenly truncated
- // until output is below this limit.
- MaxPodTerminationMessageLogLength = 1024 * 12
- // MaxContainerTerminationMessageLength is the upper bound any one container may write to
- // its termination message path. Contents above this length will be truncated.
- MaxContainerTerminationMessageLength = 1024 * 4
- // MaxContainerTerminationMessageLogLength is the maximum bytes any one container will
- // have written to its termination message when the message is read from the logs.
- MaxContainerTerminationMessageLogLength = 1024 * 2
- // MaxContainerTerminationMessageLogLines is the maximum number of previous lines of
- // log output that the termination message can contain.
- MaxContainerTerminationMessageLogLines = 80
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache.go
deleted file mode 100644
index d15852f88..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache.go
+++ /dev/null
@@ -1,96 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "sync"
- "time"
-)
-
-var (
- // TODO(yifan): Maybe set the them as parameters for NewCache().
- defaultCachePeriod = time.Second * 2
-)
-
-type RuntimeCache interface {
- GetPods() ([]*Pod, error)
- ForceUpdateIfOlder(time.Time) error
-}
-
-type podsGetter interface {
- GetPods(bool) ([]*Pod, error)
-}
-
-// NewRuntimeCache creates a container runtime cache.
-func NewRuntimeCache(getter podsGetter) (RuntimeCache, error) {
- return &runtimeCache{
- getter: getter,
- }, nil
-}
-
-// runtimeCache caches a list of pods. It records a timestamp (cacheTime) right
-// before updating the pods, so the timestamp is at most as new as the pods
-// (and can be slightly older). The timestamp always moves forward. Callers are
-// expected not to modify the pods returned from GetPods.
-type runtimeCache struct {
- sync.Mutex
- // The underlying container runtime used to update the cache.
- getter podsGetter
- // Last time when cache was updated.
- cacheTime time.Time
- // The content of the cache.
- pods []*Pod
-}
-
-// GetPods returns the cached pods if they are not outdated; otherwise, it
-// retrieves the latest pods and return them.
-func (r *runtimeCache) GetPods() ([]*Pod, error) {
- r.Lock()
- defer r.Unlock()
- if time.Since(r.cacheTime) > defaultCachePeriod {
- if err := r.updateCache(); err != nil {
- return nil, err
- }
- }
- return r.pods, nil
-}
-
-func (r *runtimeCache) ForceUpdateIfOlder(minExpectedCacheTime time.Time) error {
- r.Lock()
- defer r.Unlock()
- if r.cacheTime.Before(minExpectedCacheTime) {
- return r.updateCache()
- }
- return nil
-}
-
-func (r *runtimeCache) updateCache() error {
- pods, timestamp, err := r.getPodsWithTimestamp()
- if err != nil {
- return err
- }
- r.pods, r.cacheTime = pods, timestamp
- return nil
-}
-
-// getPodsWithTimestamp records a timestamp and retrieves pods from the getter.
-func (r *runtimeCache) getPodsWithTimestamp() ([]*Pod, time.Time, error) {
- // Always record the timestamp before getting the pods to avoid stale pods.
- timestamp := time.Now()
- pods, err := r.getter.GetPods(false)
- return pods, timestamp, err
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache_fake.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache_fake.go
deleted file mode 100644
index 59a6288d5..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/runtime_cache_fake.go
+++ /dev/null
@@ -1,45 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-// TestRunTimeCache embeds runtimeCache with some additional methods for testing.
-// It must be declared in the container package to have visibility to runtimeCache.
-// It cannot be in a "..._test.go" file in order for runtime_cache_test.go to have cross-package visibility to it.
-// (cross-package declarations in test files cannot be used from dot imports if this package is vendored)
-type TestRuntimeCache struct {
- runtimeCache
-}
-
-func (r *TestRuntimeCache) UpdateCacheWithLock() error {
- r.Lock()
- defer r.Unlock()
- return r.updateCache()
-}
-
-func (r *TestRuntimeCache) GetCachedPods() []*Pod {
- r.Lock()
- defer r.Unlock()
- return r.pods
-}
-
-func NewTestRuntimeCache(getter podsGetter) *TestRuntimeCache {
- return &TestRuntimeCache{
- runtimeCache: runtimeCache{
- getter: getter,
- },
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/container/sync_result.go b/vendor/k8s.io/kubernetes/pkg/kubelet/container/sync_result.go
deleted file mode 100644
index 0d4563303..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/container/sync_result.go
+++ /dev/null
@@ -1,128 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package container
-
-import (
- "errors"
- "fmt"
-
- utilerrors "k8s.io/apimachinery/pkg/util/errors"
-)
-
-// TODO(random-liu): We need to better organize runtime errors for introspection.
-
-// Container Terminated and Kubelet is backing off the restart
-var ErrCrashLoopBackOff = errors.New("CrashLoopBackOff")
-
-var (
- // ErrContainerNotFound returned when a container in the given pod with the
- // given container name was not found, amongst those managed by the kubelet.
- ErrContainerNotFound = errors.New("no matching container")
-)
-
-var (
- ErrRunContainer = errors.New("RunContainerError")
- ErrKillContainer = errors.New("KillContainerError")
- ErrVerifyNonRoot = errors.New("VerifyNonRootError")
- ErrRunInitContainer = errors.New("RunInitContainerError")
- ErrCreatePodSandbox = errors.New("CreatePodSandboxError")
- ErrConfigPodSandbox = errors.New("ConfigPodSandboxError")
- ErrKillPodSandbox = errors.New("KillPodSandboxError")
-)
-
-var (
- ErrSetupNetwork = errors.New("SetupNetworkError")
- ErrTeardownNetwork = errors.New("TeardownNetworkError")
-)
-
-// SyncAction indicates different kind of actions in SyncPod() and KillPod(). Now there are only actions
-// about start/kill container and setup/teardown network.
-type SyncAction string
-
-const (
- StartContainer SyncAction = "StartContainer"
- KillContainer SyncAction = "KillContainer"
- SetupNetwork SyncAction = "SetupNetwork"
- TeardownNetwork SyncAction = "TeardownNetwork"
- InitContainer SyncAction = "InitContainer"
- CreatePodSandbox SyncAction = "CreatePodSandbox"
- ConfigPodSandbox SyncAction = "ConfigPodSandbox"
- KillPodSandbox SyncAction = "KillPodSandbox"
-)
-
-// SyncResult is the result of sync action.
-type SyncResult struct {
- // The associated action of the result
- Action SyncAction
- // The target of the action, now the target can only be:
- // * Container: Target should be container name
- // * Network: Target is useless now, we just set it as pod full name now
- Target interface{}
- // Brief error reason
- Error error
- // Human readable error reason
- Message string
-}
-
-// NewSyncResult generates new SyncResult with specific Action and Target
-func NewSyncResult(action SyncAction, target interface{}) *SyncResult {
- return &SyncResult{Action: action, Target: target}
-}
-
-// Fail fails the SyncResult with specific error and message
-func (r *SyncResult) Fail(err error, msg string) {
- r.Error, r.Message = err, msg
-}
-
-// PodSyncResult is the summary result of SyncPod() and KillPod()
-type PodSyncResult struct {
- // Result of different sync actions
- SyncResults []*SyncResult
- // Error encountered in SyncPod() and KillPod() that is not already included in SyncResults
- SyncError error
-}
-
-// AddSyncResult adds multiple SyncResult to current PodSyncResult
-func (p *PodSyncResult) AddSyncResult(result ...*SyncResult) {
- p.SyncResults = append(p.SyncResults, result...)
-}
-
-// AddPodSyncResult merges a PodSyncResult to current one
-func (p *PodSyncResult) AddPodSyncResult(result PodSyncResult) {
- p.AddSyncResult(result.SyncResults...)
- p.SyncError = result.SyncError
-}
-
-// Fail fails the PodSyncResult with an error occurred in SyncPod() and KillPod() itself
-func (p *PodSyncResult) Fail(err error) {
- p.SyncError = err
-}
-
-// Error returns an error summarizing all the errors in PodSyncResult
-func (p *PodSyncResult) Error() error {
- errlist := []error{}
- if p.SyncError != nil {
- errlist = append(errlist, fmt.Errorf("failed to SyncPod: %v\n", p.SyncError))
- }
- for _, result := range p.SyncResults {
- if result.Error != nil {
- errlist = append(errlist, fmt.Errorf("failed to %q for %q with %v: %q\n", result.Action, result.Target,
- result.Error, result.Message))
- }
- }
- return utilerrors.NewAggregate(errlist)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/types/constants.go b/vendor/k8s.io/kubernetes/pkg/kubelet/types/constants.go
deleted file mode 100644
index 2c83908a7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/types/constants.go
+++ /dev/null
@@ -1,33 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package types
-
-const (
- // system default DNS resolver configuration
- ResolvConfDefault = "/etc/resolv.conf"
-
- // different container runtimes
- DockerContainerRuntime = "docker"
- RktContainerRuntime = "rkt"
- RemoteContainerRuntime = "remote"
-
- // User visible keys for managing node allocatable enforcement on the node.
- NodeAllocatableEnforcementKey = "pods"
- SystemReservedEnforcementKey = "system-reserved"
- KubeReservedEnforcementKey = "kube-reserved"
- NodeAllocatableNoneKey = "none"
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/types/doc.go b/vendor/k8s.io/kubernetes/pkg/kubelet/types/doc.go
deleted file mode 100644
index 88e345636..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/types/doc.go
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Common types in the Kubelet.
-package types // import "k8s.io/kubernetes/pkg/kubelet/types"
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/types/labels.go b/vendor/k8s.io/kubernetes/pkg/kubelet/types/labels.go
deleted file mode 100644
index 67c84f6d6..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/types/labels.go
+++ /dev/null
@@ -1,41 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package types
-
-const (
- KubernetesPodNameLabel = "io.kubernetes.pod.name"
- KubernetesPodNamespaceLabel = "io.kubernetes.pod.namespace"
- KubernetesPodUIDLabel = "io.kubernetes.pod.uid"
- KubernetesContainerNameLabel = "io.kubernetes.container.name"
- KubernetesContainerTypeLabel = "io.kubernetes.container.type"
-)
-
-func GetContainerName(labels map[string]string) string {
- return labels[KubernetesContainerNameLabel]
-}
-
-func GetPodName(labels map[string]string) string {
- return labels[KubernetesPodNameLabel]
-}
-
-func GetPodUID(labels map[string]string) string {
- return labels[KubernetesPodUIDLabel]
-}
-
-func GetPodNamespace(labels map[string]string) string {
- return labels[KubernetesPodNamespaceLabel]
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/types/pod_update.go b/vendor/k8s.io/kubernetes/pkg/kubelet/types/pod_update.go
deleted file mode 100644
index 62116985f..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/types/pod_update.go
+++ /dev/null
@@ -1,175 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package types
-
-import (
- "fmt"
-
- "k8s.io/api/core/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- kubeapi "k8s.io/kubernetes/pkg/apis/core"
- schedulerapi "k8s.io/kubernetes/pkg/scheduler/api"
-)
-
-const (
- ConfigSourceAnnotationKey = "kubernetes.io/config.source"
- ConfigMirrorAnnotationKey = v1.MirrorPodAnnotationKey
- ConfigFirstSeenAnnotationKey = "kubernetes.io/config.seen"
- ConfigHashAnnotationKey = "kubernetes.io/config.hash"
- CriticalPodAnnotationKey = "scheduler.alpha.kubernetes.io/critical-pod"
-)
-
-// PodOperation defines what changes will be made on a pod configuration.
-type PodOperation int
-
-const (
- // This is the current pod configuration
- SET PodOperation = iota
- // Pods with the given ids are new to this source
- ADD
- // Pods with the given ids are gracefully deleted from this source
- DELETE
- // Pods with the given ids have been removed from this source
- REMOVE
- // Pods with the given ids have been updated in this source
- UPDATE
- // Pods with the given ids have unexpected status in this source,
- // kubelet should reconcile status with this source
- RECONCILE
- // Pods with the given ids have been restored from a checkpoint.
- RESTORE
-
- // These constants identify the sources of pods
- // Updates from a file
- FileSource = "file"
- // Updates from querying a web page
- HTTPSource = "http"
- // Updates from Kubernetes API Server
- ApiserverSource = "api"
- // Updates from all sources
- AllSource = "*"
-
- NamespaceDefault = metav1.NamespaceDefault
-)
-
-// PodUpdate defines an operation sent on the channel. You can add or remove single services by
-// sending an array of size one and Op == ADD|REMOVE (with REMOVE, only the ID is required).
-// For setting the state of the system to a given state for this source configuration, set
-// Pods as desired and Op to SET, which will reset the system state to that specified in this
-// operation for this source channel. To remove all pods, set Pods to empty object and Op to SET.
-//
-// Additionally, Pods should never be nil - it should always point to an empty slice. While
-// functionally similar, this helps our unit tests properly check that the correct PodUpdates
-// are generated.
-type PodUpdate struct {
- Pods []*v1.Pod
- Op PodOperation
- Source string
-}
-
-// Gets all validated sources from the specified sources.
-func GetValidatedSources(sources []string) ([]string, error) {
- validated := make([]string, 0, len(sources))
- for _, source := range sources {
- switch source {
- case AllSource:
- return []string{FileSource, HTTPSource, ApiserverSource}, nil
- case FileSource, HTTPSource, ApiserverSource:
- validated = append(validated, source)
- break
- case "":
- break
- default:
- return []string{}, fmt.Errorf("unknown pod source %q", source)
- }
- }
- return validated, nil
-}
-
-// GetPodSource returns the source of the pod based on the annotation.
-func GetPodSource(pod *v1.Pod) (string, error) {
- if pod.Annotations != nil {
- if source, ok := pod.Annotations[ConfigSourceAnnotationKey]; ok {
- return source, nil
- }
- }
- return "", fmt.Errorf("cannot get source of pod %q", pod.UID)
-}
-
-// SyncPodType classifies pod updates, eg: create, update.
-type SyncPodType int
-
-const (
- // SyncPodSync is when the pod is synced to ensure desired state
- SyncPodSync SyncPodType = iota
- // SyncPodUpdate is when the pod is updated from source
- SyncPodUpdate
- // SyncPodCreate is when the pod is created from source
- SyncPodCreate
- // SyncPodKill is when the pod is killed based on a trigger internal to the kubelet for eviction.
- // If a SyncPodKill request is made to pod workers, the request is never dropped, and will always be processed.
- SyncPodKill
-)
-
-func (sp SyncPodType) String() string {
- switch sp {
- case SyncPodCreate:
- return "create"
- case SyncPodUpdate:
- return "update"
- case SyncPodSync:
- return "sync"
- case SyncPodKill:
- return "kill"
- default:
- return "unknown"
- }
-}
-
-// IsCriticalPod returns true if the pod bears the critical pod annotation key or if pod's priority is greater than
-// or equal to SystemCriticalPriority. Both the rescheduler(deprecated in 1.10) and the kubelet use this function
-// to make admission and scheduling decisions.
-func IsCriticalPod(pod *v1.Pod) bool {
- return IsCritical(pod.Namespace, pod.Annotations) || (pod.Spec.Priority != nil && IsCriticalPodBasedOnPriority(pod.Namespace, *pod.Spec.Priority))
-}
-
-// IsCritical returns true if parameters bear the critical pod annotation
-// key. The DaemonSetController use this key directly to make scheduling decisions.
-// TODO: @ravig - Deprecated. Remove this when we move to resolving critical pods based on priorityClassName.
-func IsCritical(ns string, annotations map[string]string) bool {
- // Critical pods are restricted to "kube-system" namespace as of now.
- if ns != kubeapi.NamespaceSystem {
- return false
- }
- val, ok := annotations[CriticalPodAnnotationKey]
- if ok && val == "" {
- return true
- }
- return false
-}
-
-// IsCriticalPodBasedOnPriority checks if the given pod is a critical pod based on priority resolved from pod Spec.
-func IsCriticalPodBasedOnPriority(ns string, priority int32) bool {
- // Critical pods are restricted to "kube-system" namespace as of now.
- if ns != kubeapi.NamespaceSystem {
- return false
- }
- if priority >= schedulerapi.SystemCriticalPriority {
- return true
- }
- return false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/types/types.go b/vendor/k8s.io/kubernetes/pkg/kubelet/types/types.go
deleted file mode 100644
index b0dff97a7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/types/types.go
+++ /dev/null
@@ -1,100 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package types
-
-import (
- "net/http"
- "time"
-
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/types"
-)
-
-// TODO: Reconcile custom types in kubelet/types and this subpackage
-
-type HttpGetter interface {
- Get(url string) (*http.Response, error)
-}
-
-// Timestamp wraps around time.Time and offers utilities to format and parse
-// the time using RFC3339Nano
-type Timestamp struct {
- time time.Time
-}
-
-// NewTimestamp returns a Timestamp object using the current time.
-func NewTimestamp() *Timestamp {
- return &Timestamp{time.Now()}
-}
-
-// ConvertToTimestamp takes a string, parses it using the RFC3339Nano layout,
-// and converts it to a Timestamp object.
-func ConvertToTimestamp(timeString string) *Timestamp {
- parsed, _ := time.Parse(time.RFC3339Nano, timeString)
- return &Timestamp{parsed}
-}
-
-// Get returns the time as time.Time.
-func (t *Timestamp) Get() time.Time {
- return t.time
-}
-
-// GetString returns the time in the string format using the RFC3339Nano
-// layout.
-func (t *Timestamp) GetString() string {
- return t.time.Format(time.RFC3339Nano)
-}
-
-// A type to help sort container statuses based on container names.
-type SortedContainerStatuses []v1.ContainerStatus
-
-func (s SortedContainerStatuses) Len() int { return len(s) }
-func (s SortedContainerStatuses) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
-
-func (s SortedContainerStatuses) Less(i, j int) bool {
- return s[i].Name < s[j].Name
-}
-
-// SortInitContainerStatuses ensures that statuses are in the order that their
-// init container appears in the pod spec
-func SortInitContainerStatuses(p *v1.Pod, statuses []v1.ContainerStatus) {
- containers := p.Spec.InitContainers
- current := 0
- for _, container := range containers {
- for j := current; j < len(statuses); j++ {
- if container.Name == statuses[j].Name {
- statuses[current], statuses[j] = statuses[j], statuses[current]
- current++
- break
- }
- }
- }
-}
-
-// Reservation represents reserved resources for non-pod components.
-type Reservation struct {
- // System represents resources reserved for non-kubernetes components.
- System v1.ResourceList
- // Kubernetes represents resources reserved for kubernetes system components.
- Kubernetes v1.ResourceList
-}
-
-// A pod UID which has been translated/resolved to the representation known to kubelets.
-type ResolvedPodUID types.UID
-
-// A pod UID for a mirror pod.
-type MirrorPodUID types.UID
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/pod.go b/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/pod.go
deleted file mode 100644
index 45c0d8b4b..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/pod.go
+++ /dev/null
@@ -1,72 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package format
-
-import (
- "fmt"
- "strings"
- "time"
-
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/types"
-)
-
-type podHandler func(*v1.Pod) string
-
-// Pod returns a string representing a pod in a consistent human readable format,
-// with pod UID as part of the string.
-func Pod(pod *v1.Pod) string {
- return PodDesc(pod.Name, pod.Namespace, pod.UID)
-}
-
-// PodDesc returns a string representing a pod in a consistent human readable format,
-// with pod UID as part of the string.
-func PodDesc(podName, podNamespace string, podUID types.UID) string {
- // Use underscore as the delimiter because it is not allowed in pod name
- // (DNS subdomain format), while allowed in the container name format.
- return fmt.Sprintf("%s_%s(%s)", podName, podNamespace, podUID)
-}
-
-// PodWithDeletionTimestamp is the same as Pod. In addition, it prints the
-// deletion timestamp of the pod if it's not nil.
-func PodWithDeletionTimestamp(pod *v1.Pod) string {
- var deletionTimestamp string
- if pod.DeletionTimestamp != nil {
- deletionTimestamp = ":DeletionTimestamp=" + pod.DeletionTimestamp.UTC().Format(time.RFC3339)
- }
- return Pod(pod) + deletionTimestamp
-}
-
-// Pods returns a string representation a list of pods in a human
-// readable format.
-func Pods(pods []*v1.Pod) string {
- return aggregatePods(pods, Pod)
-}
-
-// PodsWithDeletiontimestamps is the same as Pods. In addition, it prints the
-// deletion timestamps of the pods if they are not nil.
-func PodsWithDeletiontimestamps(pods []*v1.Pod) string {
- return aggregatePods(pods, PodWithDeletionTimestamp)
-}
-
-func aggregatePods(pods []*v1.Pod, handler podHandler) string {
- podStrings := make([]string, 0, len(pods))
- for _, pod := range pods {
- podStrings = append(podStrings, handler(pod))
- }
- return fmt.Sprintf(strings.Join(podStrings, ", "))
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/resources.go b/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/resources.go
deleted file mode 100644
index 2a64c5b72..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/util/format/resources.go
+++ /dev/null
@@ -1,36 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package format
-
-import (
- "fmt"
- "sort"
- "strings"
-
- "k8s.io/api/core/v1"
-)
-
-// ResourceList returns a string representation of a resource list in a human readable format.
-func ResourceList(resources v1.ResourceList) string {
- resourceStrings := make([]string, 0, len(resources))
- for key, value := range resources {
- resourceStrings = append(resourceStrings, fmt.Sprintf("%v=%v", key, value.String()))
- }
- // sort the results for consistent log output
- sort.Strings(resourceStrings)
- return strings.Join(resourceStrings, ",")
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/kubelet/util/ioutils/ioutils.go b/vendor/k8s.io/kubernetes/pkg/kubelet/util/ioutils/ioutils.go
deleted file mode 100644
index 42f1998c7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/kubelet/util/ioutils/ioutils.go
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package ioutils
-
-import "io"
-
-// writeCloserWrapper represents a WriteCloser whose closer operation is noop.
-type writeCloserWrapper struct {
- Writer io.Writer
-}
-
-func (w *writeCloserWrapper) Write(buf []byte) (int, error) {
- return w.Writer.Write(buf)
-}
-
-func (w *writeCloserWrapper) Close() error {
- return nil
-}
-
-// WriteCloserWrapper returns a writeCloserWrapper.
-func WriteCloserWrapper(w io.Writer) io.WriteCloser {
- return &writeCloserWrapper{w}
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/master/ports/doc.go b/vendor/k8s.io/kubernetes/pkg/master/ports/doc.go
deleted file mode 100644
index 5e14f82e7..000000000
--- a/vendor/k8s.io/kubernetes/pkg/master/ports/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package ports defines ports used by various pieces of the kubernetes
-// infrastructure.
-package ports // import "k8s.io/kubernetes/pkg/master/ports"
diff --git a/vendor/k8s.io/kubernetes/pkg/master/ports/ports.go b/vendor/k8s.io/kubernetes/pkg/master/ports/ports.go
deleted file mode 100644
index d4a42f603..000000000
--- a/vendor/k8s.io/kubernetes/pkg/master/ports/ports.go
+++ /dev/null
@@ -1,44 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package ports
-
-const (
- // ProxyStatusPort is the default port for the proxy metrics server.
- // May be overridden by a flag at startup.
- ProxyStatusPort = 10249
- // KubeletPort is the default port for the kubelet server on each host machine.
- // May be overridden by a flag at startup.
- KubeletPort = 10250
- // SchedulerPort is the default port for the scheduler status server.
- // May be overridden by a flag at startup.
- SchedulerPort = 10251
- // InsecureKubeControllerManagerPort is the default port for the controller manager status server.
- // May be overridden by a flag at startup.
- InsecureKubeControllerManagerPort = 10252
- // InsecureCloudControllerManagerPort is the default port for the cloud controller manager server.
- // This value may be overridden by a flag at startup.
- InsecureCloudControllerManagerPort = 10253
- // KubeletReadOnlyPort exposes basic read-only services from the kubelet.
- // May be overridden by a flag at startup.
- // This is necessary for heapster to collect monitoring stats from the kubelet
- // until heapster can transition to using the SSL endpoint.
- // TODO(roberthbailey): Remove this once we have a better solution for heapster.
- KubeletReadOnlyPort = 10255
- // ProxyHealthzPort is the default port for the proxy healthz server.
- // May be overridden by a flag at startup.
- ProxyHealthzPort = 10256
-)
diff --git a/vendor/k8s.io/kubernetes/pkg/scheduler/api/doc.go b/vendor/k8s.io/kubernetes/pkg/scheduler/api/doc.go
deleted file mode 100644
index c768a8c92..000000000
--- a/vendor/k8s.io/kubernetes/pkg/scheduler/api/doc.go
+++ /dev/null
@@ -1,20 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// +k8s:deepcopy-gen=package
-
-// Package api contains scheduler API objects.
-package api // import "k8s.io/kubernetes/pkg/scheduler/api"
diff --git a/vendor/k8s.io/kubernetes/pkg/scheduler/api/register.go b/vendor/k8s.io/kubernetes/pkg/scheduler/api/register.go
deleted file mode 100644
index 4852cd559..000000000
--- a/vendor/k8s.io/kubernetes/pkg/scheduler/api/register.go
+++ /dev/null
@@ -1,55 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package api
-
-import (
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
-)
-
-// Scheme is the default instance of runtime.Scheme to which types in the Kubernetes API are already registered.
-// TODO: remove this, scheduler should not have its own scheme.
-var Scheme = runtime.NewScheme()
-
-// SchemeGroupVersion is group version used to register these objects
-// TODO this should be in the "scheduler" group
-var SchemeGroupVersion = schema.GroupVersion{Group: "", Version: runtime.APIVersionInternal}
-
-var (
- // SchemeBuilder defines a SchemeBuilder object.
- SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes)
- // AddToScheme is used to add stored functions to scheme.
- AddToScheme = SchemeBuilder.AddToScheme
-)
-
-func init() {
- if err := addKnownTypes(Scheme); err != nil {
- // Programmer error.
- panic(err)
- }
-}
-
-func addKnownTypes(scheme *runtime.Scheme) error {
- if err := scheme.AddIgnoredConversionType(&metav1.TypeMeta{}, &metav1.TypeMeta{}); err != nil {
- return err
- }
- scheme.AddKnownTypes(SchemeGroupVersion,
- &Policy{},
- )
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/scheduler/api/types.go b/vendor/k8s.io/kubernetes/pkg/scheduler/api/types.go
deleted file mode 100644
index cfc8d219e..000000000
--- a/vendor/k8s.io/kubernetes/pkg/scheduler/api/types.go
+++ /dev/null
@@ -1,281 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package api
-
-import (
- "time"
-
- "k8s.io/api/core/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/types"
- restclient "k8s.io/client-go/rest"
-)
-
-const (
- // MaxUint defines the max unsigned int value.
- MaxUint = ^uint(0)
- // MaxInt defines the max signed int value.
- MaxInt = int(MaxUint >> 1)
- // MaxTotalPriority defines the max total priority value.
- MaxTotalPriority = MaxInt
- // MaxPriority defines the max priority value.
- MaxPriority = 10
- // MaxWeight defines the max weight value.
- MaxWeight = MaxInt / MaxPriority
- // HighestUserDefinablePriority is the highest priority for user defined priority classes. Priority values larger than 1 billion are reserved for Kubernetes system use.
- HighestUserDefinablePriority = int32(1000000000)
- // SystemCriticalPriority is the beginning of the range of priority values for critical system components.
- SystemCriticalPriority = 2 * HighestUserDefinablePriority
- // NOTE: In order to avoid conflict of names with user-defined priority classes, all the names must
- // start with scheduling.SystemPriorityClassPrefix which is by default "system-".
- SystemClusterCritical = "system-cluster-critical"
- SystemNodeCritical = "system-node-critical"
-)
-
-// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
-
-// Policy describes a struct of a policy resource in api.
-type Policy struct {
- metav1.TypeMeta
- // Holds the information to configure the fit predicate functions.
- // If unspecified, the default predicate functions will be applied.
- // If empty list, all predicates (except the mandatory ones) will be
- // bypassed.
- Predicates []PredicatePolicy
- // Holds the information to configure the priority functions.
- // If unspecified, the default priority functions will be applied.
- // If empty list, all priority functions will be bypassed.
- Priorities []PriorityPolicy
- // Holds the information to communicate with the extender(s)
- ExtenderConfigs []ExtenderConfig
- // RequiredDuringScheduling affinity is not symmetric, but there is an implicit PreferredDuringScheduling affinity rule
- // corresponding to every RequiredDuringScheduling affinity rule.
- // HardPodAffinitySymmetricWeight represents the weight of implicit PreferredDuringScheduling affinity rule, in the range 1-100.
- HardPodAffinitySymmetricWeight int32
-
- // When AlwaysCheckAllPredicates is set to true, scheduler checks all
- // the configured predicates even after one or more of them fails.
- // When the flag is set to false, scheduler skips checking the rest
- // of the predicates after it finds one predicate that failed.
- AlwaysCheckAllPredicates bool
-}
-
-// PredicatePolicy describes a struct of a predicate policy.
-type PredicatePolicy struct {
- // Identifier of the predicate policy
- // For a custom predicate, the name can be user-defined
- // For the Kubernetes provided predicates, the name is the identifier of the pre-defined predicate
- Name string
- // Holds the parameters to configure the given predicate
- Argument *PredicateArgument
-}
-
-// PriorityPolicy describes a struct of a priority policy.
-type PriorityPolicy struct {
- // Identifier of the priority policy
- // For a custom priority, the name can be user-defined
- // For the Kubernetes provided priority functions, the name is the identifier of the pre-defined priority function
- Name string
- // The numeric multiplier for the node scores that the priority function generates
- // The weight should be a positive integer
- Weight int
- // Holds the parameters to configure the given priority function
- Argument *PriorityArgument
-}
-
-// PredicateArgument represents the arguments to configure predicate functions in scheduler policy configuration.
-// Only one of its members may be specified
-type PredicateArgument struct {
- // The predicate that provides affinity for pods belonging to a service
- // It uses a label to identify nodes that belong to the same "group"
- ServiceAffinity *ServiceAffinity
- // The predicate that checks whether a particular node has a certain label
- // defined or not, regardless of value
- LabelsPresence *LabelsPresence
-}
-
-// PriorityArgument represents the arguments to configure priority functions in scheduler policy configuration.
-// Only one of its members may be specified
-type PriorityArgument struct {
- // The priority function that ensures a good spread (anti-affinity) for pods belonging to a service
- // It uses a label to identify nodes that belong to the same "group"
- ServiceAntiAffinity *ServiceAntiAffinity
- // The priority function that checks whether a particular node has a certain label
- // defined or not, regardless of value
- LabelPreference *LabelPreference
-}
-
-// ServiceAffinity holds the parameters that are used to configure the corresponding predicate in scheduler policy configuration.
-type ServiceAffinity struct {
- // The list of labels that identify node "groups"
- // All of the labels should match for the node to be considered a fit for hosting the pod
- Labels []string
-}
-
-// LabelsPresence holds the parameters that are used to configure the corresponding predicate in scheduler policy configuration.
-type LabelsPresence struct {
- // The list of labels that identify node "groups"
- // All of the labels should be either present (or absent) for the node to be considered a fit for hosting the pod
- Labels []string
- // The boolean flag that indicates whether the labels should be present or absent from the node
- Presence bool
-}
-
-// ServiceAntiAffinity holds the parameters that are used to configure the corresponding priority function
-type ServiceAntiAffinity struct {
- // Used to identify node "groups"
- Label string
-}
-
-// LabelPreference holds the parameters that are used to configure the corresponding priority function
-type LabelPreference struct {
- // Used to identify node "groups"
- Label string
- // This is a boolean flag
- // If true, higher priority is given to nodes that have the label
- // If false, higher priority is given to nodes that do not have the label
- Presence bool
-}
-
-// ExtenderManagedResource describes the arguments of extended resources
-// managed by an extender.
-type ExtenderManagedResource struct {
- // Name is the extended resource name.
- Name v1.ResourceName
- // IgnoredByScheduler indicates whether kube-scheduler should ignore this
- // resource when applying predicates.
- IgnoredByScheduler bool
-}
-
-// ExtenderConfig holds the parameters used to communicate with the extender. If a verb is unspecified/empty,
-// it is assumed that the extender chose not to provide that extension.
-type ExtenderConfig struct {
- // URLPrefix at which the extender is available
- URLPrefix string
- // Verb for the filter call, empty if not supported. This verb is appended to the URLPrefix when issuing the filter call to extender.
- FilterVerb string
- // Verb for the prioritize call, empty if not supported. This verb is appended to the URLPrefix when issuing the prioritize call to extender.
- PrioritizeVerb string
- // The numeric multiplier for the node scores that the prioritize call generates.
- // The weight should be a positive integer
- Weight int
- // Verb for the bind call, empty if not supported. This verb is appended to the URLPrefix when issuing the bind call to extender.
- // If this method is implemented by the extender, it is the extender's responsibility to bind the pod to apiserver. Only one extender
- // can implement this function.
- BindVerb string
- // EnableHTTPS specifies whether https should be used to communicate with the extender
- EnableHTTPS bool
- // TLSConfig specifies the transport layer security config
- TLSConfig *restclient.TLSClientConfig
- // HTTPTimeout specifies the timeout duration for a call to the extender. Filter timeout fails the scheduling of the pod. Prioritize
- // timeout is ignored, k8s/other extenders priorities are used to select the node.
- HTTPTimeout time.Duration
- // NodeCacheCapable specifies that the extender is capable of caching node information,
- // so the scheduler should only send minimal information about the eligible nodes
- // assuming that the extender already cached full details of all nodes in the cluster
- NodeCacheCapable bool
- // ManagedResources is a list of extended resources that are managed by
- // this extender.
- // - A pod will be sent to the extender on the Filter, Prioritize and Bind
- // (if the extender is the binder) phases iff the pod requests at least
- // one of the extended resources in this list. If empty or unspecified,
- // all pods will be sent to this extender.
- // - If IgnoredByScheduler is set to true for a resource, kube-scheduler
- // will skip checking the resource in predicates.
- // +optional
- ManagedResources []ExtenderManagedResource
-}
-
-// ExtenderArgs represents the arguments needed by the extender to filter/prioritize
-// nodes for a pod.
-type ExtenderArgs struct {
- // Pod being scheduled
- Pod v1.Pod
- // List of candidate nodes where the pod can be scheduled; to be populated
- // only if ExtenderConfig.NodeCacheCapable == false
- Nodes *v1.NodeList
- // List of candidate node names where the pod can be scheduled; to be
- // populated only if ExtenderConfig.NodeCacheCapable == true
- NodeNames *[]string
-}
-
-// FailedNodesMap represents the filtered out nodes, with node names and failure messages
-type FailedNodesMap map[string]string
-
-// ExtenderFilterResult represents the results of a filter call to an extender
-type ExtenderFilterResult struct {
- // Filtered set of nodes where the pod can be scheduled; to be populated
- // only if ExtenderConfig.NodeCacheCapable == false
- Nodes *v1.NodeList
- // Filtered set of nodes where the pod can be scheduled; to be populated
- // only if ExtenderConfig.NodeCacheCapable == true
- NodeNames *[]string
- // Filtered out nodes where the pod can't be scheduled and the failure messages
- FailedNodes FailedNodesMap
- // Error message indicating failure
- Error string
-}
-
-// ExtenderBindingArgs represents the arguments to an extender for binding a pod to a node.
-type ExtenderBindingArgs struct {
- // PodName is the name of the pod being bound
- PodName string
- // PodNamespace is the namespace of the pod being bound
- PodNamespace string
- // PodUID is the UID of the pod being bound
- PodUID types.UID
- // Node selected by the scheduler
- Node string
-}
-
-// ExtenderBindingResult represents the result of binding of a pod to a node from an extender.
-type ExtenderBindingResult struct {
- // Error message indicating failure
- Error string
-}
-
-// HostPriority represents the priority of scheduling to a particular host, higher priority is better.
-type HostPriority struct {
- // Name of the host
- Host string
- // Score associated with the host
- Score int
-}
-
-// HostPriorityList declares a []HostPriority type.
-type HostPriorityList []HostPriority
-
-// SystemPriorityClasses defines special priority classes which are used by system critical pods that should not be preempted by workload pods.
-var SystemPriorityClasses = map[string]int32{
- SystemClusterCritical: SystemCriticalPriority,
- SystemNodeCritical: SystemCriticalPriority + 1000,
-}
-
-func (h HostPriorityList) Len() int {
- return len(h)
-}
-
-func (h HostPriorityList) Less(i, j int) bool {
- if h[i].Score == h[j].Score {
- return h[i].Host < h[j].Host
- }
- return h[i].Score < h[j].Score
-}
-
-func (h HostPriorityList) Swap(i, j int) {
- h[i], h[j] = h[j], h[i]
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/scheduler/api/zz_generated.deepcopy.go b/vendor/k8s.io/kubernetes/pkg/scheduler/api/zz_generated.deepcopy.go
deleted file mode 100644
index 1986933b9..000000000
--- a/vendor/k8s.io/kubernetes/pkg/scheduler/api/zz_generated.deepcopy.go
+++ /dev/null
@@ -1,485 +0,0 @@
-// +build !ignore_autogenerated
-
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Code generated by deepcopy-gen. DO NOT EDIT.
-
-package api
-
-import (
- v1 "k8s.io/api/core/v1"
- runtime "k8s.io/apimachinery/pkg/runtime"
- rest "k8s.io/client-go/rest"
-)
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderArgs) DeepCopyInto(out *ExtenderArgs) {
- *out = *in
- in.Pod.DeepCopyInto(&out.Pod)
- if in.Nodes != nil {
- in, out := &in.Nodes, &out.Nodes
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.NodeList)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.NodeNames != nil {
- in, out := &in.NodeNames, &out.NodeNames
- if *in == nil {
- *out = nil
- } else {
- *out = new([]string)
- if **in != nil {
- in, out := *in, *out
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderArgs.
-func (in *ExtenderArgs) DeepCopy() *ExtenderArgs {
- if in == nil {
- return nil
- }
- out := new(ExtenderArgs)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderBindingArgs) DeepCopyInto(out *ExtenderBindingArgs) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderBindingArgs.
-func (in *ExtenderBindingArgs) DeepCopy() *ExtenderBindingArgs {
- if in == nil {
- return nil
- }
- out := new(ExtenderBindingArgs)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderBindingResult) DeepCopyInto(out *ExtenderBindingResult) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderBindingResult.
-func (in *ExtenderBindingResult) DeepCopy() *ExtenderBindingResult {
- if in == nil {
- return nil
- }
- out := new(ExtenderBindingResult)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderConfig) DeepCopyInto(out *ExtenderConfig) {
- *out = *in
- if in.TLSConfig != nil {
- in, out := &in.TLSConfig, &out.TLSConfig
- if *in == nil {
- *out = nil
- } else {
- *out = new(rest.TLSClientConfig)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.ManagedResources != nil {
- in, out := &in.ManagedResources, &out.ManagedResources
- *out = make([]ExtenderManagedResource, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderConfig.
-func (in *ExtenderConfig) DeepCopy() *ExtenderConfig {
- if in == nil {
- return nil
- }
- out := new(ExtenderConfig)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderFilterResult) DeepCopyInto(out *ExtenderFilterResult) {
- *out = *in
- if in.Nodes != nil {
- in, out := &in.Nodes, &out.Nodes
- if *in == nil {
- *out = nil
- } else {
- *out = new(v1.NodeList)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.NodeNames != nil {
- in, out := &in.NodeNames, &out.NodeNames
- if *in == nil {
- *out = nil
- } else {
- *out = new([]string)
- if **in != nil {
- in, out := *in, *out
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- }
- }
- if in.FailedNodes != nil {
- in, out := &in.FailedNodes, &out.FailedNodes
- *out = make(FailedNodesMap, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderFilterResult.
-func (in *ExtenderFilterResult) DeepCopy() *ExtenderFilterResult {
- if in == nil {
- return nil
- }
- out := new(ExtenderFilterResult)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ExtenderManagedResource) DeepCopyInto(out *ExtenderManagedResource) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtenderManagedResource.
-func (in *ExtenderManagedResource) DeepCopy() *ExtenderManagedResource {
- if in == nil {
- return nil
- }
- out := new(ExtenderManagedResource)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in FailedNodesMap) DeepCopyInto(out *FailedNodesMap) {
- {
- in := &in
- *out = make(FailedNodesMap, len(*in))
- for key, val := range *in {
- (*out)[key] = val
- }
- return
- }
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FailedNodesMap.
-func (in FailedNodesMap) DeepCopy() FailedNodesMap {
- if in == nil {
- return nil
- }
- out := new(FailedNodesMap)
- in.DeepCopyInto(out)
- return *out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *HostPriority) DeepCopyInto(out *HostPriority) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HostPriority.
-func (in *HostPriority) DeepCopy() *HostPriority {
- if in == nil {
- return nil
- }
- out := new(HostPriority)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in HostPriorityList) DeepCopyInto(out *HostPriorityList) {
- {
- in := &in
- *out = make(HostPriorityList, len(*in))
- copy(*out, *in)
- return
- }
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HostPriorityList.
-func (in HostPriorityList) DeepCopy() HostPriorityList {
- if in == nil {
- return nil
- }
- out := new(HostPriorityList)
- in.DeepCopyInto(out)
- return *out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LabelPreference) DeepCopyInto(out *LabelPreference) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LabelPreference.
-func (in *LabelPreference) DeepCopy() *LabelPreference {
- if in == nil {
- return nil
- }
- out := new(LabelPreference)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *LabelsPresence) DeepCopyInto(out *LabelsPresence) {
- *out = *in
- if in.Labels != nil {
- in, out := &in.Labels, &out.Labels
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new LabelsPresence.
-func (in *LabelsPresence) DeepCopy() *LabelsPresence {
- if in == nil {
- return nil
- }
- out := new(LabelsPresence)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *Policy) DeepCopyInto(out *Policy) {
- *out = *in
- out.TypeMeta = in.TypeMeta
- if in.Predicates != nil {
- in, out := &in.Predicates, &out.Predicates
- *out = make([]PredicatePolicy, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.Priorities != nil {
- in, out := &in.Priorities, &out.Priorities
- *out = make([]PriorityPolicy, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- if in.ExtenderConfigs != nil {
- in, out := &in.ExtenderConfigs, &out.ExtenderConfigs
- *out = make([]ExtenderConfig, len(*in))
- for i := range *in {
- (*in)[i].DeepCopyInto(&(*out)[i])
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Policy.
-func (in *Policy) DeepCopy() *Policy {
- if in == nil {
- return nil
- }
- out := new(Policy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
-func (in *Policy) DeepCopyObject() runtime.Object {
- if c := in.DeepCopy(); c != nil {
- return c
- }
- return nil
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PredicateArgument) DeepCopyInto(out *PredicateArgument) {
- *out = *in
- if in.ServiceAffinity != nil {
- in, out := &in.ServiceAffinity, &out.ServiceAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(ServiceAffinity)
- (*in).DeepCopyInto(*out)
- }
- }
- if in.LabelsPresence != nil {
- in, out := &in.LabelsPresence, &out.LabelsPresence
- if *in == nil {
- *out = nil
- } else {
- *out = new(LabelsPresence)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PredicateArgument.
-func (in *PredicateArgument) DeepCopy() *PredicateArgument {
- if in == nil {
- return nil
- }
- out := new(PredicateArgument)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PredicatePolicy) DeepCopyInto(out *PredicatePolicy) {
- *out = *in
- if in.Argument != nil {
- in, out := &in.Argument, &out.Argument
- if *in == nil {
- *out = nil
- } else {
- *out = new(PredicateArgument)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PredicatePolicy.
-func (in *PredicatePolicy) DeepCopy() *PredicatePolicy {
- if in == nil {
- return nil
- }
- out := new(PredicatePolicy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PriorityArgument) DeepCopyInto(out *PriorityArgument) {
- *out = *in
- if in.ServiceAntiAffinity != nil {
- in, out := &in.ServiceAntiAffinity, &out.ServiceAntiAffinity
- if *in == nil {
- *out = nil
- } else {
- *out = new(ServiceAntiAffinity)
- **out = **in
- }
- }
- if in.LabelPreference != nil {
- in, out := &in.LabelPreference, &out.LabelPreference
- if *in == nil {
- *out = nil
- } else {
- *out = new(LabelPreference)
- **out = **in
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PriorityArgument.
-func (in *PriorityArgument) DeepCopy() *PriorityArgument {
- if in == nil {
- return nil
- }
- out := new(PriorityArgument)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *PriorityPolicy) DeepCopyInto(out *PriorityPolicy) {
- *out = *in
- if in.Argument != nil {
- in, out := &in.Argument, &out.Argument
- if *in == nil {
- *out = nil
- } else {
- *out = new(PriorityArgument)
- (*in).DeepCopyInto(*out)
- }
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PriorityPolicy.
-func (in *PriorityPolicy) DeepCopy() *PriorityPolicy {
- if in == nil {
- return nil
- }
- out := new(PriorityPolicy)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceAffinity) DeepCopyInto(out *ServiceAffinity) {
- *out = *in
- if in.Labels != nil {
- in, out := &in.Labels, &out.Labels
- *out = make([]string, len(*in))
- copy(*out, *in)
- }
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceAffinity.
-func (in *ServiceAffinity) DeepCopy() *ServiceAffinity {
- if in == nil {
- return nil
- }
- out := new(ServiceAffinity)
- in.DeepCopyInto(out)
- return out
-}
-
-// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
-func (in *ServiceAntiAffinity) DeepCopyInto(out *ServiceAntiAffinity) {
- *out = *in
- return
-}
-
-// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceAntiAffinity.
-func (in *ServiceAntiAffinity) DeepCopy() *ServiceAntiAffinity {
- if in == nil {
- return nil
- }
- out := new(ServiceAntiAffinity)
- in.DeepCopyInto(out)
- return out
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/security/apparmor/helpers.go b/vendor/k8s.io/kubernetes/pkg/security/apparmor/helpers.go
deleted file mode 100644
index 5352f1332..000000000
--- a/vendor/k8s.io/kubernetes/pkg/security/apparmor/helpers.go
+++ /dev/null
@@ -1,80 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package apparmor
-
-import (
- "strings"
-
- "k8s.io/api/core/v1"
-)
-
-// TODO: Move these values into the API package.
-const (
- // The prefix to an annotation key specifying a container profile.
- ContainerAnnotationKeyPrefix = "container.apparmor.security.beta.kubernetes.io/"
- // The annotation key specifying the default AppArmor profile.
- DefaultProfileAnnotationKey = "apparmor.security.beta.kubernetes.io/defaultProfileName"
- // The annotation key specifying the allowed AppArmor profiles.
- AllowedProfilesAnnotationKey = "apparmor.security.beta.kubernetes.io/allowedProfileNames"
-
- // The profile specifying the runtime default.
- ProfileRuntimeDefault = "runtime/default"
- // The prefix for specifying profiles loaded on the node.
- ProfileNamePrefix = "localhost/"
-
- // Unconfined profile
- ProfileNameUnconfined = "unconfined"
-)
-
-// Checks whether app armor is required for pod to be run.
-func isRequired(pod *v1.Pod) bool {
- for key, value := range pod.Annotations {
- if strings.HasPrefix(key, ContainerAnnotationKeyPrefix) {
- return value != ProfileNameUnconfined
- }
- }
- return false
-}
-
-// Returns the name of the profile to use with the container.
-func GetProfileName(pod *v1.Pod, containerName string) string {
- return GetProfileNameFromPodAnnotations(pod.Annotations, containerName)
-}
-
-// GetProfileNameFromPodAnnotations gets the name of the profile to use with container from
-// pod annotations
-func GetProfileNameFromPodAnnotations(annotations map[string]string, containerName string) string {
- return annotations[ContainerAnnotationKeyPrefix+containerName]
-}
-
-// Sets the name of the profile to use with the container.
-func SetProfileName(pod *v1.Pod, containerName, profileName string) error {
- if pod.Annotations == nil {
- pod.Annotations = map[string]string{}
- }
- pod.Annotations[ContainerAnnotationKeyPrefix+containerName] = profileName
- return nil
-}
-
-// Sets the name of the profile to use with the container.
-func SetProfileNameFromPodAnnotations(annotations map[string]string, containerName, profileName string) error {
- if annotations == nil {
- return nil
- }
- annotations[ContainerAnnotationKeyPrefix+containerName] = profileName
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate.go b/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate.go
deleted file mode 100644
index 740698f20..000000000
--- a/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate.go
+++ /dev/null
@@ -1,229 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package apparmor
-
-import (
- "bufio"
- "errors"
- "fmt"
- "io/ioutil"
- "os"
- "path"
- "strings"
-
- "k8s.io/api/core/v1"
- utilfeature "k8s.io/apiserver/pkg/util/feature"
- "k8s.io/kubernetes/pkg/features"
- kubetypes "k8s.io/kubernetes/pkg/kubelet/types"
- utilfile "k8s.io/kubernetes/pkg/util/file"
-)
-
-// Whether AppArmor should be disabled by default.
-// Set to true if the wrong build tags are set (see validate_disabled.go).
-var isDisabledBuild bool
-
-// Interface for validating that a pod with with an AppArmor profile can be run by a Node.
-type Validator interface {
- Validate(pod *v1.Pod) error
- ValidateHost() error
-}
-
-func NewValidator(runtime string) Validator {
- if err := validateHost(runtime); err != nil {
- return &validator{validateHostErr: err}
- }
- appArmorFS, err := getAppArmorFS()
- if err != nil {
- return &validator{
- validateHostErr: fmt.Errorf("error finding AppArmor FS: %v", err),
- }
- }
- return &validator{
- appArmorFS: appArmorFS,
- }
-}
-
-type validator struct {
- validateHostErr error
- appArmorFS string
-}
-
-func (v *validator) Validate(pod *v1.Pod) error {
- if !isRequired(pod) {
- return nil
- }
-
- if v.ValidateHost() != nil {
- return v.validateHostErr
- }
-
- loadedProfiles, err := v.getLoadedProfiles()
- if err != nil {
- return fmt.Errorf("could not read loaded profiles: %v", err)
- }
-
- for _, container := range pod.Spec.InitContainers {
- if err := validateProfile(GetProfileName(pod, container.Name), loadedProfiles); err != nil {
- return err
- }
- }
- for _, container := range pod.Spec.Containers {
- if err := validateProfile(GetProfileName(pod, container.Name), loadedProfiles); err != nil {
- return err
- }
- }
-
- return nil
-}
-
-func (v *validator) ValidateHost() error {
- return v.validateHostErr
-}
-
-// Verify that the host and runtime is capable of enforcing AppArmor profiles.
-func validateHost(runtime string) error {
- // Check feature-gates
- if !utilfeature.DefaultFeatureGate.Enabled(features.AppArmor) {
- return errors.New("AppArmor disabled by feature-gate")
- }
-
- // Check build support.
- if isDisabledBuild {
- return errors.New("Binary not compiled for linux")
- }
-
- // Check kernel support.
- if !IsAppArmorEnabled() {
- return errors.New("AppArmor is not enabled on the host")
- }
-
- // Check runtime support. Currently only Docker is supported.
- if runtime != kubetypes.DockerContainerRuntime && runtime != kubetypes.RemoteContainerRuntime {
- return fmt.Errorf("AppArmor is only enabled for 'docker' and 'remote' runtimes. Found: %q.", runtime)
- }
-
- return nil
-}
-
-// Verify that the profile is valid and loaded.
-func validateProfile(profile string, loadedProfiles map[string]bool) error {
- if err := ValidateProfileFormat(profile); err != nil {
- return err
- }
-
- if strings.HasPrefix(profile, ProfileNamePrefix) {
- profileName := strings.TrimPrefix(profile, ProfileNamePrefix)
- if !loadedProfiles[profileName] {
- return fmt.Errorf("profile %q is not loaded", profileName)
- }
- }
-
- return nil
-}
-
-func ValidateProfileFormat(profile string) error {
- if profile == "" || profile == ProfileRuntimeDefault || profile == ProfileNameUnconfined {
- return nil
- }
- if !strings.HasPrefix(profile, ProfileNamePrefix) {
- return fmt.Errorf("invalid AppArmor profile name: %q", profile)
- }
- return nil
-}
-
-func (v *validator) getLoadedProfiles() (map[string]bool, error) {
- profilesPath := path.Join(v.appArmorFS, "profiles")
- profilesFile, err := os.Open(profilesPath)
- if err != nil {
- return nil, fmt.Errorf("failed to open %s: %v", profilesPath, err)
- }
- defer profilesFile.Close()
-
- profiles := map[string]bool{}
- scanner := bufio.NewScanner(profilesFile)
- for scanner.Scan() {
- profileName := parseProfileName(scanner.Text())
- if profileName == "" {
- // Unknown line format; skip it.
- continue
- }
- profiles[profileName] = true
- }
- return profiles, nil
-}
-
-// The profiles file is formatted with one profile per line, matching a form:
-// namespace://profile-name (mode)
-// profile-name (mode)
-// Where mode is {enforce, complain, kill}. The "namespace://" is only included for namespaced
-// profiles. For the purposes of Kubernetes, we consider the namespace part of the profile name.
-func parseProfileName(profileLine string) string {
- modeIndex := strings.IndexRune(profileLine, '(')
- if modeIndex < 0 {
- return ""
- }
- return strings.TrimSpace(profileLine[:modeIndex])
-}
-
-func getAppArmorFS() (string, error) {
- mountsFile, err := os.Open("/proc/mounts")
- if err != nil {
- return "", fmt.Errorf("could not open /proc/mounts: %v", err)
- }
- defer mountsFile.Close()
-
- scanner := bufio.NewScanner(mountsFile)
- for scanner.Scan() {
- fields := strings.Fields(scanner.Text())
- if len(fields) < 3 {
- // Unknown line format; skip it.
- continue
- }
- if fields[2] == "securityfs" {
- appArmorFS := path.Join(fields[1], "apparmor")
- if ok, err := utilfile.FileExists(appArmorFS); !ok {
- msg := fmt.Sprintf("path %s does not exist", appArmorFS)
- if err != nil {
- return "", fmt.Errorf("%s: %v", msg, err)
- } else {
- return "", errors.New(msg)
- }
- } else {
- return appArmorFS, nil
- }
- }
- }
- if err := scanner.Err(); err != nil {
- return "", fmt.Errorf("error scanning mounts: %v", err)
- }
-
- return "", errors.New("securityfs not found")
-}
-
-// IsAppArmorEnabled returns true if apparmor is enabled for the host.
-// This function is forked from
-// https://github.com/opencontainers/runc/blob/1a81e9ab1f138c091fe5c86d0883f87716088527/libcontainer/apparmor/apparmor.go
-// to avoid the libapparmor dependency.
-func IsAppArmorEnabled() bool {
- if _, err := os.Stat("/sys/kernel/security/apparmor"); err == nil && os.Getenv("container") == "" {
- if _, err = os.Stat("/sbin/apparmor_parser"); err == nil {
- buf, err := ioutil.ReadFile("/sys/module/apparmor/parameters/enabled")
- return err == nil && len(buf) > 1 && buf[0] == 'Y'
- }
- }
- return false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate_disabled.go b/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate_disabled.go
deleted file mode 100644
index 875054a94..000000000
--- a/vendor/k8s.io/kubernetes/pkg/security/apparmor/validate_disabled.go
+++ /dev/null
@@ -1,24 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package apparmor
-
-func init() {
- // If Kubernetes was not built for linux, apparmor is always disabled.
- isDisabledBuild = true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/serviceaccount/claims.go b/vendor/k8s.io/kubernetes/pkg/serviceaccount/claims.go
deleted file mode 100644
index c3ed29c4c..000000000
--- a/vendor/k8s.io/kubernetes/pkg/serviceaccount/claims.go
+++ /dev/null
@@ -1,186 +0,0 @@
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package serviceaccount
-
-import (
- "errors"
- "fmt"
- "time"
-
- "github.com/golang/glog"
- apiserverserviceaccount "k8s.io/apiserver/pkg/authentication/serviceaccount"
- "k8s.io/kubernetes/pkg/apis/core"
-
- "gopkg.in/square/go-jose.v2/jwt"
-)
-
-// time.Now stubbed out to allow testing
-var now = time.Now
-
-type privateClaims struct {
- Kubernetes kubernetes `json:"kubernetes.io,omitempty"`
-}
-
-type kubernetes struct {
- Namespace string `json:"namespace,omitempty"`
- Svcacct ref `json:"serviceaccount,omitempty"`
- Pod *ref `json:"pod,omitempty"`
- Secret *ref `json:"secret,omitempty"`
-}
-
-type ref struct {
- Name string `json:"name,omitempty"`
- UID string `json:"uid,omitempty"`
-}
-
-func Claims(sa core.ServiceAccount, pod *core.Pod, secret *core.Secret, expirationSeconds int64, audience []string) (*jwt.Claims, interface{}) {
- now := now()
- sc := &jwt.Claims{
- Subject: apiserverserviceaccount.MakeUsername(sa.Namespace, sa.Name),
- Audience: jwt.Audience(audience),
- IssuedAt: jwt.NewNumericDate(now),
- NotBefore: jwt.NewNumericDate(now),
- Expiry: jwt.NewNumericDate(now.Add(time.Duration(expirationSeconds) * time.Second)),
- }
- pc := &privateClaims{
- Kubernetes: kubernetes{
- Namespace: sa.Namespace,
- Svcacct: ref{
- Name: sa.Name,
- UID: string(sa.UID),
- },
- },
- }
- switch {
- case pod != nil:
- pc.Kubernetes.Pod = &ref{
- Name: pod.Name,
- UID: string(pod.UID),
- }
- case secret != nil:
- pc.Kubernetes.Secret = &ref{
- Name: secret.Name,
- UID: string(secret.UID),
- }
- }
- return sc, pc
-}
-
-func NewValidator(audiences []string, getter ServiceAccountTokenGetter) Validator {
- return &validator{
- auds: audiences,
- getter: getter,
- }
-}
-
-type validator struct {
- auds []string
- getter ServiceAccountTokenGetter
-}
-
-var _ = Validator(&validator{})
-
-func (v *validator) Validate(_ string, public *jwt.Claims, privateObj interface{}) (string, string, string, error) {
- private, ok := privateObj.(*privateClaims)
- if !ok {
- glog.Errorf("jwt validator expected private claim of type *privateClaims but got: %T", privateObj)
- return "", "", "", errors.New("Token could not be validated.")
- }
- err := public.Validate(jwt.Expected{
- Time: now(),
- })
- switch {
- case err == nil:
- case err == jwt.ErrExpired:
- return "", "", "", errors.New("Token has expired.")
- default:
- glog.Errorf("unexpected validation error: %T", err)
- return "", "", "", errors.New("Token could not be validated.")
- }
-
- var audValid bool
-
- for _, aud := range v.auds {
- audValid = public.Audience.Contains(aud)
- if audValid {
- break
- }
- }
-
- if !audValid {
- return "", "", "", errors.New("Token is invalid for this audience.")
- }
-
- namespace := private.Kubernetes.Namespace
- saref := private.Kubernetes.Svcacct
- podref := private.Kubernetes.Pod
- secref := private.Kubernetes.Secret
- // Make sure service account still exists (name and UID)
- serviceAccount, err := v.getter.GetServiceAccount(namespace, saref.Name)
- if err != nil {
- glog.V(4).Infof("Could not retrieve service account %s/%s: %v", namespace, saref.Name, err)
- return "", "", "", err
- }
- if serviceAccount.DeletionTimestamp != nil {
- glog.V(4).Infof("Service account has been deleted %s/%s", namespace, saref.Name)
- return "", "", "", fmt.Errorf("ServiceAccount %s/%s has been deleted", namespace, saref.Name)
- }
- if string(serviceAccount.UID) != saref.UID {
- glog.V(4).Infof("Service account UID no longer matches %s/%s: %q != %q", namespace, saref.Name, string(serviceAccount.UID), saref.UID)
- return "", "", "", fmt.Errorf("ServiceAccount UID (%s) does not match claim (%s)", serviceAccount.UID, saref.UID)
- }
-
- if secref != nil {
- // Make sure token hasn't been invalidated by deletion of the secret
- secret, err := v.getter.GetSecret(namespace, secref.Name)
- if err != nil {
- glog.V(4).Infof("Could not retrieve bound secret %s/%s for service account %s/%s: %v", namespace, secref.Name, namespace, saref.Name, err)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if secret.DeletionTimestamp != nil {
- glog.V(4).Infof("Bound secret is deleted and awaiting removal: %s/%s for service account %s/%s", namespace, secref.Name, namespace, saref.Name)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if string(secref.UID) != secref.UID {
- glog.V(4).Infof("Secret UID no longer matches %s/%s: %q != %q", namespace, secref.Name, string(serviceAccount.UID), secref.UID)
- return "", "", "", fmt.Errorf("Secret UID (%s) does not match claim (%s)", secret.UID, secref.UID)
- }
- }
-
- if podref != nil {
- // Make sure token hasn't been invalidated by deletion of the pod
- pod, err := v.getter.GetPod(namespace, podref.Name)
- if err != nil {
- glog.V(4).Infof("Could not retrieve bound secret %s/%s for service account %s/%s: %v", namespace, podref.Name, namespace, saref.Name, err)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if pod.DeletionTimestamp != nil {
- glog.V(4).Infof("Bound pod is deleted and awaiting removal: %s/%s for service account %s/%s", namespace, podref.Name, namespace, saref.Name)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if string(podref.UID) != podref.UID {
- glog.V(4).Infof("Pod UID no longer matches %s/%s: %q != %q", namespace, podref.Name, string(serviceAccount.UID), podref.UID)
- return "", "", "", fmt.Errorf("Pod UID (%s) does not match claim (%s)", pod.UID, podref.UID)
- }
- }
-
- return private.Kubernetes.Namespace, private.Kubernetes.Svcacct.Name, private.Kubernetes.Svcacct.UID, nil
-}
-
-func (v *validator) NewPrivateClaims() interface{} {
- return &privateClaims{}
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/serviceaccount/jwt.go b/vendor/k8s.io/kubernetes/pkg/serviceaccount/jwt.go
deleted file mode 100644
index 01c369a31..000000000
--- a/vendor/k8s.io/kubernetes/pkg/serviceaccount/jwt.go
+++ /dev/null
@@ -1,212 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package serviceaccount
-
-import (
- "crypto/ecdsa"
- "crypto/elliptic"
- "crypto/rsa"
- "encoding/base64"
- "encoding/json"
- "errors"
- "fmt"
- "strings"
-
- "k8s.io/api/core/v1"
- utilerrors "k8s.io/apimachinery/pkg/util/errors"
- "k8s.io/apiserver/pkg/authentication/authenticator"
- "k8s.io/apiserver/pkg/authentication/user"
-
- jose "gopkg.in/square/go-jose.v2"
- "gopkg.in/square/go-jose.v2/jwt"
-)
-
-// ServiceAccountTokenGetter defines functions to retrieve a named service account and secret
-type ServiceAccountTokenGetter interface {
- GetServiceAccount(namespace, name string) (*v1.ServiceAccount, error)
- GetPod(namespace, name string) (*v1.Pod, error)
- GetSecret(namespace, name string) (*v1.Secret, error)
-}
-
-type TokenGenerator interface {
- // GenerateToken generates a token which will identify the given
- // ServiceAccount. privateClaims is an interface that will be
- // serialized into the JWT payload JSON encoding at the root level of
- // the payload object. Public claims take precedent over private
- // claims i.e. if both claims and privateClaims have an "exp" field,
- // the value in claims will be used.
- GenerateToken(claims *jwt.Claims, privateClaims interface{}) (string, error)
-}
-
-// JWTTokenGenerator returns a TokenGenerator that generates signed JWT tokens, using the given privateKey.
-// privateKey is a PEM-encoded byte array of a private RSA key.
-// JWTTokenAuthenticator()
-func JWTTokenGenerator(iss string, privateKey interface{}) TokenGenerator {
- return &jwtTokenGenerator{
- iss: iss,
- privateKey: privateKey,
- }
-}
-
-type jwtTokenGenerator struct {
- iss string
- privateKey interface{}
-}
-
-func (j *jwtTokenGenerator) GenerateToken(claims *jwt.Claims, privateClaims interface{}) (string, error) {
- var alg jose.SignatureAlgorithm
- switch privateKey := j.privateKey.(type) {
- case *rsa.PrivateKey:
- alg = jose.RS256
- case *ecdsa.PrivateKey:
- switch privateKey.Curve {
- case elliptic.P256():
- alg = jose.ES256
- case elliptic.P384():
- alg = jose.ES384
- case elliptic.P521():
- alg = jose.ES512
- default:
- return "", fmt.Errorf("unknown private key curve, must be 256, 384, or 521")
- }
- default:
- return "", fmt.Errorf("unknown private key type %T, must be *rsa.PrivateKey or *ecdsa.PrivateKey", j.privateKey)
- }
-
- signer, err := jose.NewSigner(
- jose.SigningKey{
- Algorithm: alg,
- Key: j.privateKey,
- },
- nil,
- )
- if err != nil {
- return "", err
- }
-
- // claims are applied in reverse precedence
- return jwt.Signed(signer).
- Claims(privateClaims).
- Claims(claims).
- Claims(&jwt.Claims{
- Issuer: j.iss,
- }).
- CompactSerialize()
-}
-
-// JWTTokenAuthenticator authenticates tokens as JWT tokens produced by JWTTokenGenerator
-// Token signatures are verified using each of the given public keys until one works (allowing key rotation)
-// If lookup is true, the service account and secret referenced as claims inside the token are retrieved and verified with the provided ServiceAccountTokenGetter
-func JWTTokenAuthenticator(iss string, keys []interface{}, validator Validator) authenticator.Token {
- return &jwtTokenAuthenticator{
- iss: iss,
- keys: keys,
- validator: validator,
- }
-}
-
-type jwtTokenAuthenticator struct {
- iss string
- keys []interface{}
- validator Validator
-}
-
-// Validator is called by the JWT token authentictaor to apply domain specific
-// validation to a token and extract user information.
-type Validator interface {
- // Validate validates a token and returns user information or an error.
- // Validator can assume that the issuer and signature of a token are already
- // verified when this function is called.
- Validate(tokenData string, public *jwt.Claims, private interface{}) (namespace, name, uid string, err error)
- // NewPrivateClaims returns a struct that the authenticator should
- // deserialize the JWT payload into. The authenticator may then pass this
- // struct back to the Validator as the 'private' argument to a Validate()
- // call. This struct should contain fields for any private claims that the
- // Validator requires to validate the JWT.
- NewPrivateClaims() interface{}
-}
-
-var errMismatchedSigningMethod = errors.New("invalid signing method")
-
-func (j *jwtTokenAuthenticator) AuthenticateToken(tokenData string) (user.Info, bool, error) {
- if !j.hasCorrectIssuer(tokenData) {
- return nil, false, nil
- }
-
- tok, err := jwt.ParseSigned(tokenData)
- if err != nil {
- return nil, false, nil
- }
-
- public := &jwt.Claims{}
- private := j.validator.NewPrivateClaims()
-
- var (
- found bool
- errlist []error
- )
- for _, key := range j.keys {
- if err := tok.Claims(key, public, private); err != nil {
- errlist = append(errlist, err)
- continue
- }
- found = true
- break
- }
-
- if !found {
- return nil, false, utilerrors.NewAggregate(errlist)
- }
-
- // If we get here, we have a token with a recognized signature and
- // issuer string.
- ns, name, uid, err := j.validator.Validate(tokenData, public, private)
- if err != nil {
- return nil, false, err
- }
-
- return UserInfo(ns, name, uid), true, nil
-}
-
-// hasCorrectIssuer returns true if tokenData is a valid JWT in compact
-// serialization format and the "iss" claim matches the iss field of this token
-// authenticator, and otherwise returns false.
-//
-// Note: go-jose currently does not allow access to unverified JWS payloads.
-// See https://github.com/square/go-jose/issues/169
-func (j *jwtTokenAuthenticator) hasCorrectIssuer(tokenData string) bool {
- parts := strings.Split(tokenData, ".")
- if len(parts) != 3 {
- return false
- }
- payload, err := base64.RawURLEncoding.DecodeString(parts[1])
- if err != nil {
- return false
- }
- claims := struct {
- // WARNING: this JWT is not verified. Do not trust these claims.
- Issuer string `json:"iss"`
- }{}
- if err := json.Unmarshal(payload, &claims); err != nil {
- return false
- }
- if claims.Issuer != j.iss {
- return false
- }
- return true
-
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/serviceaccount/legacy.go b/vendor/k8s.io/kubernetes/pkg/serviceaccount/legacy.go
deleted file mode 100644
index 5055db7cb..000000000
--- a/vendor/k8s.io/kubernetes/pkg/serviceaccount/legacy.go
+++ /dev/null
@@ -1,135 +0,0 @@
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package serviceaccount
-
-import (
- "bytes"
- "errors"
- "fmt"
-
- "k8s.io/api/core/v1"
- apiserverserviceaccount "k8s.io/apiserver/pkg/authentication/serviceaccount"
-
- "github.com/golang/glog"
- "gopkg.in/square/go-jose.v2/jwt"
-)
-
-func LegacyClaims(serviceAccount v1.ServiceAccount, secret v1.Secret) (*jwt.Claims, interface{}) {
- return &jwt.Claims{
- Subject: apiserverserviceaccount.MakeUsername(serviceAccount.Namespace, serviceAccount.Name),
- }, &legacyPrivateClaims{
- Namespace: serviceAccount.Namespace,
- ServiceAccountName: serviceAccount.Name,
- ServiceAccountUID: string(serviceAccount.UID),
- SecretName: secret.Name,
- }
-}
-
-const LegacyIssuer = "kubernetes/serviceaccount"
-
-type legacyPrivateClaims struct {
- ServiceAccountName string `json:"kubernetes.io/serviceaccount/service-account.name"`
- ServiceAccountUID string `json:"kubernetes.io/serviceaccount/service-account.uid"`
- SecretName string `json:"kubernetes.io/serviceaccount/secret.name"`
- Namespace string `json:"kubernetes.io/serviceaccount/namespace"`
-}
-
-func NewLegacyValidator(lookup bool, getter ServiceAccountTokenGetter) Validator {
- return &legacyValidator{
- lookup: lookup,
- getter: getter,
- }
-}
-
-type legacyValidator struct {
- lookup bool
- getter ServiceAccountTokenGetter
-}
-
-var _ = Validator(&legacyValidator{})
-
-func (v *legacyValidator) Validate(tokenData string, public *jwt.Claims, privateObj interface{}) (string, string, string, error) {
- private, ok := privateObj.(*legacyPrivateClaims)
- if !ok {
- glog.Errorf("jwt validator expected private claim of type *legacyPrivateClaims but got: %T", privateObj)
- return "", "", "", errors.New("Token could not be validated.")
- }
-
- // Make sure the claims we need exist
- if len(public.Subject) == 0 {
- return "", "", "", errors.New("sub claim is missing")
- }
- namespace := private.Namespace
- if len(namespace) == 0 {
- return "", "", "", errors.New("namespace claim is missing")
- }
- secretName := private.SecretName
- if len(secretName) == 0 {
- return "", "", "", errors.New("secretName claim is missing")
- }
- serviceAccountName := private.ServiceAccountName
- if len(serviceAccountName) == 0 {
- return "", "", "", errors.New("serviceAccountName claim is missing")
- }
- serviceAccountUID := private.ServiceAccountUID
- if len(serviceAccountUID) == 0 {
- return "", "", "", errors.New("serviceAccountUID claim is missing")
- }
-
- subjectNamespace, subjectName, err := apiserverserviceaccount.SplitUsername(public.Subject)
- if err != nil || subjectNamespace != namespace || subjectName != serviceAccountName {
- return "", "", "", errors.New("sub claim is invalid")
- }
-
- if v.lookup {
- // Make sure token hasn't been invalidated by deletion of the secret
- secret, err := v.getter.GetSecret(namespace, secretName)
- if err != nil {
- glog.V(4).Infof("Could not retrieve token %s/%s for service account %s/%s: %v", namespace, secretName, namespace, serviceAccountName, err)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if secret.DeletionTimestamp != nil {
- glog.V(4).Infof("Token is deleted and awaiting removal: %s/%s for service account %s/%s", namespace, secretName, namespace, serviceAccountName)
- return "", "", "", errors.New("Token has been invalidated")
- }
- if bytes.Compare(secret.Data[v1.ServiceAccountTokenKey], []byte(tokenData)) != 0 {
- glog.V(4).Infof("Token contents no longer matches %s/%s for service account %s/%s", namespace, secretName, namespace, serviceAccountName)
- return "", "", "", errors.New("Token does not match server's copy")
- }
-
- // Make sure service account still exists (name and UID)
- serviceAccount, err := v.getter.GetServiceAccount(namespace, serviceAccountName)
- if err != nil {
- glog.V(4).Infof("Could not retrieve service account %s/%s: %v", namespace, serviceAccountName, err)
- return "", "", "", err
- }
- if serviceAccount.DeletionTimestamp != nil {
- glog.V(4).Infof("Service account has been deleted %s/%s", namespace, serviceAccountName)
- return "", "", "", fmt.Errorf("ServiceAccount %s/%s has been deleted", namespace, serviceAccountName)
- }
- if string(serviceAccount.UID) != serviceAccountUID {
- glog.V(4).Infof("Service account UID no longer matches %s/%s: %q != %q", namespace, serviceAccountName, string(serviceAccount.UID), serviceAccountUID)
- return "", "", "", fmt.Errorf("ServiceAccount UID (%s) does not match claim (%s)", serviceAccount.UID, serviceAccountUID)
- }
- }
-
- return private.Namespace, private.ServiceAccountName, private.ServiceAccountUID, nil
-}
-
-func (v *legacyValidator) NewPrivateClaims() interface{} {
- return &legacyPrivateClaims{}
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/serviceaccount/util.go b/vendor/k8s.io/kubernetes/pkg/serviceaccount/util.go
deleted file mode 100644
index 0503c1513..000000000
--- a/vendor/k8s.io/kubernetes/pkg/serviceaccount/util.go
+++ /dev/null
@@ -1,74 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package serviceaccount
-
-import (
- "k8s.io/api/core/v1"
- apiserverserviceaccount "k8s.io/apiserver/pkg/authentication/serviceaccount"
- "k8s.io/apiserver/pkg/authentication/user"
- api "k8s.io/kubernetes/pkg/apis/core"
-)
-
-// UserInfo returns a user.Info interface for the given namespace, service account name and UID
-func UserInfo(namespace, name, uid string) user.Info {
- return &user.DefaultInfo{
- Name: apiserverserviceaccount.MakeUsername(namespace, name),
- UID: uid,
- Groups: apiserverserviceaccount.MakeGroupNames(namespace),
- }
-}
-
-// IsServiceAccountToken returns true if the secret is a valid api token for the service account
-func IsServiceAccountToken(secret *v1.Secret, sa *v1.ServiceAccount) bool {
- if secret.Type != v1.SecretTypeServiceAccountToken {
- return false
- }
-
- name := secret.Annotations[v1.ServiceAccountNameKey]
- uid := secret.Annotations[v1.ServiceAccountUIDKey]
- if name != sa.Name {
- // Name must match
- return false
- }
- if len(uid) > 0 && uid != string(sa.UID) {
- // If UID is specified, it must match
- return false
- }
-
- return true
-}
-
-// TODO: remove the duplicate code
-// InternalIsServiceAccountToken returns true if the secret is a valid api token for the service account
-func InternalIsServiceAccountToken(secret *api.Secret, sa *api.ServiceAccount) bool {
- if secret.Type != api.SecretTypeServiceAccountToken {
- return false
- }
-
- name := secret.Annotations[api.ServiceAccountNameKey]
- uid := secret.Annotations[api.ServiceAccountUIDKey]
- if name != sa.Name {
- // Name must match
- return false
- }
- if len(uid) > 0 && uid != string(sa.UID) {
- // If UID is specified, it must match
- return false
- }
-
- return true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/file/file.go b/vendor/k8s.io/kubernetes/pkg/util/file/file.go
deleted file mode 100644
index 70d26c4ef..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/file/file.go
+++ /dev/null
@@ -1,57 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package file
-
-import (
- "os"
-)
-
-// FileExists checks if specified file exists.
-func FileExists(filename string) (bool, error) {
- if _, err := os.Stat(filename); os.IsNotExist(err) {
- return false, nil
- } else if err != nil {
- return false, err
- }
- return true, nil
-}
-
-// FileOrSymlinkExists checks if specified file or symlink exists.
-func FileOrSymlinkExists(filename string) (bool, error) {
- if _, err := os.Lstat(filename); os.IsNotExist(err) {
- return false, nil
- } else if err != nil {
- return false, err
- }
- return true, nil
-}
-
-// ReadDirNoStat returns a string of files/directories contained
-// in dirname without calling lstat on them.
-func ReadDirNoStat(dirname string) ([]string, error) {
- if dirname == "" {
- dirname = "."
- }
-
- f, err := os.Open(dirname)
- if err != nil {
- return nil, err
- }
- defer f.Close()
-
- return f.Readdirnames(-1)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/hash/hash.go b/vendor/k8s.io/kubernetes/pkg/util/hash/hash.go
deleted file mode 100644
index 803f066a4..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/hash/hash.go
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package hash
-
-import (
- "hash"
-
- "github.com/davecgh/go-spew/spew"
-)
-
-// DeepHashObject writes specified object to hash using the spew library
-// which follows pointers and prints actual values of the nested objects
-// ensuring the hash does not change when a pointer changes.
-func DeepHashObject(hasher hash.Hash, objectToWrite interface{}) {
- hasher.Reset()
- printer := spew.ConfigState{
- Indent: " ",
- SortKeys: true,
- DisableMethods: true,
- SpewKeys: true,
- }
- printer.Fprintf(hasher, "%#v", objectToWrite)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/io/consistentread.go b/vendor/k8s.io/kubernetes/pkg/util/io/consistentread.go
deleted file mode 100644
index 6e1f17b09..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/io/consistentread.go
+++ /dev/null
@@ -1,45 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package io
-
-import (
- "bytes"
- "fmt"
- "io/ioutil"
-)
-
-// ConsistentRead repeatedly reads a file until it gets the same content twice.
-// This is useful when reading files in /proc that are larger than page size
-// and kernel may modify them between individual read() syscalls.
-func ConsistentRead(filename string, attempts int) ([]byte, error) {
- oldContent, err := ioutil.ReadFile(filename)
- if err != nil {
- return nil, err
- }
- for i := 0; i < attempts; i++ {
- newContent, err := ioutil.ReadFile(filename)
- if err != nil {
- return nil, err
- }
- if bytes.Compare(oldContent, newContent) == 0 {
- return newContent, nil
- }
- // Files are different, continue reading
- oldContent = newContent
- }
- return nil, fmt.Errorf("could not get consistent content of %s after %d attempts", filename, attempts)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/io/writer.go b/vendor/k8s.io/kubernetes/pkg/util/io/writer.go
deleted file mode 100644
index 8d1d9964e..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/io/writer.go
+++ /dev/null
@@ -1,78 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package io
-
-import (
- "bytes"
- "fmt"
- "io/ioutil"
- "os"
-
- "k8s.io/kubernetes/pkg/util/nsenter"
-
- "github.com/golang/glog"
-)
-
-// Writer is an interface which allows to write data to a file.
-type Writer interface {
- // WriteFile mimics ioutil.WriteFile.
- WriteFile(filename string, data []byte, perm os.FileMode) error
-}
-
-// StdWriter implements Writer interface and uses standard libraries
-// for writing data to files.
-type StdWriter struct {
-}
-
-// WriteFile directly calls ioutil.WriteFile.
-func (writer *StdWriter) WriteFile(filename string, data []byte, perm os.FileMode) error {
- return ioutil.WriteFile(filename, data, perm)
-}
-
-// NsenterWriter is implementation of Writer interface that allows writing data
-// to file using nsenter command.
-// If a program (e.g. kubelet) runs in a container it may want to write data to
-// a mounted device. Since in Docker, mount propagation mode is set to private,
-// it will not see the mounted device in its own namespace. To work around this
-// limitation one has to first enter hosts namespace (by using 'nsenter') and
-// only then write data.
-type NsenterWriter struct{}
-
-// WriteFile calls 'nsenter cat - > <the file>' and 'nsenter chmod' to create a
-// file on the host.
-func (writer *NsenterWriter) WriteFile(filename string, data []byte, perm os.FileMode) error {
- ne := nsenter.NewNsenter()
- echoArgs := []string{"-c", fmt.Sprintf("cat > %s", filename)}
- glog.V(5).Infof("nsenter: write data to file %s by nsenter", filename)
- command := ne.Exec("sh", echoArgs)
- command.SetStdin(bytes.NewBuffer(data))
- outputBytes, err := command.CombinedOutput()
- if err != nil {
- glog.Errorf("Output from writing to %q: %v", filename, string(outputBytes))
- return err
- }
-
- chmodArgs := []string{fmt.Sprintf("%o", perm), filename}
- glog.V(5).Infof("nsenter: change permissions of file %s to %s", filename, chmodArgs[0])
- outputBytes, err = ne.Exec("chmod", chmodArgs).CombinedOutput()
- if err != nil {
- glog.Errorf("Output from chmod command: %v", string(outputBytes))
- return err
- }
-
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/doc.go b/vendor/k8s.io/kubernetes/pkg/util/mount/doc.go
deleted file mode 100644
index 15179e53f..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/doc.go
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package mount defines an interface to mounting filesystems.
-package mount // import "k8s.io/kubernetes/pkg/util/mount"
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/exec.go b/vendor/k8s.io/kubernetes/pkg/util/mount/exec.go
deleted file mode 100644
index 716cda0a0..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/exec.go
+++ /dev/null
@@ -1,50 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import "k8s.io/utils/exec"
-
-func NewOsExec() Exec {
- return &osExec{}
-}
-
-// Real implementation of Exec interface that uses simple util.Exec
-type osExec struct{}
-
-var _ Exec = &osExec{}
-
-func (e *osExec) Run(cmd string, args ...string) ([]byte, error) {
- exe := exec.New()
- return exe.Command(cmd, args...).CombinedOutput()
-}
-
-func NewFakeExec(run runHook) *FakeExec {
- return &FakeExec{runHook: run}
-}
-
-// Fake for testing.
-type FakeExec struct {
- runHook runHook
-}
-type runHook func(cmd string, args ...string) ([]byte, error)
-
-func (f *FakeExec) Run(cmd string, args ...string) ([]byte, error) {
- if f.runHook != nil {
- return f.runHook(cmd, args...)
- }
- return nil, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount.go b/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount.go
deleted file mode 100644
index b12a2be38..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount.go
+++ /dev/null
@@ -1,140 +0,0 @@
-// +build linux
-
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "fmt"
-
- "github.com/golang/glog"
-)
-
-// ExecMounter is a mounter that uses provided Exec interface to mount and
-// unmount a filesystem. For all other calls it uses a wrapped mounter.
-type execMounter struct {
- wrappedMounter Interface
- exec Exec
-}
-
-func NewExecMounter(exec Exec, wrapped Interface) Interface {
- return &execMounter{
- wrappedMounter: wrapped,
- exec: exec,
- }
-}
-
-// execMounter implements mount.Interface
-var _ Interface = &execMounter{}
-
-// Mount runs mount(8) using given exec interface.
-func (m *execMounter) Mount(source string, target string, fstype string, options []string) error {
- bind, bindRemountOpts := isBind(options)
-
- if bind {
- err := m.doExecMount(source, target, fstype, []string{"bind"})
- if err != nil {
- return err
- }
- return m.doExecMount(source, target, fstype, bindRemountOpts)
- }
-
- return m.doExecMount(source, target, fstype, options)
-}
-
-// doExecMount calls exec(mount <what> <where>) using given exec interface.
-func (m *execMounter) doExecMount(source, target, fstype string, options []string) error {
- glog.V(5).Infof("Exec Mounting %s %s %s %v", source, target, fstype, options)
- mountArgs := makeMountArgs(source, target, fstype, options)
- output, err := m.exec.Run("mount", mountArgs...)
- glog.V(5).Infof("Exec mounted %v: %v: %s", mountArgs, err, string(output))
- if err != nil {
- return fmt.Errorf("mount failed: %v\nMounting command: %s\nMounting arguments: %s %s %s %v\nOutput: %s\n",
- err, "mount", source, target, fstype, options, string(output))
- }
-
- return err
-}
-
-// Unmount runs umount(8) using given exec interface.
-func (m *execMounter) Unmount(target string) error {
- outputBytes, err := m.exec.Run("umount", target)
- if err == nil {
- glog.V(5).Infof("Exec unmounted %s: %s", target, string(outputBytes))
- } else {
- glog.V(5).Infof("Failed to exec unmount %s: err: %q, umount output: %s", target, err, string(outputBytes))
- }
-
- return err
-}
-
-// List returns a list of all mounted filesystems.
-func (m *execMounter) List() ([]MountPoint, error) {
- return m.wrappedMounter.List()
-}
-
-// IsLikelyNotMountPoint determines whether a path is a mountpoint.
-func (m *execMounter) IsLikelyNotMountPoint(file string) (bool, error) {
- return m.wrappedMounter.IsLikelyNotMountPoint(file)
-}
-
-// DeviceOpened checks if block device in use by calling Open with O_EXCL flag.
-// Returns true if open returns errno EBUSY, and false if errno is nil.
-// Returns an error if errno is any error other than EBUSY.
-// Returns with error if pathname is not a device.
-func (m *execMounter) DeviceOpened(pathname string) (bool, error) {
- return m.wrappedMounter.DeviceOpened(pathname)
-}
-
-// PathIsDevice uses FileInfo returned from os.Stat to check if path refers
-// to a device.
-func (m *execMounter) PathIsDevice(pathname string) (bool, error) {
- return m.wrappedMounter.PathIsDevice(pathname)
-}
-
-//GetDeviceNameFromMount given a mount point, find the volume id from checking /proc/mounts
-func (m *execMounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return m.wrappedMounter.GetDeviceNameFromMount(mountPath, pluginDir)
-}
-
-func (m *execMounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return m.wrappedMounter.IsMountPointMatch(mp, dir)
-}
-
-func (m *execMounter) IsNotMountPoint(dir string) (bool, error) {
- return m.wrappedMounter.IsNotMountPoint(dir)
-}
-
-func (m *execMounter) MakeRShared(path string) error {
- return m.wrappedMounter.MakeRShared(path)
-}
-
-func (m *execMounter) GetFileType(pathname string) (FileType, error) {
- return m.wrappedMounter.GetFileType(pathname)
-}
-
-func (m *execMounter) MakeFile(pathname string) error {
- return m.wrappedMounter.MakeFile(pathname)
-}
-
-func (m *execMounter) MakeDir(pathname string) error {
- return m.wrappedMounter.MakeDir(pathname)
-}
-
-func (m *execMounter) ExistsPath(pathname string) bool {
- return m.wrappedMounter.ExistsPath(pathname)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount_unsupported.go b/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount_unsupported.go
deleted file mode 100644
index 136704b23..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/exec_mount_unsupported.go
+++ /dev/null
@@ -1,87 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "errors"
-)
-
-type execMounter struct{}
-
-// ExecMounter is a mounter that uses provided Exec interface to mount and
-// unmount a filesystem. For all other calls it uses a wrapped mounter.
-func NewExecMounter(exec Exec, wrapped Interface) Interface {
- return &execMounter{}
-}
-
-func (mounter *execMounter) Mount(source string, target string, fstype string, options []string) error {
- return nil
-}
-
-func (mounter *execMounter) Unmount(target string) error {
- return nil
-}
-
-func (mounter *execMounter) List() ([]MountPoint, error) {
- return []MountPoint{}, nil
-}
-
-func (mounter *execMounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return (mp.Path == dir)
-}
-
-func (mounter *execMounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(mounter, dir)
-}
-
-func (mounter *execMounter) IsLikelyNotMountPoint(file string) (bool, error) {
- return true, nil
-}
-
-func (mounter *execMounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return "", nil
-}
-
-func (mounter *execMounter) DeviceOpened(pathname string) (bool, error) {
- return false, nil
-}
-
-func (mounter *execMounter) PathIsDevice(pathname string) (bool, error) {
- return true, nil
-}
-
-func (mounter *execMounter) MakeRShared(path string) error {
- return nil
-}
-
-func (mounter *execMounter) GetFileType(pathname string) (FileType, error) {
- return FileType("fake"), errors.New("not implemented")
-}
-
-func (mounter *execMounter) MakeDir(pathname string) error {
- return nil
-}
-
-func (mounter *execMounter) MakeFile(pathname string) error {
- return nil
-}
-
-func (mounter *execMounter) ExistsPath(pathname string) bool {
- return true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/fake.go b/vendor/k8s.io/kubernetes/pkg/util/mount/fake.go
deleted file mode 100644
index f4e2e411d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/fake.go
+++ /dev/null
@@ -1,199 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "os"
- "path/filepath"
- "sync"
-
- "github.com/golang/glog"
-)
-
-// FakeMounter implements mount.Interface for tests.
-type FakeMounter struct {
- MountPoints []MountPoint
- Log []FakeAction
- // Some tests run things in parallel, make sure the mounter does not produce
- // any golang's DATA RACE warnings.
- mutex sync.Mutex
-}
-
-var _ Interface = &FakeMounter{}
-
-// Values for FakeAction.Action
-const FakeActionMount = "mount"
-const FakeActionUnmount = "unmount"
-
-// FakeAction objects are logged every time a fake mount or unmount is called.
-type FakeAction struct {
- Action string // "mount" or "unmount"
- Target string // applies to both mount and unmount actions
- Source string // applies only to "mount" actions
- FSType string // applies only to "mount" actions
-}
-
-func (f *FakeMounter) ResetLog() {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- f.Log = []FakeAction{}
-}
-
-func (f *FakeMounter) Mount(source string, target string, fstype string, options []string) error {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- // find 'bind' option
- for _, option := range options {
- if option == "bind" {
- // This is a bind-mount. In order to mimic linux behaviour, we must
- // use the original device of the bind-mount as the real source.
- // E.g. when mounted /dev/sda like this:
- // $ mount /dev/sda /mnt/test
- // $ mount -o bind /mnt/test /mnt/bound
- // then /proc/mount contains:
- // /dev/sda /mnt/test
- // /dev/sda /mnt/bound
- // (and not /mnt/test /mnt/bound)
- // I.e. we must use /dev/sda as source instead of /mnt/test in the
- // bind mount.
- for _, mnt := range f.MountPoints {
- if source == mnt.Path {
- source = mnt.Device
- break
- }
- }
- break
- }
- }
-
- // If target is a symlink, get its absolute path
- absTarget, err := filepath.EvalSymlinks(target)
- if err != nil {
- absTarget = target
- }
-
- f.MountPoints = append(f.MountPoints, MountPoint{Device: source, Path: absTarget, Type: fstype})
- glog.V(5).Infof("Fake mounter: mounted %s to %s", source, absTarget)
- f.Log = append(f.Log, FakeAction{Action: FakeActionMount, Target: absTarget, Source: source, FSType: fstype})
- return nil
-}
-
-func (f *FakeMounter) Unmount(target string) error {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- // If target is a symlink, get its absolute path
- absTarget, err := filepath.EvalSymlinks(target)
- if err != nil {
- absTarget = target
- }
-
- newMountpoints := []MountPoint{}
- for _, mp := range f.MountPoints {
- if mp.Path == absTarget {
- glog.V(5).Infof("Fake mounter: unmounted %s from %s", mp.Device, absTarget)
- // Don't copy it to newMountpoints
- continue
- }
- newMountpoints = append(newMountpoints, MountPoint{Device: mp.Device, Path: mp.Path, Type: mp.Type})
- }
- f.MountPoints = newMountpoints
- f.Log = append(f.Log, FakeAction{Action: FakeActionUnmount, Target: absTarget})
- return nil
-}
-
-func (f *FakeMounter) List() ([]MountPoint, error) {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- return f.MountPoints, nil
-}
-
-func (f *FakeMounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return mp.Path == dir
-}
-
-func (f *FakeMounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(f, dir)
-}
-
-func (f *FakeMounter) IsLikelyNotMountPoint(file string) (bool, error) {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- _, err := os.Stat(file)
- if err != nil {
- return true, err
- }
-
- // If file is a symlink, get its absolute path
- absFile, err := filepath.EvalSymlinks(file)
- if err != nil {
- absFile = file
- }
-
- for _, mp := range f.MountPoints {
- if mp.Path == absFile {
- glog.V(5).Infof("isLikelyNotMountPoint for %s: mounted %s, false", file, mp.Path)
- return false, nil
- }
- }
- glog.V(5).Infof("isLikelyNotMountPoint for %s: true", file)
- return true, nil
-}
-
-func (f *FakeMounter) DeviceOpened(pathname string) (bool, error) {
- f.mutex.Lock()
- defer f.mutex.Unlock()
-
- for _, mp := range f.MountPoints {
- if mp.Device == pathname {
- return true, nil
- }
- }
- return false, nil
-}
-
-func (f *FakeMounter) PathIsDevice(pathname string) (bool, error) {
- return true, nil
-}
-
-func (f *FakeMounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return getDeviceNameFromMount(f, mountPath, pluginDir)
-}
-
-func (f *FakeMounter) MakeRShared(path string) error {
- return nil
-}
-
-func (f *FakeMounter) GetFileType(pathname string) (FileType, error) {
- return FileType("fake"), nil
-}
-
-func (f *FakeMounter) MakeDir(pathname string) error {
- return nil
-}
-
-func (f *FakeMounter) MakeFile(pathname string) error {
- return nil
-}
-
-func (f *FakeMounter) ExistsPath(pathname string) bool {
- return false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/mount.go b/vendor/k8s.io/kubernetes/pkg/util/mount/mount.go
deleted file mode 100644
index 0f9bac03f..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/mount.go
+++ /dev/null
@@ -1,276 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// TODO(thockin): This whole pkg is pretty linux-centric. As soon as we have
-// an alternate platform, we will need to abstract further.
-package mount
-
-import (
- "os"
- "path/filepath"
- "strings"
-)
-
-type FileType string
-
-const (
- // Default mount command if mounter path is not specified
- defaultMountCommand = "mount"
- MountsInGlobalPDPath = "mounts"
- FileTypeDirectory FileType = "Directory"
- FileTypeFile FileType = "File"
- FileTypeSocket FileType = "Socket"
- FileTypeCharDev FileType = "CharDevice"
- FileTypeBlockDev FileType = "BlockDevice"
-)
-
-type Interface interface {
- // Mount mounts source to target as fstype with given options.
- Mount(source string, target string, fstype string, options []string) error
- // Unmount unmounts given target.
- Unmount(target string) error
- // List returns a list of all mounted filesystems. This can be large.
- // On some platforms, reading mounts is not guaranteed consistent (i.e.
- // it could change between chunked reads). This is guaranteed to be
- // consistent.
- List() ([]MountPoint, error)
- // IsMountPointMatch determines if the mountpoint matches the dir
- IsMountPointMatch(mp MountPoint, dir string) bool
- // IsNotMountPoint determines if a directory is a mountpoint.
- // It should return ErrNotExist when the directory does not exist.
- // IsNotMountPoint is more expensive than IsLikelyNotMountPoint.
- // IsNotMountPoint detects bind mounts in linux.
- // IsNotMountPoint enumerates all the mountpoints using List() and
- // the list of mountpoints may be large, then it uses
- // IsMountPointMatch to evaluate whether the directory is a mountpoint
- IsNotMountPoint(file string) (bool, error)
- // IsLikelyNotMountPoint uses heuristics to determine if a directory
- // is a mountpoint.
- // It should return ErrNotExist when the directory does not exist.
- // IsLikelyNotMountPoint does NOT properly detect all mountpoint types
- // most notably linux bind mounts.
- IsLikelyNotMountPoint(file string) (bool, error)
- // DeviceOpened determines if the device is in use elsewhere
- // on the system, i.e. still mounted.
- DeviceOpened(pathname string) (bool, error)
- // PathIsDevice determines if a path is a device.
- PathIsDevice(pathname string) (bool, error)
- // GetDeviceNameFromMount finds the device name by checking the mount path
- // to get the global mount path which matches its plugin directory
- GetDeviceNameFromMount(mountPath, pluginDir string) (string, error)
- // MakeRShared checks that given path is on a mount with 'rshared' mount
- // propagation. If not, it bind-mounts the path as rshared.
- MakeRShared(path string) error
- // GetFileType checks for file/directory/socket/block/character devices.
- // Will operate in the host mount namespace if kubelet is running in a container
- GetFileType(pathname string) (FileType, error)
- // MakeFile creates an empty file.
- // Will operate in the host mount namespace if kubelet is running in a container
- MakeFile(pathname string) error
- // MakeDir creates a new directory.
- // Will operate in the host mount namespace if kubelet is running in a container
- MakeDir(pathname string) error
- // ExistsPath checks whether the path exists.
- // Will operate in the host mount namespace if kubelet is running in a container
- ExistsPath(pathname string) bool
-}
-
-// Exec executes command where mount utilities are. This can be either the host,
-// container where kubelet runs or even a remote pod with mount utilities.
-// Usual pkg/util/exec interface is not used because kubelet.RunInContainer does
-// not provide stdin/stdout/stderr streams.
-type Exec interface {
- // Run executes a command and returns its stdout + stderr combined in one
- // stream.
- Run(cmd string, args ...string) ([]byte, error)
-}
-
-// Compile-time check to ensure all Mounter implementations satisfy
-// the mount interface
-var _ Interface = &Mounter{}
-
-// This represents a single line in /proc/mounts or /etc/fstab.
-type MountPoint struct {
- Device string
- Path string
- Type string
- Opts []string
- Freq int
- Pass int
-}
-
-// SafeFormatAndMount probes a device to see if it is formatted.
-// Namely it checks to see if a file system is present. If so it
-// mounts it otherwise the device is formatted first then mounted.
-type SafeFormatAndMount struct {
- Interface
- Exec
-}
-
-// FormatAndMount formats the given disk, if needed, and mounts it.
-// That is if the disk is not formatted and it is not being mounted as
-// read-only it will format it first then mount it. Otherwise, if the
-// disk is already formatted or it is being mounted as read-only, it
-// will be mounted without formatting.
-func (mounter *SafeFormatAndMount) FormatAndMount(source string, target string, fstype string, options []string) error {
- return mounter.formatAndMount(source, target, fstype, options)
-}
-
-// GetMountRefsByDev finds all references to the device provided
-// by mountPath; returns a list of paths.
-func GetMountRefsByDev(mounter Interface, mountPath string) ([]string, error) {
- mps, err := mounter.List()
- if err != nil {
- return nil, err
- }
- slTarget, err := filepath.EvalSymlinks(mountPath)
- if err != nil {
- slTarget = mountPath
- }
-
- // Finding the device mounted to mountPath
- diskDev := ""
- for i := range mps {
- if slTarget == mps[i].Path {
- diskDev = mps[i].Device
- break
- }
- }
-
- // Find all references to the device.
- var refs []string
- for i := range mps {
- if mps[i].Device == diskDev || mps[i].Device == slTarget {
- if mps[i].Path != slTarget {
- refs = append(refs, mps[i].Path)
- }
- }
- }
- return refs, nil
-}
-
-// GetDeviceNameFromMount: given a mnt point, find the device from /proc/mounts
-// returns the device name, reference count, and error code
-func GetDeviceNameFromMount(mounter Interface, mountPath string) (string, int, error) {
- mps, err := mounter.List()
- if err != nil {
- return "", 0, err
- }
-
- // Find the device name.
- // FIXME if multiple devices mounted on the same mount path, only the first one is returned
- device := ""
- // If mountPath is symlink, need get its target path.
- slTarget, err := filepath.EvalSymlinks(mountPath)
- if err != nil {
- slTarget = mountPath
- }
- for i := range mps {
- if mps[i].Path == slTarget {
- device = mps[i].Device
- break
- }
- }
-
- // Find all references to the device.
- refCount := 0
- for i := range mps {
- if mps[i].Device == device {
- refCount++
- }
- }
- return device, refCount, nil
-}
-
-// IsNotMountPoint determines if a directory is a mountpoint.
-// It should return ErrNotExist when the directory does not exist.
-// This method uses the List() of all mountpoints
-// It is more extensive than IsLikelyNotMountPoint
-// and it detects bind mounts in linux
-func IsNotMountPoint(mounter Interface, file string) (bool, error) {
- // IsLikelyNotMountPoint provides a quick check
- // to determine whether file IS A mountpoint
- notMnt, notMntErr := mounter.IsLikelyNotMountPoint(file)
- if notMntErr != nil && os.IsPermission(notMntErr) {
- // We were not allowed to do the simple stat() check, e.g. on NFS with
- // root_squash. Fall back to /proc/mounts check below.
- notMnt = true
- notMntErr = nil
- }
- if notMntErr != nil {
- return notMnt, notMntErr
- }
- // identified as mountpoint, so return this fact
- if notMnt == false {
- return notMnt, nil
- }
- // check all mountpoints since IsLikelyNotMountPoint
- // is not reliable for some mountpoint types
- mountPoints, mountPointsErr := mounter.List()
- if mountPointsErr != nil {
- return notMnt, mountPointsErr
- }
- for _, mp := range mountPoints {
- if mounter.IsMountPointMatch(mp, file) {
- notMnt = false
- break
- }
- }
- return notMnt, nil
-}
-
-// isBind detects whether a bind mount is being requested and makes the remount options to
-// use in case of bind mount, due to the fact that bind mount doesn't respect mount options.
-// The list equals:
-// options - 'bind' + 'remount' (no duplicate)
-func isBind(options []string) (bool, []string) {
- bindRemountOpts := []string{"remount"}
- bind := false
-
- if len(options) != 0 {
- for _, option := range options {
- switch option {
- case "bind":
- bind = true
- break
- case "remount":
- break
- default:
- bindRemountOpts = append(bindRemountOpts, option)
- }
- }
- }
-
- return bind, bindRemountOpts
-}
-
-// TODO: this is a workaround for the unmount device issue caused by gci mounter.
-// In GCI cluster, if gci mounter is used for mounting, the container started by mounter
-// script will cause additional mounts created in the container. Since these mounts are
-// irrelevant to the original mounts, they should be not considered when checking the
-// mount references. Current solution is to filter out those mount paths that contain
-// the string of original mount path.
-// Plan to work on better approach to solve this issue.
-
-func HasMountRefs(mountPath string, mountRefs []string) bool {
- count := 0
- for _, ref := range mountRefs {
- if !strings.Contains(ref, mountPath) {
- count = count + 1
- }
- }
- return count > 0
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_linux.go b/vendor/k8s.io/kubernetes/pkg/util/mount/mount_linux.go
deleted file mode 100644
index 11835432d..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_linux.go
+++ /dev/null
@@ -1,700 +0,0 @@
-// +build linux
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "errors"
- "fmt"
- "os"
- "os/exec"
- "path"
- "path/filepath"
- "strconv"
- "strings"
- "syscall"
-
- "github.com/golang/glog"
- "golang.org/x/sys/unix"
- "k8s.io/apimachinery/pkg/util/sets"
- utilio "k8s.io/kubernetes/pkg/util/io"
- utilexec "k8s.io/utils/exec"
-)
-
-const (
- // How many times to retry for a consistent read of /proc/mounts.
- maxListTries = 3
- // Number of fields per line in /proc/mounts as per the fstab man page.
- expectedNumFieldsPerLine = 6
- // Location of the mount file to use
- procMountsPath = "/proc/mounts"
- // Location of the mountinfo file
- procMountInfoPath = "/proc/self/mountinfo"
- // 'fsck' found errors and corrected them
- fsckErrorsCorrected = 1
- // 'fsck' found errors but exited without correcting them
- fsckErrorsUncorrected = 4
-)
-
-// Mounter provides the default implementation of mount.Interface
-// for the linux platform. This implementation assumes that the
-// kubelet is running in the host's root mount namespace.
-type Mounter struct {
- mounterPath string
- withSystemd bool
-}
-
-// New returns a mount.Interface for the current system.
-// It provides options to override the default mounter behavior.
-// mounterPath allows using an alternative to `/bin/mount` for mounting.
-func New(mounterPath string) Interface {
- return &Mounter{
- mounterPath: mounterPath,
- withSystemd: detectSystemd(),
- }
-}
-
-// Mount mounts source to target as fstype with given options. 'source' and 'fstype' must
-// be an empty string in case it's not required, e.g. for remount, or for auto filesystem
-// type, where kernel handles fstype for you. The mount 'options' is a list of options,
-// currently come from mount(8), e.g. "ro", "remount", "bind", etc. If no more option is
-// required, call Mount with an empty string list or nil.
-func (mounter *Mounter) Mount(source string, target string, fstype string, options []string) error {
- // Path to mounter binary if containerized mounter is needed. Otherwise, it is set to empty.
- // All Linux distros are expected to be shipped with a mount utility that a support bind mounts.
- mounterPath := ""
- bind, bindRemountOpts := isBind(options)
- if bind {
- err := mounter.doMount(mounterPath, defaultMountCommand, source, target, fstype, []string{"bind"})
- if err != nil {
- return err
- }
- return mounter.doMount(mounterPath, defaultMountCommand, source, target, fstype, bindRemountOpts)
- }
- // The list of filesystems that require containerized mounter on GCI image cluster
- fsTypesNeedMounter := sets.NewString("nfs", "glusterfs", "ceph", "cifs")
- if fsTypesNeedMounter.Has(fstype) {
- mounterPath = mounter.mounterPath
- }
- return mounter.doMount(mounterPath, defaultMountCommand, source, target, fstype, options)
-}
-
-// doMount runs the mount command. mounterPath is the path to mounter binary if containerized mounter is used.
-func (m *Mounter) doMount(mounterPath string, mountCmd string, source string, target string, fstype string, options []string) error {
- mountArgs := makeMountArgs(source, target, fstype, options)
- if len(mounterPath) > 0 {
- mountArgs = append([]string{mountCmd}, mountArgs...)
- mountCmd = mounterPath
- }
-
- if m.withSystemd {
- // Try to run mount via systemd-run --scope. This will escape the
- // service where kubelet runs and any fuse daemons will be started in a
- // specific scope. kubelet service than can be restarted without killing
- // these fuse daemons.
- //
- // Complete command line (when mounterPath is not used):
- // systemd-run --description=... --scope -- mount -t <type> <what> <where>
- //
- // Expected flow:
- // * systemd-run creates a transient scope (=~ cgroup) and executes its
- // argument (/bin/mount) there.
- // * mount does its job, forks a fuse daemon if necessary and finishes.
- // (systemd-run --scope finishes at this point, returning mount's exit
- // code and stdout/stderr - thats one of --scope benefits).
- // * systemd keeps the fuse daemon running in the scope (i.e. in its own
- // cgroup) until the fuse daemon dies (another --scope benefit).
- // Kubelet service can be restarted and the fuse daemon survives.
- // * When the fuse daemon dies (e.g. during unmount) systemd removes the
- // scope automatically.
- //
- // systemd-mount is not used because it's too new for older distros
- // (CentOS 7, Debian Jessie).
- mountCmd, mountArgs = addSystemdScope("systemd-run", target, mountCmd, mountArgs)
- } else {
- // No systemd-run on the host (or we failed to check it), assume kubelet
- // does not run as a systemd service.
- // No code here, mountCmd and mountArgs are already populated.
- }
-
- glog.V(4).Infof("Mounting cmd (%s) with arguments (%s)", mountCmd, mountArgs)
- command := exec.Command(mountCmd, mountArgs...)
- output, err := command.CombinedOutput()
- if err != nil {
- args := strings.Join(mountArgs, " ")
- glog.Errorf("Mount failed: %v\nMounting command: %s\nMounting arguments: %s\nOutput: %s\n", err, mountCmd, args, string(output))
- return fmt.Errorf("mount failed: %v\nMounting command: %s\nMounting arguments: %s\nOutput: %s\n",
- err, mountCmd, args, string(output))
- }
- return err
-}
-
-// GetMountRefs finds all other references to the device referenced
-// by mountPath; returns a list of paths.
-func GetMountRefs(mounter Interface, mountPath string) ([]string, error) {
- mps, err := mounter.List()
- if err != nil {
- return nil, err
- }
- // Find the device name.
- deviceName := ""
- // If mountPath is symlink, need get its target path.
- slTarget, err := filepath.EvalSymlinks(mountPath)
- if err != nil {
- slTarget = mountPath
- }
- for i := range mps {
- if mps[i].Path == slTarget {
- deviceName = mps[i].Device
- break
- }
- }
-
- // Find all references to the device.
- var refs []string
- if deviceName == "" {
- glog.Warningf("could not determine device for path: %q", mountPath)
- } else {
- for i := range mps {
- if mps[i].Device == deviceName && mps[i].Path != slTarget {
- refs = append(refs, mps[i].Path)
- }
- }
- }
- return refs, nil
-}
-
-// detectSystemd returns true if OS runs with systemd as init. When not sure
-// (permission errors, ...), it returns false.
-// There may be different ways how to detect systemd, this one makes sure that
-// systemd-runs (needed by Mount()) works.
-func detectSystemd() bool {
- if _, err := exec.LookPath("systemd-run"); err != nil {
- glog.V(2).Infof("Detected OS without systemd")
- return false
- }
- // Try to run systemd-run --scope /bin/true, that should be enough
- // to make sure that systemd is really running and not just installed,
- // which happens when running in a container with a systemd-based image
- // but with different pid 1.
- cmd := exec.Command("systemd-run", "--description=Kubernetes systemd probe", "--scope", "true")
- output, err := cmd.CombinedOutput()
- if err != nil {
- glog.V(2).Infof("Cannot run systemd-run, assuming non-systemd OS")
- glog.V(4).Infof("systemd-run failed with: %v", err)
- glog.V(4).Infof("systemd-run output: %s", string(output))
- return false
- }
- glog.V(2).Infof("Detected OS with systemd")
- return true
-}
-
-// makeMountArgs makes the arguments to the mount(8) command.
-func makeMountArgs(source, target, fstype string, options []string) []string {
- // Build mount command as follows:
- // mount [-t $fstype] [-o $options] [$source] $target
- mountArgs := []string{}
- if len(fstype) > 0 {
- mountArgs = append(mountArgs, "-t", fstype)
- }
- if len(options) > 0 {
- mountArgs = append(mountArgs, "-o", strings.Join(options, ","))
- }
- if len(source) > 0 {
- mountArgs = append(mountArgs, source)
- }
- mountArgs = append(mountArgs, target)
-
- return mountArgs
-}
-
-// addSystemdScope adds "system-run --scope" to given command line
-func addSystemdScope(systemdRunPath, mountName, command string, args []string) (string, []string) {
- descriptionArg := fmt.Sprintf("--description=Kubernetes transient mount for %s", mountName)
- systemdRunArgs := []string{descriptionArg, "--scope", "--", command}
- return systemdRunPath, append(systemdRunArgs, args...)
-}
-
-// Unmount unmounts the target.
-func (mounter *Mounter) Unmount(target string) error {
- glog.V(4).Infof("Unmounting %s", target)
- command := exec.Command("umount", target)
- output, err := command.CombinedOutput()
- if err != nil {
- return fmt.Errorf("Unmount failed: %v\nUnmounting arguments: %s\nOutput: %s\n", err, target, string(output))
- }
- return nil
-}
-
-// List returns a list of all mounted filesystems.
-func (*Mounter) List() ([]MountPoint, error) {
- return listProcMounts(procMountsPath)
-}
-
-func (mounter *Mounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- deletedDir := fmt.Sprintf("%s\\040(deleted)", dir)
- return ((mp.Path == dir) || (mp.Path == deletedDir))
-}
-
-func (mounter *Mounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(mounter, dir)
-}
-
-// IsLikelyNotMountPoint determines if a directory is not a mountpoint.
-// It is fast but not necessarily ALWAYS correct. If the path is in fact
-// a bind mount from one part of a mount to another it will not be detected.
-// mkdir /tmp/a /tmp/b; mount --bin /tmp/a /tmp/b; IsLikelyNotMountPoint("/tmp/b")
-// will return true. When in fact /tmp/b is a mount point. If this situation
-// if of interest to you, don't use this function...
-func (mounter *Mounter) IsLikelyNotMountPoint(file string) (bool, error) {
- stat, err := os.Stat(file)
- if err != nil {
- return true, err
- }
- rootStat, err := os.Lstat(filepath.Dir(strings.TrimSuffix(file, "/")))
- if err != nil {
- return true, err
- }
- // If the directory has a different device as parent, then it is a mountpoint.
- if stat.Sys().(*syscall.Stat_t).Dev != rootStat.Sys().(*syscall.Stat_t).Dev {
- return false, nil
- }
-
- return true, nil
-}
-
-// DeviceOpened checks if block device in use by calling Open with O_EXCL flag.
-// If pathname is not a device, log and return false with nil error.
-// If open returns errno EBUSY, return true with nil error.
-// If open returns nil, return false with nil error.
-// Otherwise, return false with error
-func (mounter *Mounter) DeviceOpened(pathname string) (bool, error) {
- return exclusiveOpenFailsOnDevice(pathname)
-}
-
-// PathIsDevice uses FileInfo returned from os.Stat to check if path refers
-// to a device.
-func (mounter *Mounter) PathIsDevice(pathname string) (bool, error) {
- pathType, err := mounter.GetFileType(pathname)
- isDevice := pathType == FileTypeCharDev || pathType == FileTypeBlockDev
- return isDevice, err
-}
-
-func exclusiveOpenFailsOnDevice(pathname string) (bool, error) {
- var isDevice bool
- finfo, err := os.Stat(pathname)
- if os.IsNotExist(err) {
- isDevice = false
- }
- // err in call to os.Stat
- if err != nil {
- return false, fmt.Errorf(
- "PathIsDevice failed for path %q: %v",
- pathname,
- err)
- }
- // path refers to a device
- if finfo.Mode()&os.ModeDevice != 0 {
- isDevice = true
- }
-
- if !isDevice {
- glog.Errorf("Path %q is not referring to a device.", pathname)
- return false, nil
- }
- fd, errno := unix.Open(pathname, unix.O_RDONLY|unix.O_EXCL, 0)
- // If the device is in use, open will return an invalid fd.
- // When this happens, it is expected that Close will fail and throw an error.
- defer unix.Close(fd)
- if errno == nil {
- // device not in use
- return false, nil
- } else if errno == unix.EBUSY {
- // device is in use
- return true, nil
- }
- // error during call to Open
- return false, errno
-}
-
-//GetDeviceNameFromMount: given a mount point, find the device name from its global mount point
-func (mounter *Mounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return getDeviceNameFromMount(mounter, mountPath, pluginDir)
-}
-
-// getDeviceNameFromMount find the device name from /proc/mounts in which
-// the mount path reference should match the given plugin directory. In case no mount path reference
-// matches, returns the volume name taken from its given mountPath
-func getDeviceNameFromMount(mounter Interface, mountPath, pluginDir string) (string, error) {
- refs, err := GetMountRefs(mounter, mountPath)
- if err != nil {
- glog.V(4).Infof("GetMountRefs failed for mount path %q: %v", mountPath, err)
- return "", err
- }
- if len(refs) == 0 {
- glog.V(4).Infof("Directory %s is not mounted", mountPath)
- return "", fmt.Errorf("directory %s is not mounted", mountPath)
- }
- basemountPath := path.Join(pluginDir, MountsInGlobalPDPath)
- for _, ref := range refs {
- if strings.HasPrefix(ref, basemountPath) {
- volumeID, err := filepath.Rel(basemountPath, ref)
- if err != nil {
- glog.Errorf("Failed to get volume id from mount %s - %v", mountPath, err)
- return "", err
- }
- return volumeID, nil
- }
- }
-
- return path.Base(mountPath), nil
-}
-
-func listProcMounts(mountFilePath string) ([]MountPoint, error) {
- content, err := utilio.ConsistentRead(mountFilePath, maxListTries)
- if err != nil {
- return nil, err
- }
- return parseProcMounts(content)
-}
-
-func parseProcMounts(content []byte) ([]MountPoint, error) {
- out := []MountPoint{}
- lines := strings.Split(string(content), "\n")
- for _, line := range lines {
- if line == "" {
- // the last split() item is empty string following the last \n
- continue
- }
- fields := strings.Fields(line)
- if len(fields) != expectedNumFieldsPerLine {
- return nil, fmt.Errorf("wrong number of fields (expected %d, got %d): %s", expectedNumFieldsPerLine, len(fields), line)
- }
-
- mp := MountPoint{
- Device: fields[0],
- Path: fields[1],
- Type: fields[2],
- Opts: strings.Split(fields[3], ","),
- }
-
- freq, err := strconv.Atoi(fields[4])
- if err != nil {
- return nil, err
- }
- mp.Freq = freq
-
- pass, err := strconv.Atoi(fields[5])
- if err != nil {
- return nil, err
- }
- mp.Pass = pass
-
- out = append(out, mp)
- }
- return out, nil
-}
-
-func (mounter *Mounter) MakeRShared(path string) error {
- return doMakeRShared(path, procMountInfoPath)
-}
-
-func (mounter *Mounter) GetFileType(pathname string) (FileType, error) {
- var pathType FileType
- finfo, err := os.Stat(pathname)
- if os.IsNotExist(err) {
- return pathType, fmt.Errorf("path %q does not exist", pathname)
- }
- // err in call to os.Stat
- if err != nil {
- return pathType, err
- }
-
- mode := finfo.Sys().(*syscall.Stat_t).Mode
- switch mode & syscall.S_IFMT {
- case syscall.S_IFSOCK:
- return FileTypeSocket, nil
- case syscall.S_IFBLK:
- return FileTypeBlockDev, nil
- case syscall.S_IFCHR:
- return FileTypeCharDev, nil
- case syscall.S_IFDIR:
- return FileTypeDirectory, nil
- case syscall.S_IFREG:
- return FileTypeFile, nil
- }
-
- return pathType, fmt.Errorf("only recognise file, directory, socket, block device and character device")
-}
-
-func (mounter *Mounter) MakeDir(pathname string) error {
- err := os.MkdirAll(pathname, os.FileMode(0755))
- if err != nil {
- if !os.IsExist(err) {
- return err
- }
- }
- return nil
-}
-
-func (mounter *Mounter) MakeFile(pathname string) error {
- f, err := os.OpenFile(pathname, os.O_CREATE, os.FileMode(0644))
- defer f.Close()
- if err != nil {
- if !os.IsExist(err) {
- return err
- }
- }
- return nil
-}
-
-func (mounter *Mounter) ExistsPath(pathname string) bool {
- _, err := os.Stat(pathname)
- if err != nil {
- return false
- }
- return true
-}
-
-// formatAndMount uses unix utils to format and mount the given disk
-func (mounter *SafeFormatAndMount) formatAndMount(source string, target string, fstype string, options []string) error {
- readOnly := false
- for _, option := range options {
- if option == "ro" {
- readOnly = true
- break
- }
- }
-
- options = append(options, "defaults")
-
- if !readOnly {
- // Run fsck on the disk to fix repairable issues, only do this for volumes requested as rw.
- glog.V(4).Infof("Checking for issues with fsck on disk: %s", source)
- args := []string{"-a", source}
- out, err := mounter.Exec.Run("fsck", args...)
- if err != nil {
- ee, isExitError := err.(utilexec.ExitError)
- switch {
- case err == utilexec.ErrExecutableNotFound:
- glog.Warningf("'fsck' not found on system; continuing mount without running 'fsck'.")
- case isExitError && ee.ExitStatus() == fsckErrorsCorrected:
- glog.Infof("Device %s has errors which were corrected by fsck.", source)
- case isExitError && ee.ExitStatus() == fsckErrorsUncorrected:
- return fmt.Errorf("'fsck' found errors on device %s but could not correct them: %s.", source, string(out))
- case isExitError && ee.ExitStatus() > fsckErrorsUncorrected:
- glog.Infof("`fsck` error %s", string(out))
- }
- }
- }
-
- // Try to mount the disk
- glog.V(4).Infof("Attempting to mount disk: %s %s %s", fstype, source, target)
- mountErr := mounter.Interface.Mount(source, target, fstype, options)
- if mountErr != nil {
- // Mount failed. This indicates either that the disk is unformatted or
- // it contains an unexpected filesystem.
- existingFormat, err := mounter.GetDiskFormat(source)
- if err != nil {
- return err
- }
- if existingFormat == "" {
- if readOnly {
- // Don't attempt to format if mounting as readonly, return an error to reflect this.
- return errors.New("failed to mount unformatted volume as read only")
- }
-
- // Disk is unformatted so format it.
- args := []string{source}
- // Use 'ext4' as the default
- if len(fstype) == 0 {
- fstype = "ext4"
- }
-
- if fstype == "ext4" || fstype == "ext3" {
- args = []string{"-F", source}
- }
- glog.Infof("Disk %q appears to be unformatted, attempting to format as type: %q with options: %v", source, fstype, args)
- _, err := mounter.Exec.Run("mkfs."+fstype, args...)
- if err == nil {
- // the disk has been formatted successfully try to mount it again.
- glog.Infof("Disk successfully formatted (mkfs): %s - %s %s", fstype, source, target)
- return mounter.Interface.Mount(source, target, fstype, options)
- }
- glog.Errorf("format of disk %q failed: type:(%q) target:(%q) options:(%q)error:(%v)", source, fstype, target, options, err)
- return err
- } else {
- // Disk is already formatted and failed to mount
- if len(fstype) == 0 || fstype == existingFormat {
- // This is mount error
- return mountErr
- } else {
- // Block device is formatted with unexpected filesystem, let the user know
- return fmt.Errorf("failed to mount the volume as %q, it already contains %s. Mount error: %v", fstype, existingFormat, mountErr)
- }
- }
- }
- return mountErr
-}
-
-// GetDiskFormat uses 'blkid' to see if the given disk is unformated
-func (mounter *SafeFormatAndMount) GetDiskFormat(disk string) (string, error) {
- args := []string{"-p", "-s", "TYPE", "-s", "PTTYPE", "-o", "export", disk}
- glog.V(4).Infof("Attempting to determine if disk %q is formatted using blkid with args: (%v)", disk, args)
- dataOut, err := mounter.Exec.Run("blkid", args...)
- output := string(dataOut)
- glog.V(4).Infof("Output: %q, err: %v", output, err)
-
- if err != nil {
- if exit, ok := err.(utilexec.ExitError); ok {
- if exit.ExitStatus() == 2 {
- // Disk device is unformatted.
- // For `blkid`, if the specified token (TYPE/PTTYPE, etc) was
- // not found, or no (specified) devices could be identified, an
- // exit code of 2 is returned.
- return "", nil
- }
- }
- glog.Errorf("Could not determine if disk %q is formatted (%v)", disk, err)
- return "", err
- }
-
- var fstype, pttype string
-
- lines := strings.Split(output, "\n")
- for _, l := range lines {
- if len(l) <= 0 {
- // Ignore empty line.
- continue
- }
- cs := strings.Split(l, "=")
- if len(cs) != 2 {
- return "", fmt.Errorf("blkid returns invalid output: %s", output)
- }
- // TYPE is filesystem type, and PTTYPE is partition table type, according
- // to https://www.kernel.org/pub/linux/utils/util-linux/v2.21/libblkid-docs/.
- if cs[0] == "TYPE" {
- fstype = cs[1]
- } else if cs[0] == "PTTYPE" {
- pttype = cs[1]
- }
- }
-
- if len(pttype) > 0 {
- glog.V(4).Infof("Disk %s detected partition table type: %s", pttype)
- // Returns a special non-empty string as filesystem type, then kubelet
- // will not format it.
- return "unknown data, probably partitions", nil
- }
-
- return fstype, nil
-}
-
-// isShared returns true, if given path is on a mount point that has shared
-// mount propagation.
-func isShared(path string, filename string) (bool, error) {
- infos, err := parseMountInfo(filename)
- if err != nil {
- return false, err
- }
-
- // process /proc/xxx/mountinfo in backward order and find the first mount
- // point that is prefix of 'path' - that's the mount where path resides
- var info *mountInfo
- for i := len(infos) - 1; i >= 0; i-- {
- if strings.HasPrefix(path, infos[i].mountPoint) {
- info = &infos[i]
- break
- }
- }
- if info == nil {
- return false, fmt.Errorf("cannot find mount point for %q", path)
- }
-
- // parse optional parameters
- for _, opt := range info.optional {
- if strings.HasPrefix(opt, "shared:") {
- return true, nil
- }
- }
- return false, nil
-}
-
-type mountInfo struct {
- mountPoint string
- // list of "optional parameters", mount propagation is one of them
- optional []string
-}
-
-// parseMountInfo parses /proc/xxx/mountinfo.
-func parseMountInfo(filename string) ([]mountInfo, error) {
- content, err := utilio.ConsistentRead(filename, maxListTries)
- if err != nil {
- return []mountInfo{}, err
- }
- contentStr := string(content)
- infos := []mountInfo{}
-
- for _, line := range strings.Split(contentStr, "\n") {
- if line == "" {
- // the last split() item is empty string following the last \n
- continue
- }
- fields := strings.Fields(line)
- if len(fields) < 7 {
- return nil, fmt.Errorf("wrong number of fields in (expected %d, got %d): %s", 8, len(fields), line)
- }
- info := mountInfo{
- mountPoint: fields[4],
- optional: []string{},
- }
- for i := 6; i < len(fields) && fields[i] != "-"; i++ {
- info.optional = append(info.optional, fields[i])
- }
- infos = append(infos, info)
- }
- return infos, nil
-}
-
-// doMakeRShared is common implementation of MakeRShared on Linux. It checks if
-// path is shared and bind-mounts it as rshared if needed. mountCmd and
-// mountArgs are expected to contain mount-like command, doMakeRShared will add
-// '--bind <path> <path>' and '--make-rshared <path>' to mountArgs.
-func doMakeRShared(path string, mountInfoFilename string) error {
- shared, err := isShared(path, mountInfoFilename)
- if err != nil {
- return err
- }
- if shared {
- glog.V(4).Infof("Directory %s is already on a shared mount", path)
- return nil
- }
-
- glog.V(2).Infof("Bind-mounting %q with shared mount propagation", path)
- // mount --bind /var/lib/kubelet /var/lib/kubelet
- if err := syscall.Mount(path, path, "" /*fstype*/, syscall.MS_BIND, "" /*data*/); err != nil {
- return fmt.Errorf("failed to bind-mount %s: %v", path, err)
- }
-
- // mount --make-rshared /var/lib/kubelet
- if err := syscall.Mount(path, path, "" /*fstype*/, syscall.MS_SHARED|syscall.MS_REC, "" /*data*/); err != nil {
- return fmt.Errorf("failed to make %s rshared: %v", path, err)
- }
-
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_unsupported.go b/vendor/k8s.io/kubernetes/pkg/util/mount/mount_unsupported.go
deleted file mode 100644
index 87d1e3748..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_unsupported.go
+++ /dev/null
@@ -1,110 +0,0 @@
-// +build !linux,!windows
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "errors"
-)
-
-type Mounter struct {
- mounterPath string
-}
-
-// New returns a mount.Interface for the current system.
-// It provides options to override the default mounter behavior.
-// mounterPath allows using an alternative to `/bin/mount` for mounting.
-func New(mounterPath string) Interface {
- return &Mounter{
- mounterPath: mounterPath,
- }
-}
-
-func (mounter *Mounter) Mount(source string, target string, fstype string, options []string) error {
- return nil
-}
-
-func (mounter *Mounter) Unmount(target string) error {
- return nil
-}
-
-// GetMountRefs finds all other references to the device referenced
-// by mountPath; returns a list of paths.
-func GetMountRefs(mounter Interface, mountPath string) ([]string, error) {
- return []string{}, nil
-}
-
-func (mounter *Mounter) List() ([]MountPoint, error) {
- return []MountPoint{}, nil
-}
-
-func (mounter *Mounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return (mp.Path == dir)
-}
-
-func (mounter *Mounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(mounter, dir)
-}
-
-func (mounter *Mounter) IsLikelyNotMountPoint(file string) (bool, error) {
- return true, nil
-}
-
-func (mounter *Mounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return "", nil
-}
-
-func getDeviceNameFromMount(mounter Interface, mountPath, pluginDir string) (string, error) {
- return "", nil
-}
-
-func (mounter *Mounter) DeviceOpened(pathname string) (bool, error) {
- return false, nil
-}
-
-func (mounter *Mounter) PathIsDevice(pathname string) (bool, error) {
- return true, nil
-}
-
-func (mounter *Mounter) MakeRShared(path string) error {
- return nil
-}
-
-func (mounter *SafeFormatAndMount) formatAndMount(source string, target string, fstype string, options []string) error {
- return mounter.Interface.Mount(source, target, fstype, options)
-}
-
-func (mounter *SafeFormatAndMount) diskLooksUnformatted(disk string) (bool, error) {
- return true, nil
-}
-
-func (mounter *Mounter) GetFileType(pathname string) (FileType, error) {
- return FileType("fake"), errors.New("not implemented")
-}
-
-func (mounter *Mounter) MakeDir(pathname string) error {
- return nil
-}
-
-func (mounter *Mounter) MakeFile(pathname string) error {
- return nil
-}
-
-func (mounter *Mounter) ExistsPath(pathname string) bool {
- return true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_windows.go b/vendor/k8s.io/kubernetes/pkg/util/mount/mount_windows.go
deleted file mode 100644
index 0c63626ea..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/mount_windows.go
+++ /dev/null
@@ -1,346 +0,0 @@
-// +build windows
-
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "fmt"
- "os"
- "os/exec"
- "path"
- "path/filepath"
- "strconv"
- "strings"
- "syscall"
-
- "github.com/golang/glog"
-)
-
-// Mounter provides the default implementation of mount.Interface
-// for the windows platform. This implementation assumes that the
-// kubelet is running in the host's root mount namespace.
-type Mounter struct {
- mounterPath string
-}
-
-// New returns a mount.Interface for the current system.
-// It provides options to override the default mounter behavior.
-// mounterPath allows using an alternative to `/bin/mount` for mounting.
-func New(mounterPath string) Interface {
- return &Mounter{
- mounterPath: mounterPath,
- }
-}
-
-// Mount : mounts source to target as NTFS with given options.
-func (mounter *Mounter) Mount(source string, target string, fstype string, options []string) error {
- target = normalizeWindowsPath(target)
-
- if source == "tmpfs" {
- glog.V(3).Infof("azureMount: mounting source (%q), target (%q), with options (%q)", source, target, options)
- return os.MkdirAll(target, 0755)
- }
-
- parentDir := filepath.Dir(target)
- if err := os.MkdirAll(parentDir, 0755); err != nil {
- return err
- }
-
- glog.V(4).Infof("azureMount: mount options(%q) source:%q, target:%q, fstype:%q, begin to mount",
- options, source, target, fstype)
- bindSource := ""
-
- // tell it's going to mount azure disk or azure file according to options
- if bind, _ := isBind(options); bind {
- // mount azure disk
- bindSource = normalizeWindowsPath(source)
- } else {
- if len(options) < 2 {
- glog.Warningf("azureMount: mount options(%q) command number(%d) less than 2, source:%q, target:%q, skip mounting",
- options, len(options), source, target)
- return nil
- }
-
- // currently only cifs mount is supported
- if strings.ToLower(fstype) != "cifs" {
- return fmt.Errorf("azureMount: only cifs mount is supported now, fstype: %q, mounting source (%q), target (%q), with options (%q)", fstype, source, target, options)
- }
-
- cmdLine := fmt.Sprintf(`$User = "%s";$PWord = ConvertTo-SecureString -String "%s" -AsPlainText -Force;`+
- `$Credential = New-Object -TypeName System.Management.Automation.PSCredential -ArgumentList $User, $PWord`,
- options[0], options[1])
-
- bindSource = source
- cmdLine += fmt.Sprintf(";New-SmbGlobalMapping -RemotePath %s -Credential $Credential", source)
-
- if output, err := exec.Command("powershell", "/c", cmdLine).CombinedOutput(); err != nil {
- return fmt.Errorf("azureMount: SmbGlobalMapping failed: %v, only SMB mount is supported now, output: %q", err, string(output))
- }
- }
-
- if output, err := exec.Command("cmd", "/c", "mklink", "/D", target, bindSource).CombinedOutput(); err != nil {
- glog.Errorf("mklink failed: %v, source(%q) target(%q) output: %q", err, bindSource, target, string(output))
- return err
- }
-
- return nil
-}
-
-// Unmount unmounts the target.
-func (mounter *Mounter) Unmount(target string) error {
- glog.V(4).Infof("azureMount: Unmount target (%q)", target)
- target = normalizeWindowsPath(target)
- if output, err := exec.Command("cmd", "/c", "rmdir", target).CombinedOutput(); err != nil {
- glog.Errorf("rmdir failed: %v, output: %q", err, string(output))
- return err
- }
- return nil
-}
-
-// GetMountRefs finds all other references to the device(drive) referenced
-// by mountPath; returns a list of paths.
-func GetMountRefs(mounter Interface, mountPath string) ([]string, error) {
- refs, err := getAllParentLinks(normalizeWindowsPath(mountPath))
- if err != nil {
- return nil, err
- }
- return refs, nil
-}
-
-// List returns a list of all mounted filesystems. todo
-func (mounter *Mounter) List() ([]MountPoint, error) {
- return []MountPoint{}, nil
-}
-
-// IsMountPointMatch determines if the mountpoint matches the dir
-func (mounter *Mounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return mp.Path == dir
-}
-
-// IsNotMountPoint determines if a directory is a mountpoint.
-func (mounter *Mounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(mounter, dir)
-}
-
-// IsLikelyNotMountPoint determines if a directory is not a mountpoint.
-func (mounter *Mounter) IsLikelyNotMountPoint(file string) (bool, error) {
- stat, err := os.Lstat(file)
- if err != nil {
- return true, err
- }
- // If current file is a symlink, then it is a mountpoint.
- if stat.Mode()&os.ModeSymlink != 0 {
- return false, nil
- }
-
- return true, nil
-}
-
-// GetDeviceNameFromMount given a mnt point, find the device
-func (mounter *Mounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return getDeviceNameFromMount(mounter, mountPath, pluginDir)
-}
-
-// getDeviceNameFromMount find the device(drive) name in which
-// the mount path reference should match the given plugin directory. In case no mount path reference
-// matches, returns the volume name taken from its given mountPath
-func getDeviceNameFromMount(mounter Interface, mountPath, pluginDir string) (string, error) {
- refs, err := GetMountRefs(mounter, mountPath)
- if err != nil {
- glog.V(4).Infof("GetMountRefs failed for mount path %q: %v", mountPath, err)
- return "", err
- }
- if len(refs) == 0 {
- return "", fmt.Errorf("directory %s is not mounted", mountPath)
- }
- basemountPath := normalizeWindowsPath(path.Join(pluginDir, MountsInGlobalPDPath))
- for _, ref := range refs {
- if strings.Contains(ref, basemountPath) {
- volumeID, err := filepath.Rel(normalizeWindowsPath(basemountPath), ref)
- if err != nil {
- glog.Errorf("Failed to get volume id from mount %s - %v", mountPath, err)
- return "", err
- }
- return volumeID, nil
- }
- }
-
- return path.Base(mountPath), nil
-}
-
-// DeviceOpened determines if the device is in use elsewhere
-func (mounter *Mounter) DeviceOpened(pathname string) (bool, error) {
- return false, nil
-}
-
-// PathIsDevice determines if a path is a device.
-func (mounter *Mounter) PathIsDevice(pathname string) (bool, error) {
- return false, nil
-}
-
-// MakeRShared checks that given path is on a mount with 'rshared' mount
-// propagation. Empty implementation here.
-func (mounter *Mounter) MakeRShared(path string) error {
- return nil
-}
-
-// GetFileType checks for sockets/block/character devices
-func (mounter *Mounter) GetFileType(pathname string) (FileType, error) {
- var pathType FileType
- info, err := os.Stat(pathname)
- if os.IsNotExist(err) {
- return pathType, fmt.Errorf("path %q does not exist", pathname)
- }
- // err in call to os.Stat
- if err != nil {
- return pathType, err
- }
-
- mode := info.Sys().(*syscall.Win32FileAttributeData).FileAttributes
- switch mode & syscall.S_IFMT {
- case syscall.S_IFSOCK:
- return FileTypeSocket, nil
- case syscall.S_IFBLK:
- return FileTypeBlockDev, nil
- case syscall.S_IFCHR:
- return FileTypeCharDev, nil
- case syscall.S_IFDIR:
- return FileTypeDirectory, nil
- case syscall.S_IFREG:
- return FileTypeFile, nil
- }
-
- return pathType, fmt.Errorf("only recognise file, directory, socket, block device and character device")
-}
-
-// MakeFile creates a new directory
-func (mounter *Mounter) MakeDir(pathname string) error {
- err := os.MkdirAll(pathname, os.FileMode(0755))
- if err != nil {
- if !os.IsExist(err) {
- return err
- }
- }
- return nil
-}
-
-// MakeFile creates an empty file
-func (mounter *Mounter) MakeFile(pathname string) error {
- f, err := os.OpenFile(pathname, os.O_CREATE, os.FileMode(0644))
- defer f.Close()
- if err != nil {
- if !os.IsExist(err) {
- return err
- }
- }
- return nil
-}
-
-// ExistsPath checks whether the path exists
-func (mounter *Mounter) ExistsPath(pathname string) bool {
- _, err := os.Stat(pathname)
- if err != nil {
- return false
- }
- return true
-}
-
-func (mounter *SafeFormatAndMount) formatAndMount(source string, target string, fstype string, options []string) error {
- // Try to mount the disk
- glog.V(4).Infof("Attempting to formatAndMount disk: %s %s %s", fstype, source, target)
-
- if err := ValidateDiskNumber(source); err != nil {
- glog.Errorf("azureMount: formatAndMount failed, err: %v\n", err)
- return err
- }
-
- driveLetter, err := getDriveLetterByDiskNumber(source, mounter.Exec)
- if err != nil {
- return err
- }
- driverPath := driveLetter + ":"
- target = normalizeWindowsPath(target)
- glog.V(4).Infof("Attempting to formatAndMount disk: %s %s %s", fstype, driverPath, target)
- if output, err := mounter.Exec.Run("cmd", "/c", "mklink", "/D", target, driverPath); err != nil {
- glog.Errorf("mklink failed: %v, output: %q", err, string(output))
- return err
- }
- return nil
-}
-
-func normalizeWindowsPath(path string) string {
- normalizedPath := strings.Replace(path, "/", "\\", -1)
- if strings.HasPrefix(normalizedPath, "\\") {
- normalizedPath = "c:" + normalizedPath
- }
- return normalizedPath
-}
-
-// ValidateDiskNumber : disk number should be a number in [0, 99]
-func ValidateDiskNumber(disk string) error {
- diskNum, err := strconv.Atoi(disk)
- if err != nil {
- return fmt.Errorf("wrong disk number format: %q, err:%v", disk, err)
- }
-
- if diskNum < 0 || diskNum > 99 {
- return fmt.Errorf("disk number out of range: %q", disk)
- }
-
- return nil
-}
-
-// Get drive letter according to windows disk number
-func getDriveLetterByDiskNumber(diskNum string, exec Exec) (string, error) {
- cmd := fmt.Sprintf("(Get-Partition -DiskNumber %s).DriveLetter", diskNum)
- output, err := exec.Run("powershell", "/c", cmd)
- if err != nil {
- return "", fmt.Errorf("azureMount: Get Drive Letter failed: %v, output: %q", err, string(output))
- }
- if len(string(output)) < 1 {
- return "", fmt.Errorf("azureMount: Get Drive Letter failed, output is empty")
- }
- return string(output)[:1], nil
-}
-
-// getAllParentLinks walks all symbolic links and return all the parent targets recursively
-func getAllParentLinks(path string) ([]string, error) {
- const maxIter = 255
- links := []string{}
- for {
- links = append(links, path)
- if len(links) > maxIter {
- return links, fmt.Errorf("unexpected length of parent links: %v", links)
- }
-
- fi, err := os.Lstat(path)
- if err != nil {
- return links, fmt.Errorf("Lstat: %v", err)
- }
- if fi.Mode()&os.ModeSymlink == 0 {
- break
- }
-
- path, err = os.Readlink(path)
- if err != nil {
- return links, fmt.Errorf("Readlink error: %v", err)
- }
- }
-
- return links, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount.go b/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount.go
deleted file mode 100644
index 99e81837f..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount.go
+++ /dev/null
@@ -1,272 +0,0 @@
-// +build linux
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "fmt"
- "os"
- "path/filepath"
- "strings"
-
- "github.com/golang/glog"
- "k8s.io/kubernetes/pkg/util/nsenter"
-)
-
-const (
- // hostProcMountsPath is the default mount path for rootfs
- hostProcMountsPath = "/rootfs/proc/1/mounts"
- // hostProcMountinfoPath is the default mount info path for rootfs
- hostProcMountinfoPath = "/rootfs/proc/1/mountinfo"
-)
-
-// Currently, all docker containers receive their own mount namespaces.
-// NsenterMounter works by executing nsenter to run commands in
-// the host's mount namespace.
-type NsenterMounter struct {
- ne *nsenter.Nsenter
-}
-
-func NewNsenterMounter() *NsenterMounter {
- return &NsenterMounter{ne: nsenter.NewNsenter()}
-}
-
-// NsenterMounter implements mount.Interface
-var _ = Interface(&NsenterMounter{})
-
-// Mount runs mount(8) in the host's root mount namespace. Aside from this
-// aspect, Mount has the same semantics as the mounter returned by mount.New()
-func (n *NsenterMounter) Mount(source string, target string, fstype string, options []string) error {
- bind, bindRemountOpts := isBind(options)
-
- if bind {
- err := n.doNsenterMount(source, target, fstype, []string{"bind"})
- if err != nil {
- return err
- }
- return n.doNsenterMount(source, target, fstype, bindRemountOpts)
- }
-
- return n.doNsenterMount(source, target, fstype, options)
-}
-
-// doNsenterMount nsenters the host's mount namespace and performs the
-// requested mount.
-func (n *NsenterMounter) doNsenterMount(source, target, fstype string, options []string) error {
- glog.V(5).Infof("nsenter mount %s %s %s %v", source, target, fstype, options)
- cmd, args := n.makeNsenterArgs(source, target, fstype, options)
- outputBytes, err := n.ne.Exec(cmd, args).CombinedOutput()
- if len(outputBytes) != 0 {
- glog.V(5).Infof("Output of mounting %s to %s: %v", source, target, string(outputBytes))
- }
- return err
-}
-
-// makeNsenterArgs makes a list of argument to nsenter in order to do the
-// requested mount.
-func (n *NsenterMounter) makeNsenterArgs(source, target, fstype string, options []string) (string, []string) {
- mountCmd := n.ne.AbsHostPath("mount")
- mountArgs := makeMountArgs(source, target, fstype, options)
-
- if systemdRunPath, hasSystemd := n.ne.SupportsSystemd(); hasSystemd {
- // Complete command line:
- // nsenter --mount=/rootfs/proc/1/ns/mnt -- /bin/systemd-run --description=... --scope -- /bin/mount -t <type> <what> <where>
- // Expected flow is:
- // * nsenter breaks out of container's mount namespace and executes
- // host's systemd-run.
- // * systemd-run creates a transient scope (=~ cgroup) and executes its
- // argument (/bin/mount) there.
- // * mount does its job, forks a fuse daemon if necessary and finishes.
- // (systemd-run --scope finishes at this point, returning mount's exit
- // code and stdout/stderr - thats one of --scope benefits).
- // * systemd keeps the fuse daemon running in the scope (i.e. in its own
- // cgroup) until the fuse daemon dies (another --scope benefit).
- // Kubelet container can be restarted and the fuse daemon survives.
- // * When the daemon dies (e.g. during unmount) systemd removes the
- // scope automatically.
- mountCmd, mountArgs = addSystemdScope(systemdRunPath, target, mountCmd, mountArgs)
- } else {
- // Fall back to simple mount when the host has no systemd.
- // Complete command line:
- // nsenter --mount=/rootfs/proc/1/ns/mnt -- /bin/mount -t <type> <what> <where>
- // Expected flow is:
- // * nsenter breaks out of container's mount namespace and executes host's /bin/mount.
- // * mount does its job, forks a fuse daemon if necessary and finishes.
- // * Any fuse daemon runs in cgroup of kubelet docker container,
- // restart of kubelet container will kill it!
-
- // No code here, mountCmd and mountArgs use /bin/mount
- }
-
- return mountCmd, mountArgs
-}
-
-// Unmount runs umount(8) in the host's mount namespace.
-func (n *NsenterMounter) Unmount(target string) error {
- args := []string{target}
- // No need to execute systemd-run here, it's enough that unmount is executed
- // in the host's mount namespace. It will finish appropriate fuse daemon(s)
- // running in any scope.
- glog.V(5).Infof("nsenter unmount args: %v", args)
- outputBytes, err := n.ne.Exec("umount", args).CombinedOutput()
- if len(outputBytes) != 0 {
- glog.V(5).Infof("Output of unmounting %s: %v", target, string(outputBytes))
- }
- return err
-}
-
-// List returns a list of all mounted filesystems in the host's mount namespace.
-func (*NsenterMounter) List() ([]MountPoint, error) {
- return listProcMounts(hostProcMountsPath)
-}
-
-func (m *NsenterMounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(m, dir)
-}
-
-func (*NsenterMounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- deletedDir := fmt.Sprintf("%s\\040(deleted)", dir)
- return (mp.Path == dir) || (mp.Path == deletedDir)
-}
-
-// IsLikelyNotMountPoint determines whether a path is a mountpoint by calling findmnt
-// in the host's root mount namespace.
-func (n *NsenterMounter) IsLikelyNotMountPoint(file string) (bool, error) {
- file, err := filepath.Abs(file)
- if err != nil {
- return true, err
- }
-
- // Check the directory exists
- if _, err = os.Stat(file); os.IsNotExist(err) {
- glog.V(5).Infof("findmnt: directory %s does not exist", file)
- return true, err
- }
- // Add --first-only option: since we are testing for the absence of a mountpoint, it is sufficient to get only
- // the first of multiple possible mountpoints using --first-only.
- // Also add fstype output to make sure that the output of target file will give the full path
- // TODO: Need more refactoring for this function. Track the solution with issue #26996
- args := []string{"-o", "target,fstype", "--noheadings", "--first-only", "--target", file}
- glog.V(5).Infof("nsenter findmnt args: %v", args)
- out, err := n.ne.Exec("findmnt", args).CombinedOutput()
- if err != nil {
- glog.V(2).Infof("Failed findmnt command for path %s: %s %v", file, out, err)
- // Different operating systems behave differently for paths which are not mount points.
- // On older versions (e.g. 2.20.1) we'd get error, on newer ones (e.g. 2.26.2) we'd get "/".
- // It's safer to assume that it's not a mount point.
- return true, nil
- }
- mountTarget, err := parseFindMnt(string(out))
- if err != nil {
- return false, err
- }
-
- glog.V(5).Infof("IsLikelyNotMountPoint findmnt output for path %s: %v:", file, mountTarget)
-
- if mountTarget == file {
- glog.V(5).Infof("IsLikelyNotMountPoint: %s is a mount point", file)
- return false, nil
- }
- glog.V(5).Infof("IsLikelyNotMountPoint: %s is not a mount point", file)
- return true, nil
-}
-
-// parse output of "findmnt -o target,fstype" and return just the target
-func parseFindMnt(out string) (string, error) {
- // cut trailing newline
- out = strings.TrimSuffix(out, "\n")
- // cut everything after the last space - it's the filesystem type
- i := strings.LastIndex(out, " ")
- if i == -1 {
- return "", fmt.Errorf("error parsing findmnt output, expected at least one space: %q", out)
- }
- return out[:i], nil
-}
-
-// DeviceOpened checks if block device in use by calling Open with O_EXCL flag.
-// Returns true if open returns errno EBUSY, and false if errno is nil.
-// Returns an error if errno is any error other than EBUSY.
-// Returns with error if pathname is not a device.
-func (n *NsenterMounter) DeviceOpened(pathname string) (bool, error) {
- return exclusiveOpenFailsOnDevice(pathname)
-}
-
-// PathIsDevice uses FileInfo returned from os.Stat to check if path refers
-// to a device.
-func (n *NsenterMounter) PathIsDevice(pathname string) (bool, error) {
- pathType, err := n.GetFileType(pathname)
- isDevice := pathType == FileTypeCharDev || pathType == FileTypeBlockDev
- return isDevice, err
-}
-
-//GetDeviceNameFromMount given a mount point, find the volume id from checking /proc/mounts
-func (n *NsenterMounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return getDeviceNameFromMount(n, mountPath, pluginDir)
-}
-
-func (n *NsenterMounter) MakeRShared(path string) error {
- return doMakeRShared(path, hostProcMountinfoPath)
-}
-
-func (mounter *NsenterMounter) GetFileType(pathname string) (FileType, error) {
- var pathType FileType
- outputBytes, err := mounter.ne.Exec("stat", []string{"-L", `--printf "%F"`, pathname}).CombinedOutput()
- if err != nil {
- return pathType, err
- }
-
- switch string(outputBytes) {
- case "socket":
- return FileTypeSocket, nil
- case "character special file":
- return FileTypeCharDev, nil
- case "block special file":
- return FileTypeBlockDev, nil
- case "directory":
- return FileTypeDirectory, nil
- case "regular file":
- return FileTypeFile, nil
- }
-
- return pathType, fmt.Errorf("only recognise file, directory, socket, block device and character device")
-}
-
-func (mounter *NsenterMounter) MakeDir(pathname string) error {
- args := []string{"-p", pathname}
- if _, err := mounter.ne.Exec("mkdir", args).CombinedOutput(); err != nil {
- return err
- }
- return nil
-}
-
-func (mounter *NsenterMounter) MakeFile(pathname string) error {
- args := []string{pathname}
- if _, err := mounter.ne.Exec("touch", args).CombinedOutput(); err != nil {
- return err
- }
- return nil
-}
-
-func (mounter *NsenterMounter) ExistsPath(pathname string) bool {
- args := []string{pathname}
- _, err := mounter.ne.Exec("ls", args).CombinedOutput()
- if err == nil {
- return true
- }
- return false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount_unsupported.go b/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount_unsupported.go
deleted file mode 100644
index f4eb692f9..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/mount/nsenter_mount_unsupported.go
+++ /dev/null
@@ -1,87 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package mount
-
-import (
- "errors"
-)
-
-type NsenterMounter struct{}
-
-func NewNsenterMounter() *NsenterMounter {
- return &NsenterMounter{}
-}
-
-var _ = Interface(&NsenterMounter{})
-
-func (*NsenterMounter) Mount(source string, target string, fstype string, options []string) error {
- return nil
-}
-
-func (*NsenterMounter) Unmount(target string) error {
- return nil
-}
-
-func (*NsenterMounter) List() ([]MountPoint, error) {
- return []MountPoint{}, nil
-}
-
-func (m *NsenterMounter) IsNotMountPoint(dir string) (bool, error) {
- return IsNotMountPoint(m, dir)
-}
-
-func (*NsenterMounter) IsMountPointMatch(mp MountPoint, dir string) bool {
- return (mp.Path == dir)
-}
-
-func (*NsenterMounter) IsLikelyNotMountPoint(file string) (bool, error) {
- return true, nil
-}
-
-func (*NsenterMounter) DeviceOpened(pathname string) (bool, error) {
- return false, nil
-}
-
-func (*NsenterMounter) PathIsDevice(pathname string) (bool, error) {
- return true, nil
-}
-
-func (*NsenterMounter) GetDeviceNameFromMount(mountPath, pluginDir string) (string, error) {
- return "", nil
-}
-
-func (*NsenterMounter) MakeRShared(path string) error {
- return nil
-}
-
-func (*NsenterMounter) GetFileType(_ string) (FileType, error) {
- return FileType("fake"), errors.New("not implemented")
-}
-
-func (*NsenterMounter) MakeDir(pathname string) error {
- return nil
-}
-
-func (*NsenterMounter) MakeFile(pathname string) error {
- return nil
-}
-
-func (*NsenterMounter) ExistsPath(pathname string) bool {
- return true
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/net/sets/doc.go b/vendor/k8s.io/kubernetes/pkg/util/net/sets/doc.go
deleted file mode 100644
index 8414f74ac..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/net/sets/doc.go
+++ /dev/null
@@ -1,28 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// This package contains hand-coded set implementations that should be similar
-// to the autogenerated ones in pkg/util/sets.
-// We can't simply use net.IPNet as a map-key in Go (because it contains a
-// []byte).
-// We could use the same workaround we use here (a string representation as the
-// key) to autogenerate sets. If we do that, or decide on an alternate
-// approach, we should replace the implementations in this package with the
-// autogenerated versions.
-// It is expected that callers will alias this import as "netsets" i.e. import
-// netsets "k8s.io/kubernetes/pkg/util/net/sets"
-
-package sets
diff --git a/vendor/k8s.io/kubernetes/pkg/util/net/sets/ipnet.go b/vendor/k8s.io/kubernetes/pkg/util/net/sets/ipnet.go
deleted file mode 100644
index 90ad58c63..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/net/sets/ipnet.go
+++ /dev/null
@@ -1,121 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package sets
-
-import (
- "net"
- "strings"
-)
-
-// IPNet maps string to net.IPNet.
-type IPNet map[string]*net.IPNet
-
-// ParseIPNets parses string slice to IPNet.
-func ParseIPNets(specs ...string) (IPNet, error) {
- ipnetset := make(IPNet)
- for _, spec := range specs {
- spec = strings.TrimSpace(spec)
- _, ipnet, err := net.ParseCIDR(spec)
- if err != nil {
- return nil, err
- }
- k := ipnet.String() // In case of normalization
- ipnetset[k] = ipnet
- }
- return ipnetset, nil
-}
-
-// Insert adds items to the set.
-func (s IPNet) Insert(items ...*net.IPNet) {
- for _, item := range items {
- s[item.String()] = item
- }
-}
-
-// Delete removes all items from the set.
-func (s IPNet) Delete(items ...*net.IPNet) {
- for _, item := range items {
- delete(s, item.String())
- }
-}
-
-// Has returns true if and only if item is contained in the set.
-func (s IPNet) Has(item *net.IPNet) bool {
- _, contained := s[item.String()]
- return contained
-}
-
-// HasAll returns true if and only if all items are contained in the set.
-func (s IPNet) HasAll(items ...*net.IPNet) bool {
- for _, item := range items {
- if !s.Has(item) {
- return false
- }
- }
- return true
-}
-
-// Difference returns a set of objects that are not in s2
-// For example:
-// s1 = {a1, a2, a3}
-// s2 = {a1, a2, a4, a5}
-// s1.Difference(s2) = {a3}
-// s2.Difference(s1) = {a4, a5}
-func (s IPNet) Difference(s2 IPNet) IPNet {
- result := make(IPNet)
- for k, i := range s {
- _, found := s2[k]
- if found {
- continue
- }
- result[k] = i
- }
- return result
-}
-
-// StringSlice returns a []string with the String representation of each element in the set.
-// Order is undefined.
-func (s IPNet) StringSlice() []string {
- a := make([]string, 0, len(s))
- for k := range s {
- a = append(a, k)
- }
- return a
-}
-
-// IsSuperset returns true if and only if s1 is a superset of s2.
-func (s IPNet) IsSuperset(s2 IPNet) bool {
- for k := range s2 {
- _, found := s[k]
- if !found {
- return false
- }
- }
- return true
-}
-
-// Equal returns true if and only if s1 is equal (as a set) to s2.
-// Two sets are equal if their membership is identical.
-// (In practice, this means same elements, order doesn't matter)
-func (s IPNet) Equal(s2 IPNet) bool {
- return len(s) == len(s2) && s.IsSuperset(s2)
-}
-
-// Len returns the size of the set.
-func (s IPNet) Len() int {
- return len(s)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter.go b/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter.go
deleted file mode 100644
index 32fbc0848..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter.go
+++ /dev/null
@@ -1,124 +0,0 @@
-// +build linux
-
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package nsenter
-
-import (
- "fmt"
- "os"
- "path/filepath"
-
- "k8s.io/utils/exec"
-
- "github.com/golang/glog"
-)
-
-const (
- hostRootFsPath = "/rootfs"
- // hostProcMountNsPath is the default mount namespace for rootfs
- hostProcMountNsPath = "/rootfs/proc/1/ns/mnt"
- // nsenterPath is the default nsenter command
- nsenterPath = "nsenter"
-)
-
-// Nsenter is part of experimental support for running the kubelet
-// in a container.
-//
-// Nsenter requires:
-//
-// 1. Docker >= 1.6 due to the dependency on the slave propagation mode
-// of the bind-mount of the kubelet root directory in the container.
-// Docker 1.5 used a private propagation mode for bind-mounts, so mounts
-// performed in the host's mount namespace do not propagate out to the
-// bind-mount in this docker version.
-// 2. The host's root filesystem must be available at /rootfs
-// 3. The nsenter binary must be on the Kubelet process' PATH in the container's
-// filesystem.
-// 4. The Kubelet process must have CAP_SYS_ADMIN (required by nsenter); at
-// the present, this effectively means that the kubelet is running in a
-// privileged container.
-// 5. The volume path used by the Kubelet must be the same inside and outside
-// the container and be writable by the container (to initialize volume)
-// contents. TODO: remove this requirement.
-// 6. The host image must have "mount", "findmnt", "umount", "stat", "touch",
-// "mkdir", "ls", "sh" and "chmod" binaries in /bin, /usr/sbin, or /usr/bin
-// 7. The host image should have systemd-run in /bin, /usr/sbin, or /usr/bin
-// For more information about mount propagation modes, see:
-// https://www.kernel.org/doc/Documentation/filesystems/sharedsubtree.txt
-type Nsenter struct {
- // a map of commands to their paths on the host filesystem
- paths map[string]string
-}
-
-// NewNsenter constructs a new instance of Nsenter
-func NewNsenter() *Nsenter {
- ne := &Nsenter{
- paths: map[string]string{
- "mount": "",
- "findmnt": "",
- "umount": "",
- "systemd-run": "",
- "stat": "",
- "touch": "",
- "mkdir": "",
- "ls": "",
- "sh": "",
- "chmod": "",
- },
- }
- // search for the required commands in other locations besides /usr/bin
- for binary := range ne.paths {
- // default to root
- ne.paths[binary] = filepath.Join("/", binary)
- for _, path := range []string{"/bin", "/usr/sbin", "/usr/bin"} {
- binPath := filepath.Join(path, binary)
- if _, err := os.Stat(filepath.Join(hostRootFsPath, binPath)); err != nil {
- continue
- }
- ne.paths[binary] = binPath
- break
- }
- // TODO: error, so that the kubelet can stop if the paths don't exist
- // (don't forget that systemd-run is optional)
- }
- return ne
-}
-
-// Exec executes nsenter commands in hostProcMountNsPath mount namespace
-func (ne *Nsenter) Exec(cmd string, args []string) exec.Cmd {
- fullArgs := append([]string{fmt.Sprintf("--mount=%s", hostProcMountNsPath), "--"},
- append([]string{ne.AbsHostPath(cmd)}, args...)...)
- glog.V(5).Infof("Running nsenter command: %v %v", nsenterPath, fullArgs)
- exec := exec.New()
- return exec.Command(nsenterPath, fullArgs...)
-}
-
-// AbsHostPath returns the absolute runnable path for a specified command
-func (ne *Nsenter) AbsHostPath(command string) string {
- path, ok := ne.paths[command]
- if !ok {
- return command
- }
- return path
-}
-
-// SupportsSystemd checks whether command systemd-run exists
-func (ne *Nsenter) SupportsSystemd() (string, bool) {
- systemdRunPath, hasSystemd := ne.paths["systemd-run"]
- return systemdRunPath, hasSystemd
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter_unsupported.go b/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter_unsupported.go
deleted file mode 100644
index 9c2130cb6..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/nsenter/nsenter_unsupported.go
+++ /dev/null
@@ -1,50 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package nsenter
-
-import (
- "k8s.io/utils/exec"
-)
-
-// Nsenter is part of experimental support for running the kubelet
-// in a container.
-type Nsenter struct {
- // a map of commands to their paths on the host filesystem
- Paths map[string]string
-}
-
-// NewNsenter constructs a new instance of Nsenter
-func NewNsenter() *Nsenter {
- return &Nsenter{}
-}
-
-// Exec executes nsenter commands in hostProcMountNsPath mount namespace
-func (ne *Nsenter) Exec(cmd string, args []string) exec.Cmd {
- return nil
-}
-
-// AbsHostPath returns the absolute runnable path for a specified command
-func (ne *Nsenter) AbsHostPath(command string) string {
- return ""
-}
-
-// SupportsSystemd checks whether command systemd-run exists
-func (ne *Nsenter) SupportsSystemd() (string, bool) {
- return "", false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/parsers/parsers.go b/vendor/k8s.io/kubernetes/pkg/util/parsers/parsers.go
deleted file mode 100644
index be35da7cc..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/parsers/parsers.go
+++ /dev/null
@@ -1,58 +0,0 @@
-/*
-Copyright 2015 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package parsers
-
-import (
- "fmt"
- // Import the crypto sha256 algorithm for the docker image parser to work
- _ "crypto/sha256"
- // Import the crypto/sha512 algorithm for the docker image parser to work with 384 and 512 sha hashes
- _ "crypto/sha512"
-
- dockerref "github.com/docker/distribution/reference"
-)
-
-const (
- DefaultImageTag = "latest"
-)
-
-// ParseImageName parses a docker image string into three parts: repo, tag and digest.
-// If both tag and digest are empty, a default image tag will be returned.
-func ParseImageName(image string) (string, string, string, error) {
- named, err := dockerref.ParseNormalizedNamed(image)
- if err != nil {
- return "", "", "", fmt.Errorf("couldn't parse image name: %v", err)
- }
-
- repoToPull := named.Name()
- var tag, digest string
-
- tagged, ok := named.(dockerref.Tagged)
- if ok {
- tag = tagged.Tag()
- }
-
- digested, ok := named.(dockerref.Digested)
- if ok {
- digest = digested.Digest().String()
- }
- // If no tag was specified, use the default "latest".
- if len(tag) == 0 && len(digest) == 0 {
- tag = DefaultImageTag
- }
- return repoToPull, tag, digest, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/pointer/pointer.go b/vendor/k8s.io/kubernetes/pkg/util/pointer/pointer.go
deleted file mode 100644
index a970bf7f5..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/pointer/pointer.go
+++ /dev/null
@@ -1,68 +0,0 @@
-/*
-Copyright 2017 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package pointer
-
-import (
- "fmt"
- "reflect"
-)
-
-// AllPtrFieldsNil tests whether all pointer fields in a struct are nil. This is useful when,
-// for example, an API struct is handled by plugins which need to distinguish
-// "no plugin accepted this spec" from "this spec is empty".
-//
-// This function is only valid for structs and pointers to structs. Any other
-// type will cause a panic. Passing a typed nil pointer will return true.
-func AllPtrFieldsNil(obj interface{}) bool {
- v := reflect.ValueOf(obj)
- if !v.IsValid() {
- panic(fmt.Sprintf("reflect.ValueOf() produced a non-valid Value for %#v", obj))
- }
- if v.Kind() == reflect.Ptr {
- if v.IsNil() {
- return true
- }
- v = v.Elem()
- }
- for i := 0; i < v.NumField(); i++ {
- if v.Field(i).Kind() == reflect.Ptr && !v.Field(i).IsNil() {
- return false
- }
- }
- return true
-}
-
-// Int32Ptr returns a pointer to an int32
-func Int32Ptr(i int32) *int32 {
- o := i
- return &o
-}
-
-// Int32PtrDerefOr dereference the int32 ptr and returns it i not nil,
-// else returns def.
-func Int32PtrDerefOr(ptr *int32, def int32) int32 {
- if ptr != nil {
- return *ptr
- }
- return def
-}
-
-// BoolPtr returns a pointer to a bool
-func BoolPtr(b bool) *bool {
- o := b
- return &o
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/util/taints/taints.go b/vendor/k8s.io/kubernetes/pkg/util/taints/taints.go
deleted file mode 100644
index 76e4bb866..000000000
--- a/vendor/k8s.io/kubernetes/pkg/util/taints/taints.go
+++ /dev/null
@@ -1,342 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// package taints implements utilites for working with taints
-package taints
-
-import (
- "fmt"
- "strings"
-
- "k8s.io/api/core/v1"
- utilerrors "k8s.io/apimachinery/pkg/util/errors"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/validation"
- api "k8s.io/kubernetes/pkg/apis/core"
- "k8s.io/kubernetes/pkg/apis/core/helper"
-)
-
-const (
- MODIFIED = "modified"
- TAINTED = "tainted"
- UNTAINTED = "untainted"
-)
-
-// parseTaint parses a taint from a string. Taint must be of the format '<key>=<value>:<effect>'.
-func parseTaint(st string) (v1.Taint, error) {
- var taint v1.Taint
- parts := strings.Split(st, "=")
- if len(parts) != 2 || len(parts[1]) == 0 || len(validation.IsQualifiedName(parts[0])) > 0 {
- return taint, fmt.Errorf("invalid taint spec: %v", st)
- }
-
- parts2 := strings.Split(parts[1], ":")
-
- errs := validation.IsValidLabelValue(parts2[0])
- if len(parts2) != 2 || len(errs) != 0 {
- return taint, fmt.Errorf("invalid taint spec: %v, %s", st, strings.Join(errs, "; "))
- }
-
- effect := v1.TaintEffect(parts2[1])
- if err := validateTaintEffect(effect); err != nil {
- return taint, err
- }
-
- taint.Key = parts[0]
- taint.Value = parts2[0]
- taint.Effect = effect
-
- return taint, nil
-}
-
-func validateTaintEffect(effect v1.TaintEffect) error {
- if effect != v1.TaintEffectNoSchedule && effect != v1.TaintEffectPreferNoSchedule && effect != v1.TaintEffectNoExecute {
- return fmt.Errorf("invalid taint effect: %v, unsupported taint effect", effect)
- }
-
- return nil
-}
-
-// NewTaintsVar wraps []api.Taint in a struct that implements flag.Value to allow taints to be
-// bound to command line flags.
-func NewTaintsVar(ptr *[]api.Taint) taintsVar {
- return taintsVar{
- ptr: ptr,
- }
-}
-
-type taintsVar struct {
- ptr *[]api.Taint
-}
-
-func (t taintsVar) Set(s string) error {
- if len(s) == 0 {
- *t.ptr = nil
- return nil
- }
- sts := strings.Split(s, ",")
- var taints []api.Taint
- for _, st := range sts {
- taint, err := parseTaint(st)
- if err != nil {
- return err
- }
- taints = append(taints, api.Taint{Key: taint.Key, Value: taint.Value, Effect: api.TaintEffect(taint.Effect)})
- }
- *t.ptr = taints
- return nil
-}
-
-func (t taintsVar) String() string {
- if len(*t.ptr) == 0 {
- return ""
- }
- var taints []string
- for _, taint := range *t.ptr {
- taints = append(taints, fmt.Sprintf("%s=%s:%s", taint.Key, taint.Value, taint.Effect))
- }
- return strings.Join(taints, ",")
-}
-
-func (t taintsVar) Type() string {
- return "[]api.Taint"
-}
-
-// ParseTaints takes a spec which is an array and creates slices for new taints to be added, taints to be deleted.
-func ParseTaints(spec []string) ([]v1.Taint, []v1.Taint, error) {
- var taints, taintsToRemove []v1.Taint
- uniqueTaints := map[v1.TaintEffect]sets.String{}
-
- for _, taintSpec := range spec {
- if strings.Index(taintSpec, "=") != -1 && strings.Index(taintSpec, ":") != -1 {
- newTaint, err := parseTaint(taintSpec)
- if err != nil {
- return nil, nil, err
- }
- // validate if taint is unique by <key, effect>
- if len(uniqueTaints[newTaint.Effect]) > 0 && uniqueTaints[newTaint.Effect].Has(newTaint.Key) {
- return nil, nil, fmt.Errorf("duplicated taints with the same key and effect: %v", newTaint)
- }
- // add taint to existingTaints for uniqueness check
- if len(uniqueTaints[newTaint.Effect]) == 0 {
- uniqueTaints[newTaint.Effect] = sets.String{}
- }
- uniqueTaints[newTaint.Effect].Insert(newTaint.Key)
-
- taints = append(taints, newTaint)
- } else if strings.HasSuffix(taintSpec, "-") {
- taintKey := taintSpec[:len(taintSpec)-1]
- var effect v1.TaintEffect
- if strings.Index(taintKey, ":") != -1 {
- parts := strings.Split(taintKey, ":")
- taintKey = parts[0]
- effect = v1.TaintEffect(parts[1])
- }
-
- // If effect is specified, need to validate it.
- if len(effect) > 0 {
- err := validateTaintEffect(effect)
- if err != nil {
- return nil, nil, err
- }
- }
- taintsToRemove = append(taintsToRemove, v1.Taint{Key: taintKey, Effect: effect})
- } else {
- return nil, nil, fmt.Errorf("unknown taint spec: %v", taintSpec)
- }
- }
- return taints, taintsToRemove, nil
-}
-
-// ReorganizeTaints returns the updated set of taints, taking into account old taints that were not updated,
-// old taints that were updated, old taints that were deleted, and new taints.
-func ReorganizeTaints(node *v1.Node, overwrite bool, taintsToAdd []v1.Taint, taintsToRemove []v1.Taint) (string, []v1.Taint, error) {
- newTaints := append([]v1.Taint{}, taintsToAdd...)
- oldTaints := node.Spec.Taints
- // add taints that already existing but not updated to newTaints
- added := addTaints(oldTaints, &newTaints)
- allErrs, deleted := deleteTaints(taintsToRemove, &newTaints)
- if (added && deleted) || overwrite {
- return MODIFIED, newTaints, utilerrors.NewAggregate(allErrs)
- } else if added {
- return TAINTED, newTaints, utilerrors.NewAggregate(allErrs)
- }
- return UNTAINTED, newTaints, utilerrors.NewAggregate(allErrs)
-}
-
-// deleteTaints deletes the given taints from the node's taintlist.
-func deleteTaints(taintsToRemove []v1.Taint, newTaints *[]v1.Taint) ([]error, bool) {
- allErrs := []error{}
- var removed bool
- for _, taintToRemove := range taintsToRemove {
- removed = false
- if len(taintToRemove.Effect) > 0 {
- *newTaints, removed = DeleteTaint(*newTaints, &taintToRemove)
- } else {
- *newTaints, removed = DeleteTaintsByKey(*newTaints, taintToRemove.Key)
- }
- if !removed {
- allErrs = append(allErrs, fmt.Errorf("taint %q not found", taintToRemove.ToString()))
- }
- }
- return allErrs, removed
-}
-
-// addTaints adds the newTaints list to existing ones and updates the newTaints List.
-// TODO: This needs a rewrite to take only the new values instead of appended newTaints list to be consistent.
-func addTaints(oldTaints []v1.Taint, newTaints *[]v1.Taint) bool {
- for _, oldTaint := range oldTaints {
- existsInNew := false
- for _, taint := range *newTaints {
- if taint.MatchTaint(&oldTaint) {
- existsInNew = true
- break
- }
- }
- if !existsInNew {
- *newTaints = append(*newTaints, oldTaint)
- }
- }
- return len(oldTaints) != len(*newTaints)
-}
-
-// CheckIfTaintsAlreadyExists checks if the node already has taints that we want to add and returns a string with taint keys.
-func CheckIfTaintsAlreadyExists(oldTaints []v1.Taint, taints []v1.Taint) string {
- var existingTaintList = make([]string, 0)
- for _, taint := range taints {
- for _, oldTaint := range oldTaints {
- if taint.Key == oldTaint.Key && taint.Effect == oldTaint.Effect {
- existingTaintList = append(existingTaintList, taint.Key)
- }
- }
- }
- return strings.Join(existingTaintList, ",")
-}
-
-// DeleteTaintsByKey removes all the taints that have the same key to given taintKey
-func DeleteTaintsByKey(taints []v1.Taint, taintKey string) ([]v1.Taint, bool) {
- newTaints := []v1.Taint{}
- deleted := false
- for i := range taints {
- if taintKey == taints[i].Key {
- deleted = true
- continue
- }
- newTaints = append(newTaints, taints[i])
- }
- return newTaints, deleted
-}
-
-// DeleteTaint removes all the the taints that have the same key and effect to given taintToDelete.
-func DeleteTaint(taints []v1.Taint, taintToDelete *v1.Taint) ([]v1.Taint, bool) {
- newTaints := []v1.Taint{}
- deleted := false
- for i := range taints {
- if taintToDelete.MatchTaint(&taints[i]) {
- deleted = true
- continue
- }
- newTaints = append(newTaints, taints[i])
- }
- return newTaints, deleted
-}
-
-// RemoveTaint tries to remove a taint from annotations list. Returns a new copy of updated Node and true if something was updated
-// false otherwise.
-func RemoveTaint(node *v1.Node, taint *v1.Taint) (*v1.Node, bool, error) {
- newNode := node.DeepCopy()
- nodeTaints := newNode.Spec.Taints
- if len(nodeTaints) == 0 {
- return newNode, false, nil
- }
-
- if !TaintExists(nodeTaints, taint) {
- return newNode, false, nil
- }
-
- newTaints, _ := DeleteTaint(nodeTaints, taint)
- newNode.Spec.Taints = newTaints
- return newNode, true, nil
-}
-
-// AddOrUpdateTaint tries to add a taint to annotations list. Returns a new copy of updated Node and true if something was updated
-// false otherwise.
-func AddOrUpdateTaint(node *v1.Node, taint *v1.Taint) (*v1.Node, bool, error) {
- newNode := node.DeepCopy()
- nodeTaints := newNode.Spec.Taints
-
- var newTaints []v1.Taint
- updated := false
- for i := range nodeTaints {
- if taint.MatchTaint(&nodeTaints[i]) {
- if helper.Semantic.DeepEqual(*taint, nodeTaints[i]) {
- return newNode, false, nil
- }
- newTaints = append(newTaints, *taint)
- updated = true
- continue
- }
-
- newTaints = append(newTaints, nodeTaints[i])
- }
-
- if !updated {
- newTaints = append(newTaints, *taint)
- }
-
- newNode.Spec.Taints = newTaints
- return newNode, true, nil
-}
-
-// TaintExists checks if the given taint exists in list of taints. Returns true if exists false otherwise.
-func TaintExists(taints []v1.Taint, taintToFind *v1.Taint) bool {
- for _, taint := range taints {
- if taint.MatchTaint(taintToFind) {
- return true
- }
- }
- return false
-}
-
-func TaintSetDiff(t1, t2 []v1.Taint) (taintsToAdd []*v1.Taint, taintsToRemove []*v1.Taint) {
- for _, taint := range t1 {
- if !TaintExists(t2, &taint) {
- t := taint
- taintsToAdd = append(taintsToAdd, &t)
- }
- }
-
- for _, taint := range t2 {
- if !TaintExists(t1, &taint) {
- t := taint
- taintsToRemove = append(taintsToRemove, &t)
- }
- }
-
- return
-}
-
-func TaintSetFilter(taints []v1.Taint, fn func(*v1.Taint) bool) []v1.Taint {
- res := []v1.Taint{}
-
- for _, taint := range taints {
- if fn(&taint) {
- res = append(res, taint)
- }
- }
-
- return res
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/doc.go b/vendor/k8s.io/kubernetes/pkg/volume/doc.go
deleted file mode 100644
index c98a5a153..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/doc.go
+++ /dev/null
@@ -1,19 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-// Package volume includes internal representations of external volume types
-// as well as utility methods required to mount/unmount volumes to kubelets.
-package volume // import "k8s.io/kubernetes/pkg/volume"
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/metrics_cached.go b/vendor/k8s.io/kubernetes/pkg/volume/metrics_cached.go
deleted file mode 100644
index ac0dc9b7a..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/metrics_cached.go
+++ /dev/null
@@ -1,74 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "sync"
- "sync/atomic"
-)
-
-var _ MetricsProvider = &cachedMetrics{}
-
-// cachedMetrics represents a MetricsProvider that wraps another provider and
-// caches the result.
-type cachedMetrics struct {
- wrapped MetricsProvider
- resultError error
- resultMetrics *Metrics
- once cacheOnce
-}
-
-// NewCachedMetrics creates a new cachedMetrics wrapping another
-// MetricsProvider and caching the results.
-func NewCachedMetrics(provider MetricsProvider) MetricsProvider {
- return &cachedMetrics{wrapped: provider}
-}
-
-// GetMetrics runs the wrapped metrics provider's GetMetrics methd once and
-// caches the result. Will not cache result if there is an error.
-// See MetricsProvider.GetMetrics
-func (md *cachedMetrics) GetMetrics() (*Metrics, error) {
- md.once.cache(func() error {
- md.resultMetrics, md.resultError = md.wrapped.GetMetrics()
- return md.resultError
- })
- return md.resultMetrics, md.resultError
-}
-
-// Copied from sync.Once but we don't want to cache the results if there is an
-// error
-type cacheOnce struct {
- m sync.Mutex
- done uint32
-}
-
-// Copied from sync.Once but we don't want to cache the results if there is an
-// error
-func (o *cacheOnce) cache(f func() error) {
- if atomic.LoadUint32(&o.done) == 1 {
- return
- }
- // Slow-path.
- o.m.Lock()
- defer o.m.Unlock()
- if o.done == 0 {
- err := f()
- if err == nil {
- atomic.StoreUint32(&o.done, 1)
- }
- }
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/metrics_du.go b/vendor/k8s.io/kubernetes/pkg/volume/metrics_du.go
deleted file mode 100644
index 88a985d5a..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/metrics_du.go
+++ /dev/null
@@ -1,99 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/kubernetes/pkg/volume/util/fs"
-)
-
-var _ MetricsProvider = &metricsDu{}
-
-// metricsDu represents a MetricsProvider that calculates the used and
-// available Volume space by executing the "du" command and gathering
-// filesystem info for the Volume path.
-type metricsDu struct {
- // the directory path the volume is mounted to.
- path string
-}
-
-// NewMetricsDu creates a new metricsDu with the Volume path.
-func NewMetricsDu(path string) MetricsProvider {
- return &metricsDu{path}
-}
-
-// GetMetrics calculates the volume usage and device free space by executing "du"
-// and gathering filesystem info for the Volume path.
-// See MetricsProvider.GetMetrics
-func (md *metricsDu) GetMetrics() (*Metrics, error) {
- metrics := &Metrics{Time: metav1.Now()}
- if md.path == "" {
- return metrics, NewNoPathDefinedError()
- }
-
- err := md.runDu(metrics)
- if err != nil {
- return metrics, err
- }
-
- err = md.runFind(metrics)
- if err != nil {
- return metrics, err
- }
-
- err = md.getFsInfo(metrics)
- if err != nil {
- return metrics, err
- }
-
- return metrics, nil
-}
-
-// runDu executes the "du" command and writes the results to metrics.Used
-func (md *metricsDu) runDu(metrics *Metrics) error {
- used, err := fs.Du(md.path)
- if err != nil {
- return err
- }
- metrics.Used = used
- return nil
-}
-
-// runFind executes the "find" command and writes the results to metrics.InodesUsed
-func (md *metricsDu) runFind(metrics *Metrics) error {
- inodesUsed, err := fs.Find(md.path)
- if err != nil {
- return err
- }
- metrics.InodesUsed = resource.NewQuantity(inodesUsed, resource.BinarySI)
- return nil
-}
-
-// getFsInfo writes metrics.Capacity and metrics.Available from the filesystem
-// info
-func (md *metricsDu) getFsInfo(metrics *Metrics) error {
- available, capacity, _, inodes, inodesFree, _, err := fs.FsInfo(md.path)
- if err != nil {
- return NewFsInfoFailedError(err)
- }
- metrics.Available = resource.NewQuantity(available, resource.BinarySI)
- metrics.Capacity = resource.NewQuantity(capacity, resource.BinarySI)
- metrics.Inodes = resource.NewQuantity(inodes, resource.BinarySI)
- metrics.InodesFree = resource.NewQuantity(inodesFree, resource.BinarySI)
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/metrics_errors.go b/vendor/k8s.io/kubernetes/pkg/volume/metrics_errors.go
deleted file mode 100644
index 50e7c2a21..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/metrics_errors.go
+++ /dev/null
@@ -1,77 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "fmt"
-)
-
-const (
- // ErrCodeNotSupported code for NotSupported Errors.
- ErrCodeNotSupported int = iota + 1
- ErrCodeNoPathDefined
- ErrCodeFsInfoFailed
-)
-
-// NewNotSupportedError creates a new MetricsError with code NotSupported.
-func NewNotSupportedError() *MetricsError {
- return &MetricsError{
- Code: ErrCodeNotSupported,
- Msg: "metrics are not supported for MetricsNil Volumes",
- }
-}
-
-// NewNoPathDefined creates a new MetricsError with code NoPathDefined.
-func NewNoPathDefinedError() *MetricsError {
- return &MetricsError{
- Code: ErrCodeNoPathDefined,
- Msg: "no path defined for disk usage metrics.",
- }
-}
-
-// NewFsInfoFailedError creates a new MetricsError with code FsInfoFailed.
-func NewFsInfoFailedError(err error) *MetricsError {
- return &MetricsError{
- Code: ErrCodeFsInfoFailed,
- Msg: fmt.Sprintf("Failed to get FsInfo due to error %v", err),
- }
-}
-
-// MetricsError to distinguish different Metrics Errors.
-type MetricsError struct {
- Code int
- Msg string
-}
-
-func (e *MetricsError) Error() string {
- return fmt.Sprintf("%s", e.Msg)
-}
-
-// IsNotSupported returns true if and only if err is "key" not found error.
-func IsNotSupported(err error) bool {
- return isErrCode(err, ErrCodeNotSupported)
-}
-
-func isErrCode(err error, code int) bool {
- if err == nil {
- return false
- }
- if e, ok := err.(*MetricsError); ok {
- return e.Code == code
- }
- return false
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/metrics_nil.go b/vendor/k8s.io/kubernetes/pkg/volume/metrics_nil.go
deleted file mode 100644
index 5438dc3de..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/metrics_nil.go
+++ /dev/null
@@ -1,30 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-var _ MetricsProvider = &MetricsNil{}
-
-// MetricsNil represents a MetricsProvider that does not support returning
-// Metrics. It serves as a placeholder for Volumes that do not yet support
-// metrics.
-type MetricsNil struct{}
-
-// GetMetrics returns an empty Metrics and an error.
-// See MetricsProvider.GetMetrics
-func (*MetricsNil) GetMetrics() (*Metrics, error) {
- return &Metrics{}, NewNotSupportedError()
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/metrics_statfs.go b/vendor/k8s.io/kubernetes/pkg/volume/metrics_statfs.go
deleted file mode 100644
index 66f99e30a..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/metrics_statfs.go
+++ /dev/null
@@ -1,69 +0,0 @@
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/kubernetes/pkg/volume/util/fs"
-)
-
-var _ MetricsProvider = &metricsStatFS{}
-
-// metricsStatFS represents a MetricsProvider that calculates the used and available
-// Volume space by stat'ing and gathering filesystem info for the Volume path.
-type metricsStatFS struct {
- // the directory path the volume is mounted to.
- path string
-}
-
-// NewMetricsStatfs creates a new metricsStatFS with the Volume path.
-func NewMetricsStatFS(path string) MetricsProvider {
- return &metricsStatFS{path}
-}
-
-// See MetricsProvider.GetMetrics
-// GetMetrics calculates the volume usage and device free space by executing "du"
-// and gathering filesystem info for the Volume path.
-func (md *metricsStatFS) GetMetrics() (*Metrics, error) {
- metrics := &Metrics{Time: metav1.Now()}
- if md.path == "" {
- return metrics, NewNoPathDefinedError()
- }
-
- err := md.getFsInfo(metrics)
- if err != nil {
- return metrics, err
- }
-
- return metrics, nil
-}
-
-// getFsInfo writes metrics.Capacity, metrics.Used and metrics.Available from the filesystem info
-func (md *metricsStatFS) getFsInfo(metrics *Metrics) error {
- available, capacity, usage, inodes, inodesFree, inodesUsed, err := fs.FsInfo(md.path)
- if err != nil {
- return NewFsInfoFailedError(err)
- }
- metrics.Available = resource.NewQuantity(available, resource.BinarySI)
- metrics.Capacity = resource.NewQuantity(capacity, resource.BinarySI)
- metrics.Used = resource.NewQuantity(usage, resource.BinarySI)
- metrics.Inodes = resource.NewQuantity(inodes, resource.BinarySI)
- metrics.InodesFree = resource.NewQuantity(inodesFree, resource.BinarySI)
- metrics.InodesUsed = resource.NewQuantity(inodesUsed, resource.BinarySI)
- return nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/plugins.go b/vendor/k8s.io/kubernetes/pkg/volume/plugins.go
deleted file mode 100644
index ec4ec5791..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/plugins.go
+++ /dev/null
@@ -1,807 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "fmt"
- "net"
- "strings"
- "sync"
-
- "github.com/golang/glog"
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/types"
- utilerrors "k8s.io/apimachinery/pkg/util/errors"
- "k8s.io/apimachinery/pkg/util/validation"
- clientset "k8s.io/client-go/kubernetes"
- "k8s.io/kubernetes/pkg/cloudprovider"
- "k8s.io/kubernetes/pkg/util/io"
- "k8s.io/kubernetes/pkg/util/mount"
- "k8s.io/kubernetes/pkg/volume/util/recyclerclient"
-)
-
-const (
- // Common parameter which can be specified in StorageClass to specify the desired FSType
- // Provisioners SHOULD implement support for this if they are block device based
- // Must be a filesystem type supported by the host operating system.
- // Ex. "ext4", "xfs", "ntfs". Default value depends on the provisioner
- VolumeParameterFSType = "fstype"
-)
-
-// VolumeOptions contains option information about a volume.
-type VolumeOptions struct {
- // The attributes below are required by volume.Provisioner
- // TODO: refactor all of this out of volumes when an admin can configure
- // many kinds of provisioners.
-
- // Reclamation policy for a persistent volume
- PersistentVolumeReclaimPolicy v1.PersistentVolumeReclaimPolicy
- // Mount options for a persistent volume
- MountOptions []string
- // Suggested PV.Name of the PersistentVolume to provision.
- // This is a generated name guaranteed to be unique in Kubernetes cluster.
- // If you choose not to use it as volume name, ensure uniqueness by either
- // combining it with your value or create unique values of your own.
- PVName string
- // PVC is reference to the claim that lead to provisioning of a new PV.
- // Provisioners *must* create a PV that would be matched by this PVC,
- // i.e. with required capacity, accessMode, labels matching PVC.Selector and
- // so on.
- PVC *v1.PersistentVolumeClaim
- // Unique name of Kubernetes cluster.
- ClusterName string
- // Tags to attach to the real volume in the cloud provider - e.g. AWS EBS
- CloudTags *map[string]string
- // Volume provisioning parameters from StorageClass
- Parameters map[string]string
-}
-
-type DynamicPluginProber interface {
- Init() error
-
- // If an update has occurred since the last probe, updated = true
- // and the list of probed plugins is returned.
- // Otherwise, update = false and probedPlugins = nil.
- //
- // If an error occurs, updated and probedPlugins are undefined.
- Probe() (updated bool, probedPlugins []VolumePlugin, err error)
-}
-
-// VolumePlugin is an interface to volume plugins that can be used on a
-// kubernetes node (e.g. by kubelet) to instantiate and manage volumes.
-type VolumePlugin interface {
- // Init initializes the plugin. This will be called exactly once
- // before any New* calls are made - implementations of plugins may
- // depend on this.
- Init(host VolumeHost) error
-
- // Name returns the plugin's name. Plugins must use namespaced names
- // such as "example.com/volume" and contain exactly one '/' character.
- // The "kubernetes.io" namespace is reserved for plugins which are
- // bundled with kubernetes.
- GetPluginName() string
-
- // GetVolumeName returns the name/ID to uniquely identifying the actual
- // backing device, directory, path, etc. referenced by the specified volume
- // spec.
- // For Attachable volumes, this value must be able to be passed back to
- // volume Detach methods to identify the device to act on.
- // If the plugin does not support the given spec, this returns an error.
- GetVolumeName(spec *Spec) (string, error)
-
- // CanSupport tests whether the plugin supports a given volume
- // specification from the API. The spec pointer should be considered
- // const.
- CanSupport(spec *Spec) bool
-
- // RequiresRemount returns true if this plugin requires mount calls to be
- // reexecuted. Atomically updating volumes, like Downward API, depend on
- // this to update the contents of the volume.
- RequiresRemount() bool
-
- // NewMounter creates a new volume.Mounter from an API specification.
- // Ownership of the spec pointer in *not* transferred.
- // - spec: The v1.Volume spec
- // - pod: The enclosing pod
- NewMounter(spec *Spec, podRef *v1.Pod, opts VolumeOptions) (Mounter, error)
-
- // NewUnmounter creates a new volume.Unmounter from recoverable state.
- // - name: The volume name, as per the v1.Volume spec.
- // - podUID: The UID of the enclosing pod
- NewUnmounter(name string, podUID types.UID) (Unmounter, error)
-
- // ConstructVolumeSpec constructs a volume spec based on the given volume name
- // and mountPath. The spec may have incomplete information due to limited
- // information from input. This function is used by volume manager to reconstruct
- // volume spec by reading the volume directories from disk
- ConstructVolumeSpec(volumeName, mountPath string) (*Spec, error)
-
- // SupportsMountOption returns true if volume plugins supports Mount options
- // Specifying mount options in a volume plugin that doesn't support
- // user specified mount options will result in error creating persistent volumes
- SupportsMountOption() bool
-
- // SupportsBulkVolumeVerification checks if volume plugin type is capable
- // of enabling bulk polling of all nodes. This can speed up verification of
- // attached volumes by quite a bit, but underlying pluging must support it.
- SupportsBulkVolumeVerification() bool
-}
-
-// PersistentVolumePlugin is an extended interface of VolumePlugin and is used
-// by volumes that want to provide long term persistence of data
-type PersistentVolumePlugin interface {
- VolumePlugin
- // GetAccessModes describes the ways a given volume can be accessed/mounted.
- GetAccessModes() []v1.PersistentVolumeAccessMode
-}
-
-// RecyclableVolumePlugin is an extended interface of VolumePlugin and is used
-// by persistent volumes that want to be recycled before being made available
-// again to new claims
-type RecyclableVolumePlugin interface {
- VolumePlugin
-
- // Recycle knows how to reclaim this
- // resource after the volume's release from a PersistentVolumeClaim.
- // Recycle will use the provided recorder to write any events that might be
- // interesting to user. It's expected that caller will pass these events to
- // the PV being recycled.
- Recycle(pvName string, spec *Spec, eventRecorder recyclerclient.RecycleEventRecorder) error
-}
-
-// DeletableVolumePlugin is an extended interface of VolumePlugin and is used
-// by persistent volumes that want to be deleted from the cluster after their
-// release from a PersistentVolumeClaim.
-type DeletableVolumePlugin interface {
- VolumePlugin
- // NewDeleter creates a new volume.Deleter which knows how to delete this
- // resource in accordance with the underlying storage provider after the
- // volume's release from a claim
- NewDeleter(spec *Spec) (Deleter, error)
-}
-
-const (
- // Name of a volume in external cloud that is being provisioned and thus
- // should be ignored by rest of Kubernetes.
- ProvisionedVolumeName = "placeholder-for-provisioning"
-)
-
-// ProvisionableVolumePlugin is an extended interface of VolumePlugin and is
-// used to create volumes for the cluster.
-type ProvisionableVolumePlugin interface {
- VolumePlugin
- // NewProvisioner creates a new volume.Provisioner which knows how to
- // create PersistentVolumes in accordance with the plugin's underlying
- // storage provider
- NewProvisioner(options VolumeOptions) (Provisioner, error)
-}
-
-// AttachableVolumePlugin is an extended interface of VolumePlugin and is used for volumes that require attachment
-// to a node before mounting.
-type AttachableVolumePlugin interface {
- VolumePlugin
- NewAttacher() (Attacher, error)
- NewDetacher() (Detacher, error)
- GetDeviceMountRefs(deviceMountPath string) ([]string, error)
-}
-
-// ExpandableVolumePlugin is an extended interface of VolumePlugin and is used for volumes that can be
-// expanded
-type ExpandableVolumePlugin interface {
- VolumePlugin
- ExpandVolumeDevice(spec *Spec, newSize resource.Quantity, oldSize resource.Quantity) (resource.Quantity, error)
- RequiresFSResize() bool
-}
-
-// BlockVolumePlugin is an extend interface of VolumePlugin and is used for block volumes support.
-type BlockVolumePlugin interface {
- VolumePlugin
- // NewBlockVolumeMapper creates a new volume.BlockVolumeMapper from an API specification.
- // Ownership of the spec pointer in *not* transferred.
- // - spec: The v1.Volume spec
- // - pod: The enclosing pod
- NewBlockVolumeMapper(spec *Spec, podRef *v1.Pod, opts VolumeOptions) (BlockVolumeMapper, error)
- // NewBlockVolumeUnmapper creates a new volume.BlockVolumeUnmapper from recoverable state.
- // - name: The volume name, as per the v1.Volume spec.
- // - podUID: The UID of the enclosing pod
- NewBlockVolumeUnmapper(name string, podUID types.UID) (BlockVolumeUnmapper, error)
- // ConstructBlockVolumeSpec constructs a volume spec based on the given
- // podUID, volume name and a pod device map path.
- // The spec may have incomplete information due to limited information
- // from input. This function is used by volume manager to reconstruct
- // volume spec by reading the volume directories from disk.
- ConstructBlockVolumeSpec(podUID types.UID, volumeName, mountPath string) (*Spec, error)
-}
-
-// VolumeHost is an interface that plugins can use to access the kubelet.
-type VolumeHost interface {
- // GetPluginDir returns the absolute path to a directory under which
- // a given plugin may store data. This directory might not actually
- // exist on disk yet. For plugin data that is per-pod, see
- // GetPodPluginDir().
- GetPluginDir(pluginName string) string
-
- // GetVolumeDevicePluginDir returns the absolute path to a directory
- // under which a given plugin may store data.
- // ex. plugins/kubernetes.io/{PluginName}/{DefaultKubeletVolumeDevicesDirName}/{volumePluginDependentPath}/
- GetVolumeDevicePluginDir(pluginName string) string
-
- // GetPodVolumeDir returns the absolute path a directory which
- // represents the named volume under the named plugin for the given
- // pod. If the specified pod does not exist, the result of this call
- // might not exist.
- GetPodVolumeDir(podUID types.UID, pluginName string, volumeName string) string
-
- // GetPodPluginDir returns the absolute path to a directory under which
- // a given plugin may store data for a given pod. If the specified pod
- // does not exist, the result of this call might not exist. This
- // directory might not actually exist on disk yet.
- GetPodPluginDir(podUID types.UID, pluginName string) string
-
- // GetPodVolumeDeviceDir returns the absolute path a directory which
- // represents the named plugin for the given pod.
- // If the specified pod does not exist, the result of this call
- // might not exist.
- // ex. pods/{podUid}/{DefaultKubeletVolumeDevicesDirName}/{escapeQualifiedPluginName}/
- GetPodVolumeDeviceDir(podUID types.UID, pluginName string) string
-
- // GetKubeClient returns a client interface
- GetKubeClient() clientset.Interface
-
- // NewWrapperMounter finds an appropriate plugin with which to handle
- // the provided spec. This is used to implement volume plugins which
- // "wrap" other plugins. For example, the "secret" volume is
- // implemented in terms of the "emptyDir" volume.
- NewWrapperMounter(volName string, spec Spec, pod *v1.Pod, opts VolumeOptions) (Mounter, error)
-
- // NewWrapperUnmounter finds an appropriate plugin with which to handle
- // the provided spec. See comments on NewWrapperMounter for more
- // context.
- NewWrapperUnmounter(volName string, spec Spec, podUID types.UID) (Unmounter, error)
-
- // Get cloud provider from kubelet.
- GetCloudProvider() cloudprovider.Interface
-
- // Get mounter interface.
- GetMounter(pluginName string) mount.Interface
-
- // Get writer interface for writing data to disk.
- GetWriter() io.Writer
-
- // Returns the hostname of the host kubelet is running on
- GetHostName() string
-
- // Returns host IP or nil in the case of error.
- GetHostIP() (net.IP, error)
-
- // Returns node allocatable.
- GetNodeAllocatable() (v1.ResourceList, error)
-
- // Returns a function that returns a secret.
- GetSecretFunc() func(namespace, name string) (*v1.Secret, error)
-
- // Returns a function that returns a configmap.
- GetConfigMapFunc() func(namespace, name string) (*v1.ConfigMap, error)
-
- // Returns an interface that should be used to execute any utilities in volume plugins
- GetExec(pluginName string) mount.Exec
-
- // Returns the labels on the node
- GetNodeLabels() (map[string]string, error)
-
- // Returns the name of the node
- GetNodeName() types.NodeName
-}
-
-// VolumePluginMgr tracks registered plugins.
-type VolumePluginMgr struct {
- mutex sync.Mutex
- plugins map[string]VolumePlugin
- prober DynamicPluginProber
- probedPlugins []VolumePlugin
- Host VolumeHost
-}
-
-// Spec is an internal representation of a volume. All API volume types translate to Spec.
-type Spec struct {
- Volume *v1.Volume
- PersistentVolume *v1.PersistentVolume
- ReadOnly bool
-}
-
-// Name returns the name of either Volume or PersistentVolume, one of which must not be nil.
-func (spec *Spec) Name() string {
- switch {
- case spec.Volume != nil:
- return spec.Volume.Name
- case spec.PersistentVolume != nil:
- return spec.PersistentVolume.Name
- default:
- return ""
- }
-}
-
-// VolumeConfig is how volume plugins receive configuration. An instance
-// specific to the plugin will be passed to the plugin's
-// ProbeVolumePlugins(config) func. Reasonable defaults will be provided by
-// the binary hosting the plugins while allowing override of those default
-// values. Those config values are then set to an instance of VolumeConfig
-// and passed to the plugin.
-//
-// Values in VolumeConfig are intended to be relevant to several plugins, but
-// not necessarily all plugins. The preference is to leverage strong typing
-// in this struct. All config items must have a descriptive but non-specific
-// name (i.e, RecyclerMinimumTimeout is OK but RecyclerMinimumTimeoutForNFS is
-// !OK). An instance of config will be given directly to the plugin, so
-// config names specific to plugins are unneeded and wrongly expose plugins in
-// this VolumeConfig struct.
-//
-// OtherAttributes is a map of string values intended for one-off
-// configuration of a plugin or config that is only relevant to a single
-// plugin. All values are passed by string and require interpretation by the
-// plugin. Passing config as strings is the least desirable option but can be
-// used for truly one-off configuration. The binary should still use strong
-// typing for this value when binding CLI values before they are passed as
-// strings in OtherAttributes.
-type VolumeConfig struct {
- // RecyclerPodTemplate is pod template that understands how to scrub clean
- // a persistent volume after its release. The template is used by plugins
- // which override specific properties of the pod in accordance with that
- // plugin. See NewPersistentVolumeRecyclerPodTemplate for the properties
- // that are expected to be overridden.
- RecyclerPodTemplate *v1.Pod
-
- // RecyclerMinimumTimeout is the minimum amount of time in seconds for the
- // recycler pod's ActiveDeadlineSeconds attribute. Added to the minimum
- // timeout is the increment per Gi of capacity.
- RecyclerMinimumTimeout int
-
- // RecyclerTimeoutIncrement is the number of seconds added to the recycler
- // pod's ActiveDeadlineSeconds for each Gi of capacity in the persistent
- // volume. Example: 5Gi volume x 30s increment = 150s + 30s minimum = 180s
- // ActiveDeadlineSeconds for recycler pod
- RecyclerTimeoutIncrement int
-
- // PVName is name of the PersistentVolume instance that is being recycled.
- // It is used to generate unique recycler pod name.
- PVName string
-
- // OtherAttributes stores config as strings. These strings are opaque to
- // the system and only understood by the binary hosting the plugin and the
- // plugin itself.
- OtherAttributes map[string]string
-
- // ProvisioningEnabled configures whether provisioning of this plugin is
- // enabled or not. Currently used only in host_path plugin.
- ProvisioningEnabled bool
-}
-
-// NewSpecFromVolume creates an Spec from an v1.Volume
-func NewSpecFromVolume(vs *v1.Volume) *Spec {
- return &Spec{
- Volume: vs,
- }
-}
-
-// NewSpecFromPersistentVolume creates an Spec from an v1.PersistentVolume
-func NewSpecFromPersistentVolume(pv *v1.PersistentVolume, readOnly bool) *Spec {
- return &Spec{
- PersistentVolume: pv,
- ReadOnly: readOnly,
- }
-}
-
-// InitPlugins initializes each plugin. All plugins must have unique names.
-// This must be called exactly once before any New* methods are called on any
-// plugins.
-func (pm *VolumePluginMgr) InitPlugins(plugins []VolumePlugin, prober DynamicPluginProber, host VolumeHost) error {
- pm.mutex.Lock()
- defer pm.mutex.Unlock()
-
- pm.Host = host
-
- if prober == nil {
- // Use a dummy prober to prevent nil deference.
- pm.prober = &dummyPluginProber{}
- } else {
- pm.prober = prober
- }
- if err := pm.prober.Init(); err != nil {
- // Prober init failure should not affect the initialization of other plugins.
- glog.Errorf("Error initializing dynamic plugin prober: %s", err)
- pm.prober = &dummyPluginProber{}
- }
-
- if pm.plugins == nil {
- pm.plugins = map[string]VolumePlugin{}
- }
-
- allErrs := []error{}
- for _, plugin := range plugins {
- name := plugin.GetPluginName()
- if errs := validation.IsQualifiedName(name); len(errs) != 0 {
- allErrs = append(allErrs, fmt.Errorf("volume plugin has invalid name: %q: %s", name, strings.Join(errs, ";")))
- continue
- }
-
- if _, found := pm.plugins[name]; found {
- allErrs = append(allErrs, fmt.Errorf("volume plugin %q was registered more than once", name))
- continue
- }
- err := plugin.Init(host)
- if err != nil {
- glog.Errorf("Failed to load volume plugin %s, error: %s", name, err.Error())
- allErrs = append(allErrs, err)
- continue
- }
- pm.plugins[name] = plugin
- glog.V(1).Infof("Loaded volume plugin %q", name)
- }
- return utilerrors.NewAggregate(allErrs)
-}
-
-func (pm *VolumePluginMgr) initProbedPlugin(probedPlugin VolumePlugin) error {
- name := probedPlugin.GetPluginName()
- if errs := validation.IsQualifiedName(name); len(errs) != 0 {
- return fmt.Errorf("volume plugin has invalid name: %q: %s", name, strings.Join(errs, ";"))
- }
-
- err := probedPlugin.Init(pm.Host)
- if err != nil {
- return fmt.Errorf("Failed to load volume plugin %s, error: %s", name, err.Error())
- }
-
- glog.V(1).Infof("Loaded volume plugin %q", name)
- return nil
-}
-
-// FindPluginBySpec looks for a plugin that can support a given volume
-// specification. If no plugins can support or more than one plugin can
-// support it, return error.
-func (pm *VolumePluginMgr) FindPluginBySpec(spec *Spec) (VolumePlugin, error) {
- pm.mutex.Lock()
- defer pm.mutex.Unlock()
-
- if spec == nil {
- return nil, fmt.Errorf("Could not find plugin because volume spec is nil")
- }
-
- matchedPluginNames := []string{}
- matches := []VolumePlugin{}
- for k, v := range pm.plugins {
- if v.CanSupport(spec) {
- matchedPluginNames = append(matchedPluginNames, k)
- matches = append(matches, v)
- }
- }
-
- pm.refreshProbedPlugins()
- for _, plugin := range pm.probedPlugins {
- if plugin.CanSupport(spec) {
- matchedPluginNames = append(matchedPluginNames, plugin.GetPluginName())
- matches = append(matches, plugin)
- }
- }
-
- if len(matches) == 0 {
- return nil, fmt.Errorf("no volume plugin matched")
- }
- if len(matches) > 1 {
- return nil, fmt.Errorf("multiple volume plugins matched: %s", strings.Join(matchedPluginNames, ","))
- }
- return matches[0], nil
-}
-
-// FindPluginByName fetches a plugin by name or by legacy name. If no plugin
-// is found, returns error.
-func (pm *VolumePluginMgr) FindPluginByName(name string) (VolumePlugin, error) {
- pm.mutex.Lock()
- defer pm.mutex.Unlock()
-
- // Once we can get rid of legacy names we can reduce this to a map lookup.
- matchedPluginNames := []string{}
- matches := []VolumePlugin{}
- for k, v := range pm.plugins {
- if v.GetPluginName() == name {
- matchedPluginNames = append(matchedPluginNames, k)
- matches = append(matches, v)
- }
- }
-
- pm.refreshProbedPlugins()
- for _, plugin := range pm.probedPlugins {
- if plugin.GetPluginName() == name {
- matchedPluginNames = append(matchedPluginNames, plugin.GetPluginName())
- matches = append(matches, plugin)
- }
- }
-
- if len(matches) == 0 {
- return nil, fmt.Errorf("no volume plugin matched")
- }
- if len(matches) > 1 {
- return nil, fmt.Errorf("multiple volume plugins matched: %s", strings.Join(matchedPluginNames, ","))
- }
- return matches[0], nil
-}
-
-// Check if probedPlugin cache update is required.
-// If it is, initialize all probed plugins and replace the cache with them.
-func (pm *VolumePluginMgr) refreshProbedPlugins() {
- updated, plugins, err := pm.prober.Probe()
- if err != nil {
- glog.Errorf("Error dynamically probing plugins: %s", err)
- return // Use cached plugins upon failure.
- }
-
- if updated {
- pm.probedPlugins = []VolumePlugin{}
- for _, plugin := range plugins {
- if err := pm.initProbedPlugin(plugin); err != nil {
- glog.Errorf("Error initializing dynamically probed plugin %s; error: %s",
- plugin.GetPluginName(), err)
- continue
- }
- pm.probedPlugins = append(pm.probedPlugins, plugin)
- }
- }
-}
-
-// FindPersistentPluginBySpec looks for a persistent volume plugin that can
-// support a given volume specification. If no plugin is found, return an
-// error
-func (pm *VolumePluginMgr) FindPersistentPluginBySpec(spec *Spec) (PersistentVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, fmt.Errorf("Could not find volume plugin for spec: %#v", spec)
- }
- if persistentVolumePlugin, ok := volumePlugin.(PersistentVolumePlugin); ok {
- return persistentVolumePlugin, nil
- }
- return nil, fmt.Errorf("no persistent volume plugin matched")
-}
-
-// FindPersistentPluginByName fetches a persistent volume plugin by name. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindPersistentPluginByName(name string) (PersistentVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
- if persistentVolumePlugin, ok := volumePlugin.(PersistentVolumePlugin); ok {
- return persistentVolumePlugin, nil
- }
- return nil, fmt.Errorf("no persistent volume plugin matched")
-}
-
-// FindRecyclablePluginByName fetches a persistent volume plugin by name. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindRecyclablePluginBySpec(spec *Spec) (RecyclableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
- if recyclableVolumePlugin, ok := volumePlugin.(RecyclableVolumePlugin); ok {
- return recyclableVolumePlugin, nil
- }
- return nil, fmt.Errorf("no recyclable volume plugin matched")
-}
-
-// FindProvisionablePluginByName fetches a persistent volume plugin by name. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindProvisionablePluginByName(name string) (ProvisionableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
- if provisionableVolumePlugin, ok := volumePlugin.(ProvisionableVolumePlugin); ok {
- return provisionableVolumePlugin, nil
- }
- return nil, fmt.Errorf("no provisionable volume plugin matched")
-}
-
-// FindDeletablePluginBySppec fetches a persistent volume plugin by spec. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindDeletablePluginBySpec(spec *Spec) (DeletableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
- if deletableVolumePlugin, ok := volumePlugin.(DeletableVolumePlugin); ok {
- return deletableVolumePlugin, nil
- }
- return nil, fmt.Errorf("no deletable volume plugin matched")
-}
-
-// FindDeletablePluginByName fetches a persistent volume plugin by name. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindDeletablePluginByName(name string) (DeletableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
- if deletableVolumePlugin, ok := volumePlugin.(DeletableVolumePlugin); ok {
- return deletableVolumePlugin, nil
- }
- return nil, fmt.Errorf("no deletable volume plugin matched")
-}
-
-// FindCreatablePluginBySpec fetches a persistent volume plugin by name. If
-// no plugin is found, returns error.
-func (pm *VolumePluginMgr) FindCreatablePluginBySpec(spec *Spec) (ProvisionableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
- if provisionableVolumePlugin, ok := volumePlugin.(ProvisionableVolumePlugin); ok {
- return provisionableVolumePlugin, nil
- }
- return nil, fmt.Errorf("no creatable volume plugin matched")
-}
-
-// FindAttachablePluginBySpec fetches a persistent volume plugin by spec.
-// Unlike the other "FindPlugin" methods, this does not return error if no
-// plugin is found. All volumes require a mounter and unmounter, but not
-// every volume will have an attacher/detacher.
-func (pm *VolumePluginMgr) FindAttachablePluginBySpec(spec *Spec) (AttachableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
- if attachableVolumePlugin, ok := volumePlugin.(AttachableVolumePlugin); ok {
- return attachableVolumePlugin, nil
- }
- return nil, nil
-}
-
-// FindAttachablePluginByName fetches an attachable volume plugin by name.
-// Unlike the other "FindPlugin" methods, this does not return error if no
-// plugin is found. All volumes require a mounter and unmounter, but not
-// every volume will have an attacher/detacher.
-func (pm *VolumePluginMgr) FindAttachablePluginByName(name string) (AttachableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
- if attachablePlugin, ok := volumePlugin.(AttachableVolumePlugin); ok {
- return attachablePlugin, nil
- }
- return nil, nil
-}
-
-// FindExpandablePluginBySpec fetches a persistent volume plugin by spec.
-func (pm *VolumePluginMgr) FindExpandablePluginBySpec(spec *Spec) (ExpandableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
-
- if expandableVolumePlugin, ok := volumePlugin.(ExpandableVolumePlugin); ok {
- return expandableVolumePlugin, nil
- }
- return nil, nil
-}
-
-// FindExpandablePluginBySpec fetches a persistent volume plugin by name.
-func (pm *VolumePluginMgr) FindExpandablePluginByName(name string) (ExpandableVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
-
- if expandableVolumePlugin, ok := volumePlugin.(ExpandableVolumePlugin); ok {
- return expandableVolumePlugin, nil
- }
- return nil, nil
-}
-
-// FindMapperPluginBySpec fetches a block volume plugin by spec.
-func (pm *VolumePluginMgr) FindMapperPluginBySpec(spec *Spec) (BlockVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginBySpec(spec)
- if err != nil {
- return nil, err
- }
-
- if blockVolumePlugin, ok := volumePlugin.(BlockVolumePlugin); ok {
- return blockVolumePlugin, nil
- }
- return nil, nil
-}
-
-// FindMapperPluginByName fetches a block volume plugin by name.
-func (pm *VolumePluginMgr) FindMapperPluginByName(name string) (BlockVolumePlugin, error) {
- volumePlugin, err := pm.FindPluginByName(name)
- if err != nil {
- return nil, err
- }
-
- if blockVolumePlugin, ok := volumePlugin.(BlockVolumePlugin); ok {
- return blockVolumePlugin, nil
- }
- return nil, nil
-}
-
-// NewPersistentVolumeRecyclerPodTemplate creates a template for a recycler
-// pod. By default, a recycler pod simply runs "rm -rf" on a volume and tests
-// for emptiness. Most attributes of the template will be correct for most
-// plugin implementations. The following attributes can be overridden per
-// plugin via configuration:
-//
-// 1. pod.Spec.Volumes[0].VolumeSource must be overridden. Recycler
-// implementations without a valid VolumeSource will fail.
-// 2. pod.GenerateName helps distinguish recycler pods by name. Recommended.
-// Default is "pv-recycler-".
-// 3. pod.Spec.ActiveDeadlineSeconds gives the recycler pod a maximum timeout
-// before failing. Recommended. Default is 60 seconds.
-//
-// See HostPath and NFS for working recycler examples
-func NewPersistentVolumeRecyclerPodTemplate() *v1.Pod {
- timeout := int64(60)
- pod := &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- GenerateName: "pv-recycler-",
- Namespace: metav1.NamespaceDefault,
- },
- Spec: v1.PodSpec{
- ActiveDeadlineSeconds: &timeout,
- RestartPolicy: v1.RestartPolicyNever,
- Volumes: []v1.Volume{
- {
- Name: "vol",
- // IMPORTANT! All plugins using this template MUST
- // override pod.Spec.Volumes[0].VolumeSource Recycler
- // implementations without a valid VolumeSource will fail.
- VolumeSource: v1.VolumeSource{},
- },
- },
- Containers: []v1.Container{
- {
- Name: "pv-recycler",
- Image: "busybox:1.27",
- Command: []string{"/bin/sh"},
- Args: []string{"-c", "test -e /scrub && rm -rf /scrub/..?* /scrub/.[!.]* /scrub/* && test -z \"$(ls -A /scrub)\" || exit 1"},
- VolumeMounts: []v1.VolumeMount{
- {
- Name: "vol",
- MountPath: "/scrub",
- },
- },
- },
- },
- },
- }
- return pod
-}
-
-// Check validity of recycle pod template
-// List of checks:
-// - at least one volume is defined in the recycle pod template
-// If successful, returns nil
-// if unsuccessful, returns an error.
-func ValidateRecyclerPodTemplate(pod *v1.Pod) error {
- if len(pod.Spec.Volumes) < 1 {
- return fmt.Errorf("does not contain any volume(s)")
- }
- return nil
-}
-
-type dummyPluginProber struct{}
-
-func (*dummyPluginProber) Init() error { return nil }
-func (*dummyPluginProber) Probe() (bool, []VolumePlugin, error) { return false, nil, nil }
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs.go b/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs.go
deleted file mode 100644
index bbb4b0105..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs.go
+++ /dev/null
@@ -1,97 +0,0 @@
-// +build linux darwin
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package fs
-
-import (
- "bytes"
- "fmt"
- "os/exec"
- "strings"
-
- "golang.org/x/sys/unix"
-
- "k8s.io/apimachinery/pkg/api/resource"
-)
-
-// FSInfo linux returns (available bytes, byte capacity, byte usage, total inodes, inodes free, inode usage, error)
-// for the filesystem that path resides upon.
-func FsInfo(path string) (int64, int64, int64, int64, int64, int64, error) {
- statfs := &unix.Statfs_t{}
- err := unix.Statfs(path, statfs)
- if err != nil {
- return 0, 0, 0, 0, 0, 0, err
- }
-
- // Available is blocks available * fragment size
- available := int64(statfs.Bavail) * int64(statfs.Bsize)
-
- // Capacity is total block count * fragment size
- capacity := int64(statfs.Blocks) * int64(statfs.Bsize)
-
- // Usage is block being used * fragment size (aka block size).
- usage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)
-
- inodes := int64(statfs.Files)
- inodesFree := int64(statfs.Ffree)
- inodesUsed := inodes - inodesFree
-
- return available, capacity, usage, inodes, inodesFree, inodesUsed, nil
-}
-
-func Du(path string) (*resource.Quantity, error) {
- // Uses the same niceness level as cadvisor.fs does when running du
- // Uses -B 1 to always scale to a blocksize of 1 byte
- out, err := exec.Command("nice", "-n", "19", "du", "-s", "-B", "1", path).CombinedOutput()
- if err != nil {
- return nil, fmt.Errorf("failed command 'du' ($ nice -n 19 du -s -B 1) on path %s with error %v", path, err)
- }
- used, err := resource.ParseQuantity(strings.Fields(string(out))[0])
- if err != nil {
- return nil, fmt.Errorf("failed to parse 'du' output %s due to error %v", out, err)
- }
- used.Format = resource.BinarySI
- return &used, nil
-}
-
-// Find uses the equivalent of the command `find <path> -dev -printf '.' | wc -c` to count files and directories.
-// While this is not an exact measure of inodes used, it is a very good approximation.
-func Find(path string) (int64, error) {
- if path == "" {
- return 0, fmt.Errorf("invalid directory")
- }
- var counter byteCounter
- var stderr bytes.Buffer
- findCmd := exec.Command("find", path, "-xdev", "-printf", ".")
- findCmd.Stdout, findCmd.Stderr = &counter, &stderr
- if err := findCmd.Start(); err != nil {
- return 0, fmt.Errorf("failed to exec cmd %v - %v; stderr: %v", findCmd.Args, err, stderr.String())
- }
- if err := findCmd.Wait(); err != nil {
- return 0, fmt.Errorf("cmd %v failed. stderr: %s; err: %v", findCmd.Args, stderr.String(), err)
- }
- return counter.bytesWritten, nil
-}
-
-// Simple io.Writer implementation that counts how many bytes were written.
-type byteCounter struct{ bytesWritten int64 }
-
-func (b *byteCounter) Write(p []byte) (int, error) {
- b.bytesWritten += int64(len(p))
- return len(p), nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs_unsupported.go b/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs_unsupported.go
deleted file mode 100644
index da41fc8ee..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/util/fs/fs_unsupported.go
+++ /dev/null
@@ -1,38 +0,0 @@
-// +build !linux,!darwin
-
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package fs
-
-import (
- "fmt"
-
- "k8s.io/apimachinery/pkg/api/resource"
-)
-
-// FSInfo unsupported returns 0 values for available and capacity and an error.
-func FsInfo(path string) (int64, int64, int64, int64, int64, int64, error) {
- return 0, 0, 0, 0, 0, 0, fmt.Errorf("FsInfo not supported for this build.")
-}
-
-func Du(path string) (*resource.Quantity, error) {
- return nil, fmt.Errorf("Du not supported for this build.")
-}
-
-func Find(path string) (int64, error) {
- return 0, fmt.Errorf("Find not supported for this build.")
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/util/recyclerclient/recycler_client.go b/vendor/k8s.io/kubernetes/pkg/volume/util/recyclerclient/recycler_client.go
deleted file mode 100644
index 1af6465c6..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/util/recyclerclient/recycler_client.go
+++ /dev/null
@@ -1,252 +0,0 @@
-/*
-Copyright 2018 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package recyclerclient
-
-import (
- "fmt"
-
- "github.com/golang/glog"
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/api/errors"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/fields"
- "k8s.io/apimachinery/pkg/watch"
- clientset "k8s.io/client-go/kubernetes"
-)
-
-type RecycleEventRecorder func(eventtype, message string)
-
-// RecycleVolumeByWatchingPodUntilCompletion is intended for use with volume
-// Recyclers. This function will save the given Pod to the API and watch it
-// until it completes, fails, or the pod's ActiveDeadlineSeconds is exceeded,
-// whichever comes first. An attempt to delete a recycler pod is always
-// attempted before returning.
-//
-// In case there is a pod with the same namespace+name already running, this
-// function deletes it as it is not able to judge if it is an old recycler
-// or user has forged a fake recycler to block Kubernetes from recycling.//
-//
-// pod - the pod designed by a volume plugin to recycle the volume. pod.Name
-// will be overwritten with unique name based on PV.Name.
-// client - kube client for API operations.
-func RecycleVolumeByWatchingPodUntilCompletion(pvName string, pod *v1.Pod, kubeClient clientset.Interface, recorder RecycleEventRecorder) error {
- return internalRecycleVolumeByWatchingPodUntilCompletion(pvName, pod, newRecyclerClient(kubeClient, recorder))
-}
-
-// same as above func comments, except 'recyclerClient' is a narrower pod API
-// interface to ease testing
-func internalRecycleVolumeByWatchingPodUntilCompletion(pvName string, pod *v1.Pod, recyclerClient recyclerClient) error {
- glog.V(5).Infof("creating recycler pod for volume %s\n", pod.Name)
-
- // Generate unique name for the recycler pod - we need to get "already
- // exists" error when a previous controller has already started recycling
- // the volume. Here we assume that pv.Name is already unique.
- pod.Name = "recycler-for-" + pvName
- pod.GenerateName = ""
-
- stopChannel := make(chan struct{})
- defer close(stopChannel)
- podCh, err := recyclerClient.WatchPod(pod.Name, pod.Namespace, stopChannel)
- if err != nil {
- glog.V(4).Infof("cannot start watcher for pod %s/%s: %v", pod.Namespace, pod.Name, err)
- return err
- }
-
- // Start the pod
- _, err = recyclerClient.CreatePod(pod)
- if err != nil {
- if errors.IsAlreadyExists(err) {
- deleteErr := recyclerClient.DeletePod(pod.Name, pod.Namespace)
- if deleteErr != nil {
- return fmt.Errorf("failed to delete old recycler pod %s/%s: %s", pod.Namespace, pod.Name, deleteErr)
- }
- // Recycler will try again and the old pod will be hopefully deleted
- // at that time.
- return fmt.Errorf("old recycler pod found, will retry later")
- }
- return fmt.Errorf("unexpected error creating recycler pod: %+v", err)
- }
- err = waitForPod(pod, recyclerClient, podCh)
-
- // In all cases delete the recycler pod and log its result.
- glog.V(2).Infof("deleting recycler pod %s/%s", pod.Namespace, pod.Name)
- deleteErr := recyclerClient.DeletePod(pod.Name, pod.Namespace)
- if deleteErr != nil {
- glog.Errorf("failed to delete recycler pod %s/%s: %v", pod.Namespace, pod.Name, err)
- }
-
- // Returning recycler error is preferred, the pod will be deleted again on
- // the next retry.
- if err != nil {
- return fmt.Errorf("failed to recycle volume: %s", err)
- }
-
- // Recycle succeeded but we failed to delete the recycler pod. Report it,
- // the controller will re-try recycling the PV again shortly.
- if deleteErr != nil {
- return fmt.Errorf("failed to delete recycler pod: %s", deleteErr)
- }
-
- return nil
-}
-
-// waitForPod watches the pod it until it finishes and send all events on the
-// pod to the PV.
-func waitForPod(pod *v1.Pod, recyclerClient recyclerClient, podCh <-chan watch.Event) error {
- for {
- event, ok := <-podCh
- if !ok {
- return fmt.Errorf("recycler pod %q watch channel had been closed", pod.Name)
- }
- switch event.Object.(type) {
- case *v1.Pod:
- // POD changed
- pod := event.Object.(*v1.Pod)
- glog.V(4).Infof("recycler pod update received: %s %s/%s %s", event.Type, pod.Namespace, pod.Name, pod.Status.Phase)
- switch event.Type {
- case watch.Added, watch.Modified:
- if pod.Status.Phase == v1.PodSucceeded {
- // Recycle succeeded.
- return nil
- }
- if pod.Status.Phase == v1.PodFailed {
- if pod.Status.Message != "" {
- return fmt.Errorf(pod.Status.Message)
- } else {
- return fmt.Errorf("pod failed, pod.Status.Message unknown.")
- }
- }
-
- case watch.Deleted:
- return fmt.Errorf("recycler pod was deleted")
-
- case watch.Error:
- return fmt.Errorf("recycler pod watcher failed")
- }
-
- case *v1.Event:
- // Event received
- podEvent := event.Object.(*v1.Event)
- glog.V(4).Infof("recycler event received: %s %s/%s %s/%s %s", event.Type, podEvent.Namespace, podEvent.Name, podEvent.InvolvedObject.Namespace, podEvent.InvolvedObject.Name, podEvent.Message)
- if event.Type == watch.Added {
- recyclerClient.Event(podEvent.Type, podEvent.Message)
- }
- }
- }
-}
-
-// recyclerClient abstracts access to a Pod by providing a narrower interface.
-// This makes it easier to mock a client for testing.
-type recyclerClient interface {
- CreatePod(pod *v1.Pod) (*v1.Pod, error)
- GetPod(name, namespace string) (*v1.Pod, error)
- DeletePod(name, namespace string) error
- // WatchPod returns a ListWatch for watching a pod. The stopChannel is used
- // to close the reflector backing the watch. The caller is responsible for
- // derring a close on the channel to stop the reflector.
- WatchPod(name, namespace string, stopChannel chan struct{}) (<-chan watch.Event, error)
- // Event sends an event to the volume that is being recycled.
- Event(eventtype, message string)
-}
-
-func newRecyclerClient(client clientset.Interface, recorder RecycleEventRecorder) recyclerClient {
- return &realRecyclerClient{
- client,
- recorder,
- }
-}
-
-type realRecyclerClient struct {
- client clientset.Interface
- recorder RecycleEventRecorder
-}
-
-func (c *realRecyclerClient) CreatePod(pod *v1.Pod) (*v1.Pod, error) {
- return c.client.CoreV1().Pods(pod.Namespace).Create(pod)
-}
-
-func (c *realRecyclerClient) GetPod(name, namespace string) (*v1.Pod, error) {
- return c.client.CoreV1().Pods(namespace).Get(name, metav1.GetOptions{})
-}
-
-func (c *realRecyclerClient) DeletePod(name, namespace string) error {
- return c.client.CoreV1().Pods(namespace).Delete(name, nil)
-}
-
-func (c *realRecyclerClient) Event(eventtype, message string) {
- c.recorder(eventtype, message)
-}
-
-func (c *realRecyclerClient) WatchPod(name, namespace string, stopChannel chan struct{}) (<-chan watch.Event, error) {
- podSelector, err := fields.ParseSelector("metadata.name=" + name)
- if err != nil {
- return nil, err
- }
- options := metav1.ListOptions{
- FieldSelector: podSelector.String(),
- Watch: true,
- }
-
- podWatch, err := c.client.CoreV1().Pods(namespace).Watch(options)
- if err != nil {
- return nil, err
- }
-
- eventSelector, _ := fields.ParseSelector("involvedObject.name=" + name)
- eventWatch, err := c.client.CoreV1().Events(namespace).Watch(metav1.ListOptions{
- FieldSelector: eventSelector.String(),
- Watch: true,
- })
- if err != nil {
- podWatch.Stop()
- return nil, err
- }
-
- eventCh := make(chan watch.Event, 30)
-
- go func() {
- defer eventWatch.Stop()
- defer podWatch.Stop()
- defer close(eventCh)
- var podWatchChannelClosed bool
- var eventWatchChannelClosed bool
- for {
- select {
- case _ = <-stopChannel:
- return
-
- case podEvent, ok := <-podWatch.ResultChan():
- if !ok {
- podWatchChannelClosed = true
- } else {
- eventCh <- podEvent
- }
- case eventEvent, ok := <-eventWatch.ResultChan():
- if !ok {
- eventWatchChannelClosed = true
- } else {
- eventCh <- eventEvent
- }
- }
- if podWatchChannelClosed && eventWatchChannelClosed {
- break
- }
- }
- }()
-
- return eventCh, nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/volume.go b/vendor/k8s.io/kubernetes/pkg/volume/volume.go
deleted file mode 100644
index 471963556..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/volume.go
+++ /dev/null
@@ -1,273 +0,0 @@
-/*
-Copyright 2014 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "time"
-
- "k8s.io/api/core/v1"
- "k8s.io/apimachinery/pkg/api/resource"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/types"
-)
-
-// Volume represents a directory used by pods or hosts on a node. All method
-// implementations of methods in the volume interface must be idempotent.
-type Volume interface {
- // GetPath returns the path to which the volume should be mounted for the
- // pod.
- GetPath() string
-
- // MetricsProvider embeds methods for exposing metrics (e.g.
- // used, available space).
- MetricsProvider
-}
-
-// BlockVolume interface provides methods to generate global map path
-// and pod device map path.
-type BlockVolume interface {
- // GetGlobalMapPath returns a global map path which contains
- // symbolic links associated to a block device.
- // ex. plugins/kubernetes.io/{PluginName}/{DefaultKubeletVolumeDevicesDirName}/{volumePluginDependentPath}/{pod uuid}
- GetGlobalMapPath(spec *Spec) (string, error)
- // GetPodDeviceMapPath returns a pod device map path
- // and name of a symbolic link associated to a block device.
- // ex. pods/{podUid}}/{DefaultKubeletVolumeDevicesDirName}/{escapeQualifiedPluginName}/{volumeName}
- GetPodDeviceMapPath() (string, string)
-}
-
-// MetricsProvider exposes metrics (e.g. used,available space) related to a
-// Volume.
-type MetricsProvider interface {
- // GetMetrics returns the Metrics for the Volume. Maybe expensive for
- // some implementations.
- GetMetrics() (*Metrics, error)
-}
-
-// Metrics represents the used and available bytes of the Volume.
-type Metrics struct {
- // The time at which these stats were updated.
- Time metav1.Time
-
- // Used represents the total bytes used by the Volume.
- // Note: For block devices this maybe more than the total size of the files.
- Used *resource.Quantity
-
- // Capacity represents the total capacity (bytes) of the volume's
- // underlying storage. For Volumes that share a filesystem with the host
- // (e.g. emptydir, hostpath) this is the size of the underlying storage,
- // and will not equal Used + Available as the fs is shared.
- Capacity *resource.Quantity
-
- // Available represents the storage space available (bytes) for the
- // Volume. For Volumes that share a filesystem with the host (e.g.
- // emptydir, hostpath), this is the available space on the underlying
- // storage, and is shared with host processes and other Volumes.
- Available *resource.Quantity
-
- // InodesUsed represents the total inodes used by the Volume.
- InodesUsed *resource.Quantity
-
- // Inodes represents the total number of inodes available in the volume.
- // For volumes that share a filesystem with the host (e.g. emptydir, hostpath),
- // this is the inodes available in the underlying storage,
- // and will not equal InodesUsed + InodesFree as the fs is shared.
- Inodes *resource.Quantity
-
- // InodesFree represent the inodes available for the volume. For Volumes that share
- // a filesystem with the host (e.g. emptydir, hostpath), this is the free inodes
- // on the underlying storage, and is shared with host processes and other volumes
- InodesFree *resource.Quantity
-}
-
-// Attributes represents the attributes of this mounter.
-type Attributes struct {
- ReadOnly bool
- Managed bool
- SupportsSELinux bool
-}
-
-// Mounter interface provides methods to set up/mount the volume.
-type Mounter interface {
- // Uses Interface to provide the path for Docker binds.
- Volume
-
- // CanMount is called immediately prior to Setup to check if
- // the required components (binaries, etc.) are available on
- // the underlying node to complete the subsequent SetUp (mount)
- // operation. If CanMount returns error, the mount operation is
- // aborted and an event is generated indicating that the node
- // does not have the required binaries to complete the mount.
- // If CanMount succeeds, the mount operation continues
- // normally. The CanMount check can be enabled or disabled
- // using the experimental-check-mount-binaries binary flag
- CanMount() error
-
- // SetUp prepares and mounts/unpacks the volume to a
- // self-determined directory path. The mount point and its
- // content should be owned by 'fsGroup' so that it can be
- // accessed by the pod. This may be called more than once, so
- // implementations must be idempotent.
- SetUp(fsGroup *int64) error
- // SetUpAt prepares and mounts/unpacks the volume to the
- // specified directory path, which may or may not exist yet.
- // The mount point and its content should be owned by
- // 'fsGroup' so that it can be accessed by the pod. This may
- // be called more than once, so implementations must be
- // idempotent.
- SetUpAt(dir string, fsGroup *int64) error
- // GetAttributes returns the attributes of the mounter.
- GetAttributes() Attributes
-}
-
-// Unmounter interface provides methods to cleanup/unmount the volumes.
-type Unmounter interface {
- Volume
- // TearDown unmounts the volume from a self-determined directory and
- // removes traces of the SetUp procedure.
- TearDown() error
- // TearDown unmounts the volume from the specified directory and
- // removes traces of the SetUp procedure.
- TearDownAt(dir string) error
-}
-
-// BlockVolumeMapper interface provides methods to set up/map the volume.
-type BlockVolumeMapper interface {
- BlockVolume
- // SetUpDevice prepares the volume to a self-determined directory path,
- // which may or may not exist yet and returns combination of physical
- // device path of a block volume and error.
- // If the plugin is non-attachable, it should prepare the device
- // in /dev/ (or where appropriate) and return unique device path.
- // Unique device path across kubelet node reboot is required to avoid
- // unexpected block volume destruction.
- // If the plugin is attachable, it should not do anything here,
- // just return empty string for device path.
- // Instead, attachable plugin have to return unique device path
- // at attacher.Attach() and attacher.WaitForAttach().
- // This may be called more than once, so implementations must be idempotent.
- SetUpDevice() (string, error)
-}
-
-// BlockVolumeUnmapper interface provides methods to cleanup/unmap the volumes.
-type BlockVolumeUnmapper interface {
- BlockVolume
- // TearDownDevice removes traces of the SetUpDevice procedure under
- // a self-determined directory.
- // If the plugin is non-attachable, this method detaches the volume
- // from a node.
- TearDownDevice(mapPath string, devicePath string) error
-}
-
-// Provisioner is an interface that creates templates for PersistentVolumes
-// and can create the volume as a new resource in the infrastructure provider.
-type Provisioner interface {
- // Provision creates the resource by allocating the underlying volume in a
- // storage system. This method should block until completion and returns
- // PersistentVolume representing the created storage resource.
- Provision() (*v1.PersistentVolume, error)
-}
-
-// Deleter removes the resource from the underlying storage provider. Calls
-// to this method should block until the deletion is complete. Any error
-// returned indicates the volume has failed to be reclaimed. A nil return
-// indicates success.
-type Deleter interface {
- Volume
- // This method should block until completion.
- // deletedVolumeInUseError returned from this function will not be reported
- // as error and it will be sent as "Info" event to the PV being deleted. The
- // volume controller will retry deleting the volume in the next periodic
- // sync. This can be used to postpone deletion of a volume that is being
- // detached from a node. Deletion of such volume would fail anyway and such
- // error would confuse users.
- Delete() error
-}
-
-// Attacher can attach a volume to a node.
-type Attacher interface {
- // Attaches the volume specified by the given spec to the node with the given Name.
- // On success, returns the device path where the device was attached on the
- // node.
- Attach(spec *Spec, nodeName types.NodeName) (string, error)
-
- // VolumesAreAttached checks whether the list of volumes still attached to the specified
- // node. It returns a map which maps from the volume spec to the checking result.
- // If an error is occurred during checking, the error will be returned
- VolumesAreAttached(specs []*Spec, nodeName types.NodeName) (map[*Spec]bool, error)
-
- // WaitForAttach blocks until the device is attached to this
- // node. If it successfully attaches, the path to the device
- // is returned. Otherwise, if the device does not attach after
- // the given timeout period, an error will be returned.
- WaitForAttach(spec *Spec, devicePath string, pod *v1.Pod, timeout time.Duration) (string, error)
-
- // GetDeviceMountPath returns a path where the device should
- // be mounted after it is attached. This is a global mount
- // point which should be bind mounted for individual volumes.
- GetDeviceMountPath(spec *Spec) (string, error)
-
- // MountDevice mounts the disk to a global path which
- // individual pods can then bind mount
- MountDevice(spec *Spec, devicePath string, deviceMountPath string) error
-}
-
-type BulkVolumeVerifier interface {
- // BulkVerifyVolumes checks whether the list of volumes still attached to the
- // the clusters in the node. It returns a map which maps from the volume spec to the checking result.
- // If an error occurs during check - error should be returned and volume on nodes
- // should be assumed as still attached.
- BulkVerifyVolumes(volumesByNode map[types.NodeName][]*Spec) (map[types.NodeName]map[*Spec]bool, error)
-}
-
-// Detacher can detach a volume from a node.
-type Detacher interface {
- // Detach the given volume from the node with the given Name.
- // volumeName is name of the volume as returned from plugin's
- // GetVolumeName().
- Detach(volumeName string, nodeName types.NodeName) error
-
- // UnmountDevice unmounts the global mount of the disk. This
- // should only be called once all bind mounts have been
- // unmounted.
- UnmountDevice(deviceMountPath string) error
-}
-
-// NewDeletedVolumeInUseError returns a new instance of DeletedVolumeInUseError
-// error.
-func NewDeletedVolumeInUseError(message string) error {
- return deletedVolumeInUseError(message)
-}
-
-type deletedVolumeInUseError string
-
-var _ error = deletedVolumeInUseError("")
-
-// IsDeletedVolumeInUse returns true if an error returned from Delete() is
-// deletedVolumeInUseError
-func IsDeletedVolumeInUse(err error) bool {
- switch err.(type) {
- case deletedVolumeInUseError:
- return true
- default:
- return false
- }
-}
-
-func (err deletedVolumeInUseError) Error() string {
- return string(err)
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/volume_linux.go b/vendor/k8s.io/kubernetes/pkg/volume/volume_linux.go
deleted file mode 100644
index d67ee4a95..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/volume_linux.go
+++ /dev/null
@@ -1,105 +0,0 @@
-// +build linux
-
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-import (
- "path/filepath"
- "syscall"
-
- "os"
-
- "github.com/golang/glog"
-)
-
-const (
- rwMask = os.FileMode(0660)
- roMask = os.FileMode(0440)
-)
-
-// SetVolumeOwnership modifies the given volume to be owned by
-// fsGroup, and sets SetGid so that newly created files are owned by
-// fsGroup. If fsGroup is nil nothing is done.
-func SetVolumeOwnership(mounter Mounter, fsGroup *int64) error {
-
- if fsGroup == nil {
- return nil
- }
-
- return filepath.Walk(mounter.GetPath(), func(path string, info os.FileInfo, err error) error {
- if err != nil {
- return err
- }
-
- // chown and chmod pass through to the underlying file for symlinks.
- // Symlinks have a mode of 777 but this really doesn't mean anything.
- // The permissions of the underlying file are what matter.
- // However, if one reads the mode of a symlink then chmods the symlink
- // with that mode, it changes the mode of the underlying file, overridden
- // the defaultMode and permissions initialized by the volume plugin, which
- // is not what we want; thus, we skip chown/chmod for symlinks.
- if info.Mode()&os.ModeSymlink != 0 {
- return nil
- }
-
- stat, ok := info.Sys().(*syscall.Stat_t)
- if !ok {
- return nil
- }
-
- if stat == nil {
- glog.Errorf("Got nil stat_t for path %v while setting ownership of volume", path)
- return nil
- }
-
- err = os.Chown(path, int(stat.Uid), int(*fsGroup))
- if err != nil {
- glog.Errorf("Chown failed on %v: %v", path, err)
- }
-
- mask := rwMask
- if mounter.GetAttributes().ReadOnly {
- mask = roMask
- }
-
- if info.IsDir() {
- mask |= os.ModeSetgid
- }
-
- err = os.Chmod(path, info.Mode()|mask)
- if err != nil {
- glog.Errorf("Chmod failed on %v: %v", path, err)
- }
-
- return nil
- })
-}
-
-// IsSameFSGroup is called only for requests to mount an already mounted
-// volume. It checks if fsGroup of new mount request is the same or not.
-// It returns false if it not the same. It also returns current Gid of a path
-// provided for dir variable.
-func IsSameFSGroup(dir string, fsGroup int64) (bool, int, error) {
- info, err := os.Stat(dir)
- if err != nil {
- glog.Errorf("Error getting stats for %s (%v)", dir, err)
- return false, 0, err
- }
- s := info.Sys().(*syscall.Stat_t)
- return int(s.Gid) == int(fsGroup), int(s.Gid), nil
-}
diff --git a/vendor/k8s.io/kubernetes/pkg/volume/volume_unsupported.go b/vendor/k8s.io/kubernetes/pkg/volume/volume_unsupported.go
deleted file mode 100644
index 46a6aeaf0..000000000
--- a/vendor/k8s.io/kubernetes/pkg/volume/volume_unsupported.go
+++ /dev/null
@@ -1,27 +0,0 @@
-// +build !linux
-
-/*
-Copyright 2016 The Kubernetes Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-*/
-
-package volume
-
-func SetVolumeOwnership(mounter Mounter, fsGroup *int64) error {
- return nil
-}
-
-func IsSameFSGroup(dir string, fsGroup int64) (bool, int, error) {
- return true, int(fsGroup), nil
-}
diff --git a/vendor/k8s.io/kubernetes/third_party/forked/golang/LICENSE b/vendor/k8s.io/kubernetes/third_party/forked/golang/LICENSE
deleted file mode 100644
index 744875676..000000000
--- a/vendor/k8s.io/kubernetes/third_party/forked/golang/LICENSE
+++ /dev/null
@@ -1,27 +0,0 @@
-Copyright (c) 2012 The Go Authors. All rights reserved.
-
-Redistribution and use in source and binary forms, with or without
-modification, are permitted provided that the following conditions are
-met:
-
- * Redistributions of source code must retain the above copyright
-notice, this list of conditions and the following disclaimer.
- * Redistributions in binary form must reproduce the above
-copyright notice, this list of conditions and the following disclaimer
-in the documentation and/or other materials provided with the
-distribution.
- * Neither the name of Google Inc. nor the names of its
-contributors may be used to endorse or promote products derived from
-this software without specific prior written permission.
-
-THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
diff --git a/vendor/k8s.io/kubernetes/third_party/forked/golang/PATENTS b/vendor/k8s.io/kubernetes/third_party/forked/golang/PATENTS
deleted file mode 100644
index 733099041..000000000
--- a/vendor/k8s.io/kubernetes/third_party/forked/golang/PATENTS
+++ /dev/null
@@ -1,22 +0,0 @@
-Additional IP Rights Grant (Patents)
-
-"This implementation" means the copyrightable works distributed by
-Google as part of the Go project.
-
-Google hereby grants to You a perpetual, worldwide, non-exclusive,
-no-charge, royalty-free, irrevocable (except as stated in this section)
-patent license to make, have made, use, offer to sell, sell, import,
-transfer and otherwise run, modify and propagate the contents of this
-implementation of Go, where such license applies only to those patent
-claims, both currently owned or controlled by Google and acquired in
-the future, licensable by Google that are necessarily infringed by this
-implementation of Go. This grant does not include claims that would be
-infringed only as a consequence of further modification of this
-implementation. If you or your agent or exclusive licensee institute or
-order or agree to the institution of patent litigation against any
-entity (including a cross-claim or counterclaim in a lawsuit) alleging
-that this implementation of Go or any code incorporated within this
-implementation of Go constitutes direct or contributory patent
-infringement, or inducement of patent infringement, then any patent
-rights granted to you under this License for this implementation of Go
-shall terminate as of the date such litigation is filed.
diff --git a/vendor/k8s.io/kubernetes/third_party/forked/golang/expansion/expand.go b/vendor/k8s.io/kubernetes/third_party/forked/golang/expansion/expand.go
deleted file mode 100644
index 6bf0ea8ce..000000000
--- a/vendor/k8s.io/kubernetes/third_party/forked/golang/expansion/expand.go
+++ /dev/null
@@ -1,102 +0,0 @@
-package expansion
-
-import (
- "bytes"
-)
-
-const (
- operator = '$'
- referenceOpener = '('
- referenceCloser = ')'
-)
-
-// syntaxWrap returns the input string wrapped by the expansion syntax.
-func syntaxWrap(input string) string {
- return string(operator) + string(referenceOpener) + input + string(referenceCloser)
-}
-
-// MappingFuncFor returns a mapping function for use with Expand that
-// implements the expansion semantics defined in the expansion spec; it
-// returns the input string wrapped in the expansion syntax if no mapping
-// for the input is found.
-func MappingFuncFor(context ...map[string]string) func(string) string {
- return func(input string) string {
- for _, vars := range context {
- val, ok := vars[input]
- if ok {
- return val
- }
- }
-
- return syntaxWrap(input)
- }
-}
-
-// Expand replaces variable references in the input string according to
-// the expansion spec using the given mapping function to resolve the
-// values of variables.
-func Expand(input string, mapping func(string) string) string {
- var buf bytes.Buffer
- checkpoint := 0
- for cursor := 0; cursor < len(input); cursor++ {
- if input[cursor] == operator && cursor+1 < len(input) {
- // Copy the portion of the input string since the last
- // checkpoint into the buffer
- buf.WriteString(input[checkpoint:cursor])
-
- // Attempt to read the variable name as defined by the
- // syntax from the input string
- read, isVar, advance := tryReadVariableName(input[cursor+1:])
-
- if isVar {
- // We were able to read a variable name correctly;
- // apply the mapping to the variable name and copy the
- // bytes into the buffer
- buf.WriteString(mapping(read))
- } else {
- // Not a variable name; copy the read bytes into the buffer
- buf.WriteString(read)
- }
-
- // Advance the cursor in the input string to account for
- // bytes consumed to read the variable name expression
- cursor += advance
-
- // Advance the checkpoint in the input string
- checkpoint = cursor + 1
- }
- }
-
- // Return the buffer and any remaining unwritten bytes in the
- // input string.
- return buf.String() + input[checkpoint:]
-}
-
-// tryReadVariableName attempts to read a variable name from the input
-// string and returns the content read from the input, whether that content
-// represents a variable name to perform mapping on, and the number of bytes
-// consumed in the input string.
-//
-// The input string is assumed not to contain the initial operator.
-func tryReadVariableName(input string) (string, bool, int) {
- switch input[0] {
- case operator:
- // Escaped operator; return it.
- return input[0:1], false, 1
- case referenceOpener:
- // Scan to expression closer
- for i := 1; i < len(input); i++ {
- if input[i] == referenceCloser {
- return input[1:i], true, i + 1
- }
- }
-
- // Incomplete reference; return it.
- return string(operator) + string(referenceOpener), false, 1
- default:
- // Not the beginning of an expression, ie, an operator
- // that doesn't begin an expression. Return the operator
- // and the first rune in the string.
- return (string(operator) + string(input[0])), false, 1
- }
-}
diff --git a/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/compiler/plugin.proto b/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/compiler/plugin.proto
deleted file mode 100644
index 5116167e0..000000000
--- a/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/compiler/plugin.proto
+++ /dev/null
@@ -1,150 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-// Author: kenton@google.com (Kenton Varda)
-//
-// WARNING: The plugin interface is currently EXPERIMENTAL and is subject to
-// change.
-//
-// protoc (aka the Protocol Compiler) can be extended via plugins. A plugin is
-// just a program that reads a CodeGeneratorRequest from stdin and writes a
-// CodeGeneratorResponse to stdout.
-//
-// Plugins written using C++ can use google/protobuf/compiler/plugin.h instead
-// of dealing with the raw protocol defined here.
-//
-// A plugin executable needs only to be placed somewhere in the path. The
-// plugin should be named "protoc-gen-$NAME", and will then be used when the
-// flag "--${NAME}_out" is passed to protoc.
-
-syntax = "proto2";
-package google.protobuf.compiler;
-option java_package = "com.google.protobuf.compiler";
-option java_outer_classname = "PluginProtos";
-
-option go_package = "plugin_go";
-
-import "google/protobuf/descriptor.proto";
-
-// An encoded CodeGeneratorRequest is written to the plugin's stdin.
-message CodeGeneratorRequest {
- // The .proto files that were explicitly listed on the command-line. The
- // code generator should generate code only for these files. Each file's
- // descriptor will be included in proto_file, below.
- repeated string file_to_generate = 1;
-
- // The generator parameter passed on the command-line.
- optional string parameter = 2;
-
- // FileDescriptorProtos for all files in files_to_generate and everything
- // they import. The files will appear in topological order, so each file
- // appears before any file that imports it.
- //
- // protoc guarantees that all proto_files will be written after
- // the fields above, even though this is not technically guaranteed by the
- // protobuf wire format. This theoretically could allow a plugin to stream
- // in the FileDescriptorProtos and handle them one by one rather than read
- // the entire set into memory at once. However, as of this writing, this
- // is not similarly optimized on protoc's end -- it will store all fields in
- // memory at once before sending them to the plugin.
- repeated FileDescriptorProto proto_file = 15;
-}
-
-// The plugin writes an encoded CodeGeneratorResponse to stdout.
-message CodeGeneratorResponse {
- // Error message. If non-empty, code generation failed. The plugin process
- // should exit with status code zero even if it reports an error in this way.
- //
- // This should be used to indicate errors in .proto files which prevent the
- // code generator from generating correct code. Errors which indicate a
- // problem in protoc itself -- such as the input CodeGeneratorRequest being
- // unparseable -- should be reported by writing a message to stderr and
- // exiting with a non-zero status code.
- optional string error = 1;
-
- // Represents a single generated file.
- message File {
- // The file name, relative to the output directory. The name must not
- // contain "." or ".." components and must be relative, not be absolute (so,
- // the file cannot lie outside the output directory). "/" must be used as
- // the path separator, not "\".
- //
- // If the name is omitted, the content will be appended to the previous
- // file. This allows the generator to break large files into small chunks,
- // and allows the generated text to be streamed back to protoc so that large
- // files need not reside completely in memory at one time. Note that as of
- // this writing protoc does not optimize for this -- it will read the entire
- // CodeGeneratorResponse before writing files to disk.
- optional string name = 1;
-
- // If non-empty, indicates that the named file should already exist, and the
- // content here is to be inserted into that file at a defined insertion
- // point. This feature allows a code generator to extend the output
- // produced by another code generator. The original generator may provide
- // insertion points by placing special annotations in the file that look
- // like:
- // @@protoc_insertion_point(NAME)
- // The annotation can have arbitrary text before and after it on the line,
- // which allows it to be placed in a comment. NAME should be replaced with
- // an identifier naming the point -- this is what other generators will use
- // as the insertion_point. Code inserted at this point will be placed
- // immediately above the line containing the insertion point (thus multiple
- // insertions to the same point will come out in the order they were added).
- // The double-@ is intended to make it unlikely that the generated code
- // could contain things that look like insertion points by accident.
- //
- // For example, the C++ code generator places the following line in the
- // .pb.h files that it generates:
- // // @@protoc_insertion_point(namespace_scope)
- // This line appears within the scope of the file's package namespace, but
- // outside of any particular class. Another plugin can then specify the
- // insertion_point "namespace_scope" to generate additional classes or
- // other declarations that should be placed in this scope.
- //
- // Note that if the line containing the insertion point begins with
- // whitespace, the same whitespace will be added to every line of the
- // inserted text. This is useful for languages like Python, where
- // indentation matters. In these languages, the insertion point comment
- // should be indented the same amount as any inserted code will need to be
- // in order to work correctly in that context.
- //
- // The code generator that generates the initial file and the one which
- // inserts into it must both run as part of a single invocation of protoc.
- // Code generators are executed in the order in which they appear on the
- // command line.
- //
- // If |insertion_point| is present, |name| must also be present.
- optional string insertion_point = 2;
-
- // The file contents.
- optional string content = 15;
- }
- repeated File file = 15;
-} \ No newline at end of file
diff --git a/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/descriptor.proto b/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/descriptor.proto
deleted file mode 100644
index 14594a8c6..000000000
--- a/vendor/k8s.io/kubernetes/third_party/protobuf/google/protobuf/descriptor.proto
+++ /dev/null
@@ -1,779 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc. All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-// Author: kenton@google.com (Kenton Varda)
-// Based on original Protocol Buffers design by
-// Sanjay Ghemawat, Jeff Dean, and others.
-//
-// The messages in this file describe the definitions found in .proto files.
-// A valid .proto file can be translated directly to a FileDescriptorProto
-// without any other information (e.g. without reading its imports).
-
-
-syntax = "proto2";
-
-package google.protobuf;
-option go_package = "descriptor";
-option java_package = "com.google.protobuf";
-option java_outer_classname = "DescriptorProtos";
-option csharp_namespace = "Google.Protobuf.Reflection";
-option objc_class_prefix = "GPB";
-
-// descriptor.proto must be optimized for speed because reflection-based
-// algorithms don't work during bootstrapping.
-option optimize_for = SPEED;
-
-// The protocol compiler can output a FileDescriptorSet containing the .proto
-// files it parses.
-message FileDescriptorSet {
- repeated FileDescriptorProto file = 1;
-}
-
-// Describes a complete .proto file.
-message FileDescriptorProto {
- optional string name = 1; // file name, relative to root of source tree
- optional string package = 2; // e.g. "foo", "foo.bar", etc.
-
- // Names of files imported by this file.
- repeated string dependency = 3;
- // Indexes of the public imported files in the dependency list above.
- repeated int32 public_dependency = 10;
- // Indexes of the weak imported files in the dependency list.
- // For Google-internal migration only. Do not use.
- repeated int32 weak_dependency = 11;
-
- // All top-level definitions in this file.
- repeated DescriptorProto message_type = 4;
- repeated EnumDescriptorProto enum_type = 5;
- repeated ServiceDescriptorProto service = 6;
- repeated FieldDescriptorProto extension = 7;
-
- optional FileOptions options = 8;
-
- // This field contains optional information about the original source code.
- // You may safely remove this entire field without harming runtime
- // functionality of the descriptors -- the information is needed only by
- // development tools.
- optional SourceCodeInfo source_code_info = 9;
-
- // The syntax of the proto file.
- // The supported values are "proto2" and "proto3".
- optional string syntax = 12;
-}
-
-// Describes a message type.
-message DescriptorProto {
- optional string name = 1;
-
- repeated FieldDescriptorProto field = 2;
- repeated FieldDescriptorProto extension = 6;
-
- repeated DescriptorProto nested_type = 3;
- repeated EnumDescriptorProto enum_type = 4;
-
- message ExtensionRange {
- optional int32 start = 1;
- optional int32 end = 2;
- }
- repeated ExtensionRange extension_range = 5;
-
- repeated OneofDescriptorProto oneof_decl = 8;
-
- optional MessageOptions options = 7;
-
- // Range of reserved tag numbers. Reserved tag numbers may not be used by
- // fields or extension ranges in the same message. Reserved ranges may
- // not overlap.
- message ReservedRange {
- optional int32 start = 1; // Inclusive.
- optional int32 end = 2; // Exclusive.
- }
- repeated ReservedRange reserved_range = 9;
- // Reserved field names, which may not be used by fields in the same message.
- // A given name may only be reserved once.
- repeated string reserved_name = 10;
-}
-
-// Describes a field within a message.
-message FieldDescriptorProto {
- enum Type {
- // 0 is reserved for errors.
- // Order is weird for historical reasons.
- TYPE_DOUBLE = 1;
- TYPE_FLOAT = 2;
- // Not ZigZag encoded. Negative numbers take 10 bytes. Use TYPE_SINT64 if
- // negative values are likely.
- TYPE_INT64 = 3;
- TYPE_UINT64 = 4;
- // Not ZigZag encoded. Negative numbers take 10 bytes. Use TYPE_SINT32 if
- // negative values are likely.
- TYPE_INT32 = 5;
- TYPE_FIXED64 = 6;
- TYPE_FIXED32 = 7;
- TYPE_BOOL = 8;
- TYPE_STRING = 9;
- TYPE_GROUP = 10; // Tag-delimited aggregate.
- TYPE_MESSAGE = 11; // Length-delimited aggregate.
-
- // New in version 2.
- TYPE_BYTES = 12;
- TYPE_UINT32 = 13;
- TYPE_ENUM = 14;
- TYPE_SFIXED32 = 15;
- TYPE_SFIXED64 = 16;
- TYPE_SINT32 = 17; // Uses ZigZag encoding.
- TYPE_SINT64 = 18; // Uses ZigZag encoding.
- };
-
- enum Label {
- // 0 is reserved for errors
- LABEL_OPTIONAL = 1;
- LABEL_REQUIRED = 2;
- LABEL_REPEATED = 3;
- // TODO(sanjay): Should we add LABEL_MAP?
- };
-
- optional string name = 1;
- optional int32 number = 3;
- optional Label label = 4;
-
- // If type_name is set, this need not be set. If both this and type_name
- // are set, this must be one of TYPE_ENUM, TYPE_MESSAGE or TYPE_GROUP.
- optional Type type = 5;
-
- // For message and enum types, this is the name of the type. If the name
- // starts with a '.', it is fully-qualified. Otherwise, C++-like scoping
- // rules are used to find the type (i.e. first the nested types within this
- // message are searched, then within the parent, on up to the root
- // namespace).
- optional string type_name = 6;
-
- // For extensions, this is the name of the type being extended. It is
- // resolved in the same manner as type_name.
- optional string extendee = 2;
-
- // For numeric types, contains the original text representation of the value.
- // For booleans, "true" or "false".
- // For strings, contains the default text contents (not escaped in any way).
- // For bytes, contains the C escaped value. All bytes >= 128 are escaped.
- // TODO(kenton): Base-64 encode?
- optional string default_value = 7;
-
- // If set, gives the index of a oneof in the containing type's oneof_decl
- // list. This field is a member of that oneof.
- optional int32 oneof_index = 9;
-
- // JSON name of this field. The value is set by protocol compiler. If the
- // user has set a "json_name" option on this field, that option's value
- // will be used. Otherwise, it's deduced from the field's name by converting
- // it to camelCase.
- optional string json_name = 10;
-
- optional FieldOptions options = 8;
-}
-
-// Describes a oneof.
-message OneofDescriptorProto {
- optional string name = 1;
-}
-
-// Describes an enum type.
-message EnumDescriptorProto {
- optional string name = 1;
-
- repeated EnumValueDescriptorProto value = 2;
-
- optional EnumOptions options = 3;
-}
-
-// Describes a value within an enum.
-message EnumValueDescriptorProto {
- optional string name = 1;
- optional int32 number = 2;
-
- optional EnumValueOptions options = 3;
-}
-
-// Describes a service.
-message ServiceDescriptorProto {
- optional string name = 1;
- repeated MethodDescriptorProto method = 2;
-
- optional ServiceOptions options = 3;
-}
-
-// Describes a method of a service.
-message MethodDescriptorProto {
- optional string name = 1;
-
- // Input and output type names. These are resolved in the same way as
- // FieldDescriptorProto.type_name, but must refer to a message type.
- optional string input_type = 2;
- optional string output_type = 3;
-
- optional MethodOptions options = 4;
-
- // Identifies if client streams multiple client messages
- optional bool client_streaming = 5 [default=false];
- // Identifies if server streams multiple server messages
- optional bool server_streaming = 6 [default=false];
-}
-
-
-// ===================================================================
-// Options
-
-// Each of the definitions above may have "options" attached. These are
-// just annotations which may cause code to be generated slightly differently
-// or may contain hints for code that manipulates protocol messages.
-//
-// Clients may define custom options as extensions of the *Options messages.
-// These extensions may not yet be known at parsing time, so the parser cannot
-// store the values in them. Instead it stores them in a field in the *Options
-// message called uninterpreted_option. This field must have the same name
-// across all *Options messages. We then use this field to populate the
-// extensions when we build a descriptor, at which point all protos have been
-// parsed and so all extensions are known.
-//
-// Extension numbers for custom options may be chosen as follows:
-// * For options which will only be used within a single application or
-// organization, or for experimental options, use field numbers 50000
-// through 99999. It is up to you to ensure that you do not use the
-// same number for multiple options.
-// * For options which will be published and used publicly by multiple
-// independent entities, e-mail protobuf-global-extension-registry@google.com
-// to reserve extension numbers. Simply provide your project name (e.g.
-// Objective-C plugin) and your project website (if available) -- there's no
-// need to explain how you intend to use them. Usually you only need one
-// extension number. You can declare multiple options with only one extension
-// number by putting them in a sub-message. See the Custom Options section of
-// the docs for examples:
-// https://developers.google.com/protocol-buffers/docs/proto#options
-// If this turns out to be popular, a web service will be set up
-// to automatically assign option numbers.
-
-
-message FileOptions {
-
- // Sets the Java package where classes generated from this .proto will be
- // placed. By default, the proto package is used, but this is often
- // inappropriate because proto packages do not normally start with backwards
- // domain names.
- optional string java_package = 1;
-
-
- // If set, all the classes from the .proto file are wrapped in a single
- // outer class with the given name. This applies to both Proto1
- // (equivalent to the old "--one_java_file" option) and Proto2 (where
- // a .proto always translates to a single class, but you may want to
- // explicitly choose the class name).
- optional string java_outer_classname = 8;
-
- // If set true, then the Java code generator will generate a separate .java
- // file for each top-level message, enum, and service defined in the .proto
- // file. Thus, these types will *not* be nested inside the outer class
- // named by java_outer_classname. However, the outer class will still be
- // generated to contain the file's getDescriptor() method as well as any
- // top-level extensions defined in the file.
- optional bool java_multiple_files = 10 [default=false];
-
- // If set true, then the Java code generator will generate equals() and
- // hashCode() methods for all messages defined in the .proto file.
- // This increases generated code size, potentially substantially for large
- // protos, which may harm a memory-constrained application.
- // - In the full runtime this is a speed optimization, as the
- // AbstractMessage base class includes reflection-based implementations of
- // these methods.
- // - In the lite runtime, setting this option changes the semantics of
- // equals() and hashCode() to more closely match those of the full runtime;
- // the generated methods compute their results based on field values rather
- // than object identity. (Implementations should not assume that hashcodes
- // will be consistent across runtimes or versions of the protocol compiler.)
- optional bool java_generate_equals_and_hash = 20 [default=false];
-
- // If set true, then the Java2 code generator will generate code that
- // throws an exception whenever an attempt is made to assign a non-UTF-8
- // byte sequence to a string field.
- // Message reflection will do the same.
- // However, an extension field still accepts non-UTF-8 byte sequences.
- // This option has no effect on when used with the lite runtime.
- optional bool java_string_check_utf8 = 27 [default=false];
-
-
- // Generated classes can be optimized for speed or code size.
- enum OptimizeMode {
- SPEED = 1; // Generate complete code for parsing, serialization,
- // etc.
- CODE_SIZE = 2; // Use ReflectionOps to implement these methods.
- LITE_RUNTIME = 3; // Generate code using MessageLite and the lite runtime.
- }
- optional OptimizeMode optimize_for = 9 [default=SPEED];
-
- // Sets the Go package where structs generated from this .proto will be
- // placed. If omitted, the Go package will be derived from the following:
- // - The basename of the package import path, if provided.
- // - Otherwise, the package statement in the .proto file, if present.
- // - Otherwise, the basename of the .proto file, without extension.
- optional string go_package = 11;
-
-
-
- // Should generic services be generated in each language? "Generic" services
- // are not specific to any particular RPC system. They are generated by the
- // main code generators in each language (without additional plugins).
- // Generic services were the only kind of service generation supported by
- // early versions of google.protobuf.
- //
- // Generic services are now considered deprecated in favor of using plugins
- // that generate code specific to your particular RPC system. Therefore,
- // these default to false. Old code which depends on generic services should
- // explicitly set them to true.
- optional bool cc_generic_services = 16 [default=false];
- optional bool java_generic_services = 17 [default=false];
- optional bool py_generic_services = 18 [default=false];
-
- // Is this file deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for everything in the file, or it will be completely ignored; in the very
- // least, this is a formalization for deprecating files.
- optional bool deprecated = 23 [default=false];
-
- // Enables the use of arenas for the proto messages in this file. This applies
- // only to generated classes for C++.
- optional bool cc_enable_arenas = 31 [default=false];
-
-
- // Sets the objective c class prefix which is prepended to all objective c
- // generated classes from this .proto. There is no default.
- optional string objc_class_prefix = 36;
-
- // Namespace for generated classes; defaults to the package.
- optional string csharp_namespace = 37;
-
- // Whether the nano proto compiler should generate in the deprecated non-nano
- // suffixed package.
- optional bool javanano_use_deprecated_package = 38;
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message MessageOptions {
- // Set true to use the old proto1 MessageSet wire format for extensions.
- // This is provided for backwards-compatibility with the MessageSet wire
- // format. You should not use this for any other reason: It's less
- // efficient, has fewer features, and is more complicated.
- //
- // The message must be defined exactly as follows:
- // message Foo {
- // option message_set_wire_format = true;
- // extensions 4 to max;
- // }
- // Note that the message cannot have any defined fields; MessageSets only
- // have extensions.
- //
- // All extensions of your type must be singular messages; e.g. they cannot
- // be int32s, enums, or repeated messages.
- //
- // Because this is an option, the above two restrictions are not enforced by
- // the protocol compiler.
- optional bool message_set_wire_format = 1 [default=false];
-
- // Disables the generation of the standard "descriptor()" accessor, which can
- // conflict with a field of the same name. This is meant to make migration
- // from proto1 easier; new code should avoid fields named "descriptor".
- optional bool no_standard_descriptor_accessor = 2 [default=false];
-
- // Is this message deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for the message, or it will be completely ignored; in the very least,
- // this is a formalization for deprecating messages.
- optional bool deprecated = 3 [default=false];
-
- // Whether the message is an automatically generated map entry type for the
- // maps field.
- //
- // For maps fields:
- // map<KeyType, ValueType> map_field = 1;
- // The parsed descriptor looks like:
- // message MapFieldEntry {
- // option map_entry = true;
- // optional KeyType key = 1;
- // optional ValueType value = 2;
- // }
- // repeated MapFieldEntry map_field = 1;
- //
- // Implementations may choose not to generate the map_entry=true message, but
- // use a native map in the target language to hold the keys and values.
- // The reflection APIs in such implementions still need to work as
- // if the field is a repeated message field.
- //
- // NOTE: Do not set the option in .proto files. Always use the maps syntax
- // instead. The option should only be implicitly set by the proto compiler
- // parser.
- optional bool map_entry = 7;
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message FieldOptions {
- // The ctype option instructs the C++ code generator to use a different
- // representation of the field than it normally would. See the specific
- // options below. This option is not yet implemented in the open source
- // release -- sorry, we'll try to include it in a future version!
- optional CType ctype = 1 [default = STRING];
- enum CType {
- // Default mode.
- STRING = 0;
-
- CORD = 1;
-
- STRING_PIECE = 2;
- }
- // The packed option can be enabled for repeated primitive fields to enable
- // a more efficient representation on the wire. Rather than repeatedly
- // writing the tag and type for each element, the entire array is encoded as
- // a single length-delimited blob. In proto3, only explicit setting it to
- // false will avoid using packed encoding.
- optional bool packed = 2;
-
-
- // The jstype option determines the JavaScript type used for values of the
- // field. The option is permitted only for 64 bit integral and fixed types
- // (int64, uint64, sint64, fixed64, sfixed64). By default these types are
- // represented as JavaScript strings. This avoids loss of precision that can
- // happen when a large value is converted to a floating point JavaScript
- // numbers. Specifying JS_NUMBER for the jstype causes the generated
- // JavaScript code to use the JavaScript "number" type instead of strings.
- // This option is an enum to permit additional types to be added,
- // e.g. goog.math.Integer.
- optional JSType jstype = 6 [default = JS_NORMAL];
- enum JSType {
- // Use the default type.
- JS_NORMAL = 0;
-
- // Use JavaScript strings.
- JS_STRING = 1;
-
- // Use JavaScript numbers.
- JS_NUMBER = 2;
- }
-
- // Should this field be parsed lazily? Lazy applies only to message-type
- // fields. It means that when the outer message is initially parsed, the
- // inner message's contents will not be parsed but instead stored in encoded
- // form. The inner message will actually be parsed when it is first accessed.
- //
- // This is only a hint. Implementations are free to choose whether to use
- // eager or lazy parsing regardless of the value of this option. However,
- // setting this option true suggests that the protocol author believes that
- // using lazy parsing on this field is worth the additional bookkeeping
- // overhead typically needed to implement it.
- //
- // This option does not affect the public interface of any generated code;
- // all method signatures remain the same. Furthermore, thread-safety of the
- // interface is not affected by this option; const methods remain safe to
- // call from multiple threads concurrently, while non-const methods continue
- // to require exclusive access.
- //
- //
- // Note that implementations may choose not to check required fields within
- // a lazy sub-message. That is, calling IsInitialized() on the outher message
- // may return true even if the inner message has missing required fields.
- // This is necessary because otherwise the inner message would have to be
- // parsed in order to perform the check, defeating the purpose of lazy
- // parsing. An implementation which chooses not to check required fields
- // must be consistent about it. That is, for any particular sub-message, the
- // implementation must either *always* check its required fields, or *never*
- // check its required fields, regardless of whether or not the message has
- // been parsed.
- optional bool lazy = 5 [default=false];
-
- // Is this field deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for accessors, or it will be completely ignored; in the very least, this
- // is a formalization for deprecating fields.
- optional bool deprecated = 3 [default=false];
-
- // For Google-internal migration only. Do not use.
- optional bool weak = 10 [default=false];
-
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message EnumOptions {
-
- // Set this option to true to allow mapping different tag names to the same
- // value.
- optional bool allow_alias = 2;
-
- // Is this enum deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for the enum, or it will be completely ignored; in the very least, this
- // is a formalization for deprecating enums.
- optional bool deprecated = 3 [default=false];
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message EnumValueOptions {
- // Is this enum value deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for the enum value, or it will be completely ignored; in the very least,
- // this is a formalization for deprecating enum values.
- optional bool deprecated = 1 [default=false];
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message ServiceOptions {
-
- // Note: Field numbers 1 through 32 are reserved for Google's internal RPC
- // framework. We apologize for hoarding these numbers to ourselves, but
- // we were already using them long before we decided to release Protocol
- // Buffers.
-
- // Is this service deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for the service, or it will be completely ignored; in the very least,
- // this is a formalization for deprecating services.
- optional bool deprecated = 33 [default=false];
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-message MethodOptions {
-
- // Note: Field numbers 1 through 32 are reserved for Google's internal RPC
- // framework. We apologize for hoarding these numbers to ourselves, but
- // we were already using them long before we decided to release Protocol
- // Buffers.
-
- // Is this method deprecated?
- // Depending on the target platform, this can emit Deprecated annotations
- // for the method, or it will be completely ignored; in the very least,
- // this is a formalization for deprecating methods.
- optional bool deprecated = 33 [default=false];
-
- // The parser stores options it doesn't recognize here. See above.
- repeated UninterpretedOption uninterpreted_option = 999;
-
- // Clients can define custom options in extensions of this message. See above.
- extensions 1000 to max;
-}
-
-
-// A message representing a option the parser does not recognize. This only
-// appears in options protos created by the compiler::Parser class.
-// DescriptorPool resolves these when building Descriptor objects. Therefore,
-// options protos in descriptor objects (e.g. returned by Descriptor::options(),
-// or produced by Descriptor::CopyTo()) will never have UninterpretedOptions
-// in them.
-message UninterpretedOption {
- // The name of the uninterpreted option. Each string represents a segment in
- // a dot-separated name. is_extension is true iff a segment represents an
- // extension (denoted with parentheses in options specs in .proto files).
- // E.g.,{ ["foo", false], ["bar.baz", true], ["qux", false] } represents
- // "foo.(bar.baz).qux".
- message NamePart {
- required string name_part = 1;
- required bool is_extension = 2;
- }
- repeated NamePart name = 2;
-
- // The value of the uninterpreted option, in whatever type the tokenizer
- // identified it as during parsing. Exactly one of these should be set.
- optional string identifier_value = 3;
- optional uint64 positive_int_value = 4;
- optional int64 negative_int_value = 5;
- optional double double_value = 6;
- optional bytes string_value = 7;
- optional string aggregate_value = 8;
-}
-
-// ===================================================================
-// Optional source code info
-
-// Encapsulates information about the original source file from which a
-// FileDescriptorProto was generated.
-message SourceCodeInfo {
- // A Location identifies a piece of source code in a .proto file which
- // corresponds to a particular definition. This information is intended
- // to be useful to IDEs, code indexers, documentation generators, and similar
- // tools.
- //
- // For example, say we have a file like:
- // message Foo {
- // optional string foo = 1;
- // }
- // Let's look at just the field definition:
- // optional string foo = 1;
- // ^ ^^ ^^ ^ ^^^
- // a bc de f ghi
- // We have the following locations:
- // span path represents
- // [a,i) [ 4, 0, 2, 0 ] The whole field definition.
- // [a,b) [ 4, 0, 2, 0, 4 ] The label (optional).
- // [c,d) [ 4, 0, 2, 0, 5 ] The type (string).
- // [e,f) [ 4, 0, 2, 0, 1 ] The name (foo).
- // [g,h) [ 4, 0, 2, 0, 3 ] The number (1).
- //
- // Notes:
- // - A location may refer to a repeated field itself (i.e. not to any
- // particular index within it). This is used whenever a set of elements are
- // logically enclosed in a single code segment. For example, an entire
- // extend block (possibly containing multiple extension definitions) will
- // have an outer location whose path refers to the "extensions" repeated
- // field without an index.
- // - Multiple locations may have the same path. This happens when a single
- // logical declaration is spread out across multiple places. The most
- // obvious example is the "extend" block again -- there may be multiple
- // extend blocks in the same scope, each of which will have the same path.
- // - A location's span is not always a subset of its parent's span. For
- // example, the "extendee" of an extension declaration appears at the
- // beginning of the "extend" block and is shared by all extensions within
- // the block.
- // - Just because a location's span is a subset of some other location's span
- // does not mean that it is a descendent. For example, a "group" defines
- // both a type and a field in a single declaration. Thus, the locations
- // corresponding to the type and field and their components will overlap.
- // - Code which tries to interpret locations should probably be designed to
- // ignore those that it doesn't understand, as more types of locations could
- // be recorded in the future.
- repeated Location location = 1;
- message Location {
- // Identifies which part of the FileDescriptorProto was defined at this
- // location.
- //
- // Each element is a field number or an index. They form a path from
- // the root FileDescriptorProto to the place where the definition. For
- // example, this path:
- // [ 4, 3, 2, 7, 1 ]
- // refers to:
- // file.message_type(3) // 4, 3
- // .field(7) // 2, 7
- // .name() // 1
- // This is because FileDescriptorProto.message_type has field number 4:
- // repeated DescriptorProto message_type = 4;
- // and DescriptorProto.field has field number 2:
- // repeated FieldDescriptorProto field = 2;
- // and FieldDescriptorProto.name has field number 1:
- // optional string name = 1;
- //
- // Thus, the above path gives the location of a field name. If we removed
- // the last element:
- // [ 4, 3, 2, 7 ]
- // this path refers to the whole field declaration (from the beginning
- // of the label to the terminating semicolon).
- repeated int32 path = 1 [packed=true];
-
- // Always has exactly three or four elements: start line, start column,
- // end line (optional, otherwise assumed same as start line), end column.
- // These are packed into a single field for efficiency. Note that line
- // and column numbers are zero-based -- typically you will want to add
- // 1 to each before displaying to a user.
- repeated int32 span = 2 [packed=true];
-
- // If this SourceCodeInfo represents a complete declaration, these are any
- // comments appearing before and after the declaration which appear to be
- // attached to the declaration.
- //
- // A series of line comments appearing on consecutive lines, with no other
- // tokens appearing on those lines, will be treated as a single comment.
- //
- // leading_detached_comments will keep paragraphs of comments that appear
- // before (but not connected to) the current element. Each paragraph,
- // separated by empty lines, will be one comment element in the repeated
- // field.
- //
- // Only the comment content is provided; comment markers (e.g. //) are
- // stripped out. For block comments, leading whitespace and an asterisk
- // will be stripped from the beginning of each line other than the first.
- // Newlines are included in the output.
- //
- // Examples:
- //
- // optional int32 foo = 1; // Comment attached to foo.
- // // Comment attached to bar.
- // optional int32 bar = 2;
- //
- // optional string baz = 3;
- // // Comment attached to baz.
- // // Another line attached to baz.
- //
- // // Comment attached to qux.
- // //
- // // Another line attached to qux.
- // optional double qux = 4;
- //
- // // Detached comment for corge. This is not leading or trailing comments
- // // to qux or corge because there are blank lines separating it from
- // // both.
- //
- // // Detached comment for corge paragraph 2.
- //
- // optional string corge = 5;
- // /* Block comment attached
- // * to corge. Leading asterisks
- // * will be removed. */
- // /* Block comment attached to
- // * grault. */
- // optional int32 grault = 6;
- //
- // // ignored detached comments.
- optional string leading_comments = 3;
- optional string trailing_comments = 4;
- repeated string leading_detached_comments = 6;
- }
-} \ No newline at end of file