By default, Kubernetes has a certificate validity period of 1 years when initializing a cluster. Manually generating certificates can avoid this problem.
Pull git code
git clone https://github.com/fandaye/k8s-tls.git CD k8s-tls/2. Edit the configuration file ' Apiserver.json ' file in the Hosts section, add the corresponding Kubernetes master node hostname and IP address to ', ' number interv
1. First to the outside network to download the image to pour into the various nodes2. Download the Yaml file and create the appMKDIR-P ~/k8s/heapster cd ~/k8s/heapster wget https://raw.githubusercontent.com/kubernetes/heapster/master/deploy/kube-config/influxdb/ Grafana.yaml wget https:// Raw.githubusercontent.com/kubernetes/heapster/master/ Deploy/
Brief introductionIt is a new, container-based, distributed solution that solves the problem of load balancing and deployment in traditional system architectures based on a powerful automation mechanism, thus saving 30% development costs, followed by a complete clustering capability,Including service registration, service discovery, fault detection and repair, service rolling upgrade, online expansion, health monitoring and other capabilities. Using Kubernet
Planning Nodes Install ubuntu 14.04 LTS Prepare password-less SSH login Establish Ssh-key certificate, switch to root account, use command ssh-keygen-t RSAGenerating public/private RSA key pair.Enter file in which to save the key (/ROOT/.SSH/ID_RSA):Enter passphrase (empty for no passphrase):Enter same Passphrase again:Your identification has been saved In/root/.ssh/id_rsa.Your public key has been saved in/root/.ssh/id_rsa.pub. Distribute Id_rsa.pub to authorized_keys inside The root directory
Environment Construction _c5-6 Adjusting the micro-service adaptation Mesos5-7 Micro-Service Deployment _a5-8 Micro-Service Deployment _b5-9 Micro-Service Deployment _c6th Chapter Service Orchestration-dockerswarmThe process of learning is similar to Mesos, which is to start with the concept, understand the architecture design of swarm, understand its service discovery, service orchestration, and then build swarm cluster environment, then adjust our microservices to adapt to swarm, and finally
with the conditions of the dispatch. 1) Nodeselector: Directed scheduling Kubernetes Master Scheduler service (Kube-scheduler process) is responsible for the implementation of POD scheduling, the entire process through a series of complex algorithms, Finally, the optimal target node is calculated for each pod, and we often do not know which node the pod will eventually be dispatched to. In fact, we need to
:
-Mountpath: /etc/ssl/certs
name:ca-certificates
readonly:true
-Mountpath:/varName:Grafana-Storage env:
-Name:Influxdb_host value:Monitoring-Influxdb-Name:Gf_server_http_port value: " the"# The following env variables is required to make Grafana accessible via # the Kubernetes API-ServerProxy. on Production clusters,We recommend # Removing these env variables,Setup Auth forGrafana,and expose the Grafana # service using a loadbalancer
mountpath:/logs
volumes:
-name:app-logs Emptudir
: {}
View Log kubectl logs 5. Configuration management of Pod
The Kubernetes v1.2 version provides a unified cluster configuration management solution –configmap. 5.1. Configmap: Configuration management for container applications
Usage Scenario: Live as an environment variable within a container. Set startup parameters for the container startup command (set to environment variable). Mou
Build kubernetes cluster, there is a default dashboard, but this dashboard is relatively humble, can not be customized display. So we're going to use GRAFANA+HEAPSTER+INFLUXDB to build an integrated monitoring platform.What's the first thing? Get the source code from Https://github.com/kubernetes/heapster, in the Heapster/deploy/kube-config/influxdb folder will a
-controller.yamlkubectl create-f Dashboard-service.yamlOnce started, you can view the nodeport of the map by following these instructions:[[email protected] kubernetes]# kubectl Get svc--namespace=kube-systemname cluster-ip External-ip PORT (S) agekube-dns 10.254.0.100 The following access addresses are available for access:http://10.5.10.116:8080/uihttp://10.5.10.116:30097Dashboard interface650) this.width
: containers: -name:kube-tomcat image:docker.io/chenyufeng/tomcat-centos imagepullpolicy:always ports: -containerport:8080 nbSp;protocol:tcp args: # uncomment the following line to manually specify Kubernetes API server Host NBSP;NBSP;NBSP;N bsp;# If not specified, Dashboard'll attempt to auto discover the API server and con Nect # to it. Uncomment only if the default is does not work. ---apiserver-host=http://47.97.90.38:8080## Please revise for
Tail-f/var/log/messagenov1409:39:09imagejournal:w111409:39:09.943633 1controller.go:386]Resettingendpointsformasterservice " Kubernetes "to{{}{kubernetesdefault/api/v1/namespaces/default/ Endpoints/kubernetes17553a6b-c907-11e7-b180-fa163ee5f1502002302017-11-1406:43:15 +0000utcWorkaround:Append the--apiserver-count=n to the Api-server configuration file, and then restart the ApiserverThis article is from the "ipcpu" blog, make sure to keep this source
the style and logic of the above code is no longer familiar: Create a The Schedulerserver object, which passes command-line arguments, and enters the schedulerserver Run method, does not die. --"kubernetes authoritative guide" source code guidance Kubernetes Scheduler Server is implemented by the Kube-scheduler process, which runs on the
This article will introduce the creation and use of Kube-ui (also called dashboard) on the basis of the previous article. This thing is essentially the WebUI connection Master API interface, through the API to obtain information about the k8s cluster, and then on the web display, for the user is more friendly, the actual use is not very large.One, download the image file and incorporate the unified management of the local warehouse# docker pull docker
Deploy Kubernetes-dashboard on the basis of the deployment of Kubernetes, which is the official Kubernets cluster Visualizer for user management, and the deployment dashboard is actually the same as deploying other applications in Kubernets, Create dashboard required replication Controller, service services can be;Note that the overlay network of Kubernets is working properly, each node has started flannel,
spark2.3.0+kubernetes Application Deployment
Spark can be run in Kubernetes managed clusters, using native kubernetes scheduling features have been added to spark. At present, kubernetes scheduling is experimental, in future versions, Spark may have behavioral changes in configuration, container images, and portals.
(1
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.