Skip to content

Commit 4578ca4

Browse files
authored
Update README.md
Added note about size of volume(s)
1 parent 98f9693 commit 4578ca4

1 file changed

Lines changed: 3 additions & 1 deletion

File tree

  • samples/features/sql-big-data-cluster/deployment/kubeadm

samples/features/sql-big-data-cluster/deployment/kubeadm/README.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,8 @@ To install docker, you can follow the steps below:
5454
1. Execute [ubuntu/setup-k8s-prereqs.sh](ubuntu/setup-k8s-prereqs.sh/) script on each machine
5555
1. Execute [ubuntu/setup-k8s-master.sh](ubuntu/setup-k8s-master.sh/) script on the machine designated as Kubernetes master
5656
1. After successful initialization of the Kubernetes master, follow the kubeadm join commands output by the setup script on each agent machine
57-
1. Execute [ubuntu/setup-volumes.sh](ubuntu/setup-volumes.sh/) script on each agent machine to create volumes for local storage. ***NOTE: Ensure there is sufficient local storage on your agents. Each volume will use up to 6GB by default. The script creates 25 volumes. Not all of the volumes will be used since it depends on the number of pods being deployed on each agent node. It is recommended to have at least 200 GB of storage on the agent nodes. ***
57+
1. Execute [ubuntu/setup-volumes.sh](ubuntu/setup-volumes.sh/) script on each agent machine to create volumes for local storage
58+
59+
**NOTE: Ensure there is sufficient local storage on your agents. Each volume will use up to 6GB by default. The script creates 25 volumes. Not all of the volumes will be used since it depends on the number of pods being deployed on each agent node. It is recommended to have at least 200 GB of storage on the agent nodes**
5860
1. Execute ***kubectl apply -f ubuntu/local-storage-provisioner.yaml*** against the Kubernetes cluster to create the local storage provisioner. You can now use persistent volumes by setting the USE_PERSISTENT_VOLUME environment variable to ***true*** and STORAGE_CLASS_NAME environment variable to ***local-storage*** during deployment
5961
1. Now, you can deploy the SQL Server 2019 big data cluster following instructions [here](https://docs.microsoft.com/en-us/sql/big-data-cluster/deployment-guidance?view=sqlallproducts-allversions)

0 commit comments

Comments
 (0)