Connecting a Spring Boot Application to YugabyteDB Managed and Deploying It on Amazon Elastic Kubernetes Service (EKS)
Spring Boot is an extremely popular framework for building cloud native applications. It makes configuring an application easy and offers lots of starters to get you off the ground quickly. Each Spring Boot application is stand-alone and self-contained, which makes them easy to deploy in a distributed fashion – to containers or, even better, on Kubernetes.
YugabyteDB Managed (formerly Yugabyte Cloud) is a perfect match for Spring Boot applications, especially ones made highly available with Kubernetes. YugabyteDB Managed provides you with a PostgreSQL-compatible database that is resilient and scalable, just like the applications you run on k8s. You can scale your YugabyteDB cluster out (or in) with just a few clicks. So you can spend less time worrying about running your database and more time focused on building better software. It’s a cloud native distributed SQL database for cloud native applications. Sign up for a free tier of YugabyteDB Managed today.
In this post, we’ll walk you through how to connect a Spring Boot application to YugabyteDB Managed and deploy it on Kubernetes via Amazon Elastic Kubernetes Service (EKS). You can walk through our other blogs on how to connect a Spring Boot application to YugabyteDB Managed and deploy it on Kubernetes via Google Kubernetes Engine (GKE) or on minikube. We’ll be using a slightly updated version of the popular Spring Boot PetClinic sample application that has a profile making it compatible with YugabyteDB. The repository for this is at Github.
In this walkthrough, you will:
- Create a free YugabyteDB Managed account and database cluster
- Download the Spring Boot PetClinic sample application and connect it to YugabyteDB Managed
- Containerize the Spring Boot PetClinic sample application
- Deploy the Spring Boot PetClinic sample application image to EKS
Note: Anything in brackets [ ] needs to be replaced with information from your deployment.
2. Click the Create a Free Cluster button.
3. Select YugabyteDB Managed Free and click the Next button.
4. Select a cloud provider for your cluster, name your cluster, select a region, and click the Next button.
5. Copy the default admin credentials by clicking the Copy Credentials link. Then check the “I have copied the admin credentials” checkbox and click the Create Cluster button.
a. You can also set your own admin credentials by selecting the “Add your own credentials” radio button and entering your own credentials.
6. Wait for your cluster to be provisioned. Click on “Add IP Allow List” in the top section of the Clusters page.
7. Click the Create New List and Add to Cluster button.
8. In the Name text box, enter “all-ips”. In the IP Address(es) or Range text box, enter “0.0.0.0/0”. Click the Save button. This will allow traffic from all IP addresses to your cluster.
a. Allowing all IPs access to your cluster is for development purposes only. When you deploy an application to Production, you will want to specify only the specific IP addresses or ranges that need access to your cluster.
Your cluster is now created in YugabyteDB Managed and is accessible to all IP addresses. Leave this page open, as you’ll be accessing the Cloud Shell later to create an application database and user.
Note: Instructions for how to connect this application to YugabyteDB are in
spring-petclinic/src/main/resources/db/yugabytedb/petclinic_db_setup_yugabytedb.md from the repo you clone below.
1. On your computer from terminal, clone the Spring Boot PetClinic sample application:
git clone https://github.com/yugabyte/spring-petclinic.git.
$ git clone https://github.com/yugabyte/spring-petclinic.git Cloning into 'spring-petclinic'... remote: Enumerating objects: 8616, done. remote: Counting objects: 100% (18/18), done. remote: Compressing objects: 100% (18/18), done. remote: Total 8616 (delta 1), reused 13 (delta 0), pack-reused 8598 Receiving objects: 100% (8616/8616), 7.29 MiB | 19.03 MiB/s, done. Resolving deltas: 100% (3268/3268), done.
3. Copy the contents of
4. In YugabyteDB Managed, click on “Connect”. Click on Connect to your Application.
5. Save your host address and port (highlighted in the screenshot below) for later. Click the Back link in the Connect to Cluster navigation breadcrumb.
6. Click the Launch Cloud Shell button.
7. Leave the default Database Name, Database User Name, and API Type. Click the Confirm button. This will launch the Cloud Shell.
8. Enter the admin password you copied previously. After you enter the password, you will have a standard YSQL shell (exactly like a PSQL shell in PostgreSQL) that you can interact with from your browser.
Password for user admin: ysqlsh (11.2-YB-22.214.171.124-b0) SSL connection (protocol: TLSv1.2, cipher: ECDHE-RSA-AES256-GCM-SHA384, bits: 256, compression: off) Type "help" for help. yugabyte=#
9. Go to https://localhost:8080. The PetClinic sample application should be available.
The PetClinic sample application is now connected to your YugabyteDB cluster and running locally.
1. Start Docker on your computer and containerize your Spring Boot PetClinic sample application:
2. Tag your image:
`docker tag [image_id] spring-petclinic` – you can find your image id by running
`docker image ls`.
3. Run your image as a container in Docker to make sure it’s working correctly:
`docker run -d --name=spring-petclinic -p 8080:8080 -e JAVA_OPTS="-Dspring.profiles.active=yugabytedb -Dspring.datasource.url=jdbc:postgresql://[host]:[port]/petclinic?load-balance=true -Dspring.datasource.initialization-mode=never" spring-petclinic`.
a. Go to https://localhost:8080. The PetClinic sample application should be available.
The PetClinic sample application is now connected to your YugabyteDB cluster and running locally on Docker.
Deploy the Spring Boot PetClinic sample application image to Amazon Elastic Kubernetes Service (EKS)
1. Go to https://aws.amazon.com/console/ and sign up for/sign in to Amazon Web Services (AWS).
2. Type “ecr” in the search bar and click on “Elastic Container Registry”.
3. Create a repository by clicking the Get Started button.
4. Enter a repository name and click the Create Repository button.
5. On your computer from terminal, login to ECR with Docker:
`aws ecr get-login-password --region [aws_region] | docker login --username AWS --password-stdin [aws_acct_id].dkr.ecr.[aws_region].amazonaws.com`.
$ aws ecr get-login-password --region us-west-1 | docker login --username AWS --password-stdin 454529406029.dkr.ecr.us-west-1.amazonaws.com WARNING! Your password will be stored unencrypted in /Users/gavinjohnson/.docker/config.json. Configure a credential helper to remove this warning. See https://docs.docker.com/engine/reference/commandline/login/#credentials-store Login Succeeded
6. Tag your PetClinic image with your ECR repo:
`docker tag spring-petclinic:latest [aws_acct_id].dkr.ecr.[aws_region].amazonaws.com/spring-petclinic:latest`.
7. Push your PetClinic image to your repo in ECR:
`docker push [aws_acct_id].dkr.ecr.[aws_region].amazonaws.com/spring-petclinic:latest`.
$ docker push 454529406029.dkr.ecr.us-west-1.amazonaws.com/spring-petclinic:latest The push refers to repository [454529406029.dkr.ecr.us-west-1.amazonaws.com/spring-petclinic] 1dc94a70dbaa: Pushed 0d29ec96785e: Pushed 888ed16fa8d4: Pushed ...
If you go to your repo in ECR, you should see the image you just pushed.
8. Type “cloudformation” in the search bar and click on “CloudFormation”.
9. Click on Create Stack -> Create stack, With new resources (standard).
10. For Prepare Template, select Template is Ready. For Template Source, select Amazon S3 URL. For Amazon S3 URL, enter “https://s3.us-west-2.amazonaws.com/amazon-eks/cloudformation/2020-10-29/amazon-eks-vpc-private-subnets.yaml”. Click the Next button.
11. Enter a Stack Name, and click the Next button.
12. On the Configure Stack Options page, click the Next button.
13. Review your VPC stack and click the Create Stack button.
14. After your stack is created, type “eks” in the search bar and click on “Elastic Kubernetes Service”.
15. Under “Create EKS cluster” enter a cluster name and click the Next Step button.
16. Select a Cluster Service Role and click the Next button.
17. Under Networking, choose the VPC you just created (that should automatically choose the appropriate subnets) and the two security groups that are available. Under Cluster Endpoint Access, choose Public and Private. Click the Next button.
18. Select nothing for Configure Logging and click the Next button.
19. Review your EKS cluster and click the Create button.
20. After your cluster has been created, go to the Configuration tab and the Compute tab under it. Click on the Add Node Group button.
21. Give your node group a Name, select an IAM Role, and click the Next button.
22. Set your compute and scaling options and click the Next button.
23. Your subnets should automatically be selected. Click the Next button.
24. Review the details of your node group and click the Create button.
25. On your computer from terminal, configure `kubectl` to connect to your EKS cluster:
`aws eks --region [aws_region] update-kubeconfig --name [cluster_name]`.
$ aws eks --region us-west-1 update-kubeconfig --name spring-petclinic Added new context arn:aws:eks:us-west-1:454529406029:cluster/spring-petclinic to /Users/gavinjohnson/.kube/config
26. Create a new file named `manifest-eks.yml`, enter the following contents, and save:
apiVersion: v1 kind: Service metadata: name: spring-petclinic labels: run: spring-petclinic spec: selector: app: spring-petclinic ports: - protocol: TCP port: 80 targetPort: 8080 type: LoadBalancer --- apiVersion: apps/v1 kind: Deployment metadata: name: spring-petclinic labels: app: spring-petclinic spec: replicas: 2 selector: matchLabels: app: spring-petclinic template: metadata: labels: app: spring-petclinic spec: containers: - name: spring-petclinic image: [aws_acct_id].dkr.ecr.[aws_region].amazonaws.com/spring-petclinic:latest ports: - containerPort: 8080 env: - name: JAVA_OPTS value: "-Dspring.profiles.active=yugabytedb -Dspring.datasource.url=jdbc:postgresql://[host]:[port]/petclinic?load-balance=true -Dspring.datasource.initialization-mode=never"
27. Create the Service and Deployment on your EKS cluster:
`kubectl create -f manifest-eks.yml`.
$ kubectl create -f manifest-eks.yml service/spring-petclinic created deployment.apps/spring-petclinic created
28. Get the URL of the load balancer for the PetClinic app:
`kubectl get svc`.
$ kubectl get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kubernetes ClusterIP 10.100.0.1 <none> 443/TCP 20h spring-petclinic LoadBalancer 10.100.30.39 ad7029ef94fed4c06a25897baf9e3c31-1572590571.us-west-1.elb.amazonaws.com 80:30736/TCP 37s
29. If you go to the External IP address listed (https://ad7029ef94fed4c06a25897baf9e3c31-1572590571.us-west-1.elb.amazonaws.com/), the PetClinic sample application should be available.
The PetClinic sample application is now connected to your YugabyteDB cluster and running on Kubernetes on EKS.
Feel free to ask them in our YugabyteDB community Slack channel.
NOTE: This post was updated on September 17, 2021.