Skip to content

riprasad/apicurio-operate-first

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Table Of Contents

Operate First

Operate First is a community of developers, operators, and others — site reliability engineers (SRE), data scientists, Open Source practitioners, and so forth — who believe in promoting Open Source approach to operations.

Open Source software is widely available, but it faces an operations-barrier when bringing it to a production environment.

Proprietary services have attempted to address this barrier, but they undermine the Open Source development model because lessons learned from operating the code are invisible to the Open Source developers. To overcome this operations barrier in an Open Source way, we must switch to an Open Source-like approach to operations, or Open Operations (Open Ops).

Open Ops means developers and operators collaborating Openly to apply a product's operational considerations right back into the code itself. The result is an Open Source service.

For More information on operate-first, please visit https://www.operate-first.cloud/

Useful links

Infrastructure

Application

Cluster and Namespace

The projects are presently deployed on the 3scale OpenShift Cluster and overseen by the 3Scale platform's SRE team. If you encounter any challenges or have inquiries regarding the infrastructure, there's #list-3scale-sre-mas Slack channel available for you to get in touch with them.

We currently have two namespaces allocated for our use, namely rhaf-apicurio-registry and rhaf-apicurio-designer, each with the resource quotas listed below.

requests.cpu: '4'
requests.memory: 10Gi
services.loadbalancers: '0'

Please note that this is a development cluster equipped with only three OpenShift Container Platform (OCP) worker nodes. Each node has 4 CPUs and 15GB of memory available, serving the needs of OCP, 3Scale and Apicurio components. If additional capacity is required, we can consider a MODEST INCREASE in the quota only After optimizing resource requests to match actual usage. However, any adjustments will be made conservatively.

For more details on the cluster and available services, click here.

This Repository

This repository holds the manifests files that are used for deploying the apicurio projects on the operate-first cluster.

The project uses kustomize and Argo CD to enable GitOps and continuous deployments. Hence, any changes made to the manifest files will automatically be applied to the project deployments.

Manifests

The kubernetes/OpenShift manifest files for each project are organized under the manifests folder. The project uses kustomize and in particular the Overlay functionality, that allows to define basic configuration and a collection of patches(overlays) we can apply on the base definitions.

The deployments are distributed over the two namespaces within the cluster and are managed through the Kustomization.yaml file linked below.

Namespace Deployment File
rhaf-apicurio-registry kustomization.yaml
rhaf-apicurio-designer kustomization.yaml

Local Verification

To verify what is the output of the applied overlay you can run the below commands. The output is the concatenation of all the YAML documents for all of the resources we specified along with the applied overlay, patches, common labels etc.

oc kustomize manifests/registry/deployment/3scale/ | tee effective-manifest.yaml

oc kustomize manifests/api-designer/overlays/3scale/ | tee effective-manifest.yaml

If you’re using kustomize as a standalone CLI, you’ll instead run:

kustomize build manifests/registry/deployment/3scale/ | tee effective-manifest.yaml

kustomize build manifests/api-designer/overlays/3scale/ | tee effective-manifest.yaml

Continuous Deployment

This project leverages Argo CD for seamless and automated continuous deployment. Any modifications made to the manifest files trigger an automatic deployment through these Argo CD applications.

The application is set to synchronize every 5 minutes for regular updates. However, you also have the option to manually initiate a synchronization by clicking on the refresh button. This ensures that changes are promptly reflected in the deployed environment.

TODO: As of now, our permissions are limited to read-only access for the Argo applications. Any attempts to execute actions are met with denial, stemming from permission constraints. To address this, we could have the assignment of the following admin role to our AppProject.

roles:
    - name: project-admin
      description: Read/Write access to this project only
      policies:
        - p, proj:apicurio-registry:project-admin, applications, get, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, create, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, update, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, delete, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, sync, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, override, apicurio-registry/*, allow
        - p, proj:apicurio-registry:project-admin, applications, action/*, apicurio-registry/*, allow
      groups:
        - rhaf-dev

Note: The ArgoCD deployments are configured and managed by the SRE team in the 3scale platform repository here

RDS Database

The Apicurio Operate First setup leverages a private Amazon RDS (Relational Database Service) instance as its primary data storage solution.

RDS Specifics

  • DB Type: PostgreSQL
  • DB engine version: 15.4
  • Instance Type: db.t4g.small (2 vCPU and 2GB memory)
  • Instance Storage: 30 GB

Being a development AWS RDS instance, features typically needed for production workloads like MultiAZ, detailing monitoring, long backup retention, etc. are not enabled.

Connection Details

Connections string (master user, master password, RDS host...) can be found at Vault path: https://vault.apps.dev-eng-ocp4-mas.dev.3sca.net/ui/vault/secrets/secret/show/kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-master-credentials

How you should use the master credentials?

The master username and password must not be used should be used exclusively for the creation of individual user accounts, associated databases, and the granting of relevant privileges. It is imperative not to utilize them directly within your application.

This approach ensures that each application is associated with a dedicated user account, limited to accessing a specific database. Consequently, in the unfortunate event of an application compromise, any potential attacker would only gain access to the specific database associated with that particular application.

How to create new user accounts and DB?

To connect to the RDS database, a PostgreSQL client is required. However, You won't be able to connect to the RDS instance directly from your laptop, because the RDS is confined to the cluster network's access due to security measures in place.

As a solution, a PostgreSQL client (pg-client) has been provisioned within the cluster. You can utilize it to establish a connection with the database by following the steps below:

  1. Login to the openshift cluster.

  2. Exec into the pg-client pod running in the cluster.

    oc exec -it pg-client -n rhaf-apicurio-designer -- /bin/bash
    
  3. Connect to the RDS

    PGPASSWORD='<password>' psql -h <hostname> -U <user> -d postgres
    
  4. Create the new user account and DB. [Usually, the database name and the username are kept the same.. The example below illustrates the process of generating a new user and database for the apicurio application. Please make sure to substitute the database name and username according to your specific requirements.]

    CREATE USER apicurio WITH PASSWORD '<password>';
    CREATE DATABASE apicurio;
    GRANT ALL PRIVILEGES ON DATABASE apicurio TO apicurio;
    

    If your application encounters difficulties executing certain database instructions, you have the option to designate the user as the owner of the database using the following command:

    ALTER DATABASE apicurio OWNER TO apicurio;
    

Note: To confirm the successful creation of the user and database, exit the session and attempt to log in using the newly created user and database credentials.

PGPASSWORD='<password>' psql -h <hostname> -U apicurio -d apicurio

Manage Credentials

We employ HashiCorp Vault as a trusted solution for securely managing and storing sensitive information, such as secrets, credentials, and other confidential data. Vault ensures the highest level of security by providing a centralized platform for secret management, encryption, and access control. By leveraging Vault's robust features and integrations, we can securely store, retrieve, and distribute secrets to our applications and services while maintaining strict access controls and audit trails. This approach not only enhances the security posture of our deployments but also streamlines the management of secrets, making it an essential component of our infrastructure.

Vault Access

  1. Obtain a GitHub personal authentication token by following these steps:
    • Within your GitHub account, access your "Personal settings" by clicking on your profile picture in the upper right corner of the screen and then selecting "settings".

    • Proceed to "Developer settings", followed by "Personal access tokens", and click on "Generate new token". Make sure to grant at least the read:org permission.

  2. Visit the Vault Application
  3. Authenticate using your GitHub Token.

Adding Secrets

  1. Navigate to secrets > secret > kubernetes > dev-eng-mas > rhaf-apicurio-registry
  2. From the WebApplication Secrets page, click Add new secret.
  3. Enter secretname in the `Name`` field and secret-value in the Value field, and click Save.

Creating External Secrets

External Secrets is a Kubernetes extension that enables the integration of external secret management systems, such as HashiCorp Vault, with your Kubernetes clusters. By leveraging External Secrets, you can centralize the management of secrets in established external systems while benefiting from Kubernetes' native secret controller capabilities.

After successfully adding the credentials within the vault, you can proceed to define the ExternalSecret Custom Resource. Below is an illustrative YAML example pertaining to the secret named dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials, which has been incorporated into the vault.

apiVersion: external-secrets.io/v1beta1
kind: ExternalSecret
metadata:
  name: rds-apicurio-credentials
spec:
  target:
    name: rds-apicurio-credentials
  secretStoreRef:
    kind: ClusterSecretStore
    name: vault-mas  
  refreshInterval: 1m0s
  data:
    - secretKey: database
      remoteRef:
        key: kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials
        property: database
    - secretKey: host
      remoteRef:
        key: kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials
        property: host
    - secretKey: port
      remoteRef:
        key: kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials
        property: port
    - secretKey: user
      remoteRef:
        key: kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials
        property: user
    - secretKey: password
      remoteRef:
        key: kubernetes/dev-eng-mas/rhaf-apicurio-registry/dev-eng-mas-apicurio-rds-pgsql-apicurio-credentials
        property: password

About

Operate First repository for Apicurio projects.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published