Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/pingidentity/pipeline-example-application
https://github.com/pingidentity/pipeline-example-application
cicd pingone terraform
Last synced: about 1 month ago
JSON representation
- Host: GitHub
- URL: https://github.com/pingidentity/pipeline-example-application
- Owner: pingidentity
- Created: 2024-03-25T17:17:47.000Z (9 months ago)
- Default Branch: prod
- Last Pushed: 2024-10-25T20:09:25.000Z (about 2 months ago)
- Last Synced: 2024-10-26T07:31:13.307Z (about 2 months ago)
- Topics: cicd, pingone, terraform
- Language: HCL
- Homepage:
- Size: 985 KB
- Stars: 0
- Watchers: 6
- Forks: 0
- Open Issues: 3
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Ping Application Example Pipeline
This repository is intended to present a simplified reference demonstrating how a management and deployment pipeline might work for applications that depend on services managed by a central IAM platform team. As such, it is a complement to the [infrastructure](https://github.com/pingidentity/pipeline-example-infrastructure) and [platform](https://github.com/pingidentity/pipeline-example-platform) example pipeline repositories.
> [!IMPORTANT]
> This repository directly depends on a completed setup of the [pipeline-example-platform](https://github.com/pingidentity/pipeline-example-platform?tab=readme-ov-file#deploy-prod-and-qa). Please ensure you have completed the steps for configuration leading up to and including the previous link, where a `prod` and `qa` environment have been deployed.**Infrastructure** - Components dealing with deploying software onto self-managed Kubernetes infrastructure and any configuration that must be delivered directly via the filesystem.
**Platform** - Components dealing with deploying configuration to self-managed or hosted services that will be shared and leveraged by upstream applications.
**Application** - Delivery and configuration of a client application that relies on core services from Platform and Infrastructure.
The use cases and features shown in this repository are an implementation of the guidance provided from Ping Identity's [Terraform Best Practices](https://terraform.pingidentity.com/best-practices/) and [Getting Started with Configuration Promotion at Ping](https://terraform.pingidentity.com/getting-started/configuration-promotion/) documents.
In this repository, the processes and features shown in a GitOps process of developing and delivering a new application include:
- Feature Request Template
- Feature Development in an on-demand or persistent development environment
- Extracting feature configuration to be stored as code
- Validating the extracted configuration from the developer perspective
- Validating that the suggested configuration adheres to contribution guidelines
- Review process of suggested change
- Approval of change and automatic deployment into higher environments## Prerequisites
To be successful in recreating the use cases supported by this pipeline, there are initial steps that should be completed prior to configuring this repository:
- Completion of all pre-requisites and configuration steps leading to [Feature Development](https://github.com/pingidentity/pipeline-example-platform?tab=readme-ov-file#feature-development) from the example-pipeline-platform repository
- [Docker](https://docs.docker.com/engine/install/) - used to deploy the UI for a sample interface
- [tflint](https://github.com/terraform-linters/tflint) - for Terraform linting
- [dvlint](https://github.com/pingidentity/dvlint) - for Davinci flow linting
- [trivy](https://github.com/aquasecurity/trivy) - for security scanning> [!TIP]
> The last three tools are used by the pipeline in Github, and the pipeline will fail if these tests and configuration checks do not pass. Installing these tools locally and running `make devcheck` before committing changes should ensure that the pipeline will pass when changes are pushed.> [!IMPORTANT]
> For PingOne, meeting these requirements means you should have credentials for a worker app residing in the "Administrators" environment that has organization-level scoped roles. For DaVinci, you should have credentials for a user in a non-"Administrators" environment that is part of a group specifically intended to be used by command-line tools or APIs with environment-level scoped roles.### Development Environment
If you have not created a static development environment in your PingOne account, you can do so in the local copy of the *pipeline-example-platform* repository by running the following commands to instantiate one matching prod and qa:
```bash
git checkout prod
git pull origin prod
git checkout -b dev
git push origin dev
```Capture the environment ID for the development environment for use later.
> [!NOTE]
> The platform team may support ephemeral development environments rather than the static environment mentioned here. For purposes of this guide, we will assume that the development environment is shared and will be used by multiple developers.![PingOne Environments](./img/pingOneEnvs.png "PingOne Environments")
### Repository Setup
Click the **Use this template** button at the top right of this page to create your own repository. After the repository is created, clone it to your local machine to continue. The rest of this guide will assume you are working from the root of the cloned repository.
> [!NOTE]
> A pipeline will run and fail when the repository is created. This result is expected as the pipeline is attempting to deploy the application and the necessary configuration has not yet been completed.Create a `qa` branch from the `prod` branch in the repository. This branch will be used to test the changes before they are promoted to the `prod` branch. Changes to the `qa` branch in this repository are deployed to the `qa` environment in PingOne. As with the `prod` branch, the pipeline will fail due to missing configuration.
```bash
git checkout prod
git pull origin prod
git checkout -b qa
git push origin qa
```## Development Lifecycle Diagram
The use cases in this repository follow a flow similar to this diagram:
![SDLC flow](./img/generic-pipeline.png "Development Flow")
## Before You Start
There are a few items to configure before you can successfully use this repository.
### PingOne Environments
> [!IMPORTANT]
> The configurations in this sample repository rely on environments created from [pipeline-example-platform](https://github.com/pingidentity/pipeline-example-platform). For the `PINGONE_TARGET_ENVIRONMENT_ID_PROD` and `PINGONE_TARGET_ENVIRONMENT_ID_QA` variables needed down below, get the Environment ID for the `prod` and `qa` environments. The Environment ID can be found from the output at the end of a terraform apply (whether from the Github Actions pipeline, or local) or directly from the PingOne console.### Github CLI
The Github cli: `gh` will need to be configured for your repository. Run the command **gh auth login** and follow the prompts. You will need an access token for your Github account and the repository created from this template:
```bash
gh auth login? What account do you want to log into? GitHub.com
? You're already logged into github.com. Do you want to re-authenticate? Yes
? What is your preferred protocol for Git operations? HTTPS
? Authenticate Git with your GitHub credentials? Yes
? How would you like to authenticate GitHub CLI? Paste an authentication token
Tip: you can generate a Personal Access Token here https://github.com/settings/tokens
The minimum required scopes are 'repo', 'read:org', 'workflow'.
? Paste your authentication token: ****************************************
- gh config set -h github.com git_protocol https
✓ Configured git protocol
✓ Logged in as
```### Github Actions Secrets
The Github pipeline actions depend on sourcing secrets as ephemeral environment variables. To prepare the secrets in the repository:
```bash
cp secretstemplate localsecrets
```> [!CAUTION]
> `secretstemplate` is a template file while `localsecrets` contains credentials. `localsecrets` is part of *.gitignore* and should never be committed into the repository. **`secretstemplate`** is committed to the repository, so ensure that you do not edit it directly or you risk exposing your secrets.Fill in `localsecrets` accordingly, referring to the comments in the file for guidance. Many of the values needed for this file can be found in the corresponding localsecrets file from the platform repository.
After updating the file, run the following commands to upload **localsecrets** to Github:
```bash
_secrets="$(/usr/bin/base64 -i localsecrets)"
gh secret set --app actions TERRAFORM_ENV_BASE64 --body $_secrets
unset _secrets
```## Development Example Overview
To experience the developer's perspective, a walkthrough follows. The demonstration will simulate the use case of modifying a Davinci flow and promoting the change. To simplify the demonstration, a starting pre-configured flow will be created using Terraform. The UI components will be built into a Docker image and launched on your local machine. After you have deployed the flow, you will be able to make the changes necessary in the PingOne UI, export the configuration, and promote the change to the QA and Prod environments.
## Feature Development
Now that the repository and pipeline are configured, the typical git flow can be followed. You will follow steps similar to those documented in the [pipeline-example-platform "Feature Development"](https://github.com/pingidentity/pipeline-example-platform/tree/prod?tab=readme-ov-file#feature-development) section.
A notable difference between this repository and the platform example is that the application pipeline does NOT deploy "development" feature configurations. Unlike QA and Prod, feature configuration *deployment* only takes place from the local machine and the development environment ID is not stored in the repository. In a following step when the `./scripts/local_feature_deploy.sh` script runs, you will be prompted for the environment ID.
This deployment format accounts for static and ephemeral development environment setups, and further helps avoid the possibility that a developer changes a shared variable that impacts another engineer's work. For reliability, the developer must provide the environment ID for development and initial testing from the local machine, but the pipeline will handle deploying changes to QA and Prod, as those are common across teams and can be defined universally.
### Feature Development Walkthrough
1. To align with a typical developer experience, use the **Feature request** template in Github to create a GitHub Issue in the UI. GitHub Issue templates help ensure the requestor provides appropriate information on the issue. Provide a description and save the issue by clicking the "Submit New Issue" button.
![Create a new issue](./img/githubissuerequest.png "Create a new issue")
2. Select the issue and click "Create a branch" and choose "Checkout Locally" from the right-hand navigation menu.
![Create a branch](./img/createabranch.png "Create a branch")
3. When you create the branch, you are provided the commands to check out the branch locally. Run the commands as instructed.
![Check out the new branch](./img/checkoutbranch.png "Check out the new branch")
4. Deploy the sample Davinci flow to the development environment by running the following commands. You will be prompted for the development environment ID:
```bash
source localsecrets
./scripts/local_feature_deploy.sh
```> [!NOTE]
> If you want to see what Terraform will do without actually deploying, provide the `--dry-run` flag to the script. This flag generates the Terraform configuration without applying it by running `terraform plan`.> [!NOTE]
> The `local_feature_deploy.sh` script will not run against the `prod` or `qa` branches.5. Confirm the deployment by examining the Davinci flow in the PingOne console in the development environment matching the ID you provided. Click on the Davinci link from the PingOne console to open the DaVinci console and select **Flows** from the left navigation panel. Click on the **PingOne DaVinci Registration Example** flow to view the configuration.
6. The Terraform configuration also deployed a sample client application in a local docker container that can be used to try out the flow by navigating to [https://127.0.0.1:8443](https://127.0.0.1:8443). You will be presented a simple progressive profiling style form to enter an email address. If the email address is not found, you will be prompted to register the user.
> [!NOTE]
> For demo purposes, there is a self-signed certificate in the Docker image that will require you to accept the security warning in your browser to proceed.7. On the next panel, you are told to provide the email and password. There are password rules in place, but you are not informed when prompted. Try using a simple password such as `password`. The form does not indicate there is a problem, but refuses to accept the password and continue. The password must be at least 8 characters long and contain at least one uppercase letter, one lowercase letter, one number, and one special character.
8. Create a valid password. After registering the user, you will be redirected to login.
9. To improve the flow, you will add a small prompt on the registration page to indicate that the password must meet the requirements. To do so, select the **Registration Window** node in the Davinci flow editor. Replace the text in the HTML Template editor with the following code block. The only change from what is provided is the addition of the password requirements notification and some descriptive comments.
```html
We did not locate that account. Sign up now!
Password must be at least 8 characters and contain 1 uppercase, 1 lowercase, 1 digit, and 1 symbol
Register```
10. Click **Apply** to save the changes, then click **Deploy** to update the flow in the development environment.
11. Next, test your change from the client application. Browse to [https://127.0.0.1:8443](https://127.0.0.1:8443) again and provide a new email address. Notice on the registration page that you are presented the password requirements message. Completing the new user registration is optional, as you can already see the change has been applied in the interface.
12. To capture the changes for inclusion in your code, export the flow. You can do so by selecting the three dots at the top right of the DaVinci flow editor UI and clicking **Download Flow JSON**. Ensure to select **Include Variable Values** when you export.
![Export Menu](./img/exportMenu.png "Export Menu")
13. For the sake of brevity, assume that testing has been done, and you are ready to proceed. After the application is "tested", the new configuration must be added to the Terraform configuration. This addition will happen in a few steps:
a. Copy the contents of the downloaded JSON file and use them to replace the `terraform/davinci-flows/davinci-widget-reg-authn-flow.json` file contents. If you examine the changes, you will see that it involves the company ID, metadata about the file and the changes you made to the node in the flow.
b. Run the deploy script again:
```bash
source localsecrets
./scripts/local_feature_deploy.sh
```c. If you examine the output, you will see that the change to the JSON object caused Terraform to detect a delta between the state stored in S3 (with the old JSON code) and what is present in the new code:
```bash
Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the
following symbols:
~ update in-placeTerraform will perform the following actions:
# davinci_flow.registration_flow will be updated in-place
~ resource "davinci_flow" "registration_flow" {
~ flow_configuration_json = (sensitive value)
~ flow_export_json = (sensitive value)
~ flow_json = (sensitive value)
id = "a6d551f1d7aa2612f2bf6c371b0026e1"
name = "PingOne DaVinci Registration Example"
# (4 unchanged attributes hidden)# (3 unchanged blocks hidden)
}Plan: 0 to add, 1 to change, 0 to destroy.
```d. Applying the change will update and 'redeploy' the flow (even though it is exactly the same as what is there, the update to the file creates a delta for Terraform to resolve) and afterward, the state will reflect the new configuration. Type `yes` to apply the changes.
e. Run the script again to confirm that the state matches the configuration:
```bash
No changes. Your infrastructure matches the configuration.
Terraform has compared your real infrastructure against your configuration and found no differences, so no changes are
needed.Apply complete! Resources: 0 added, 0 changed, 0 destroyed.
```14. If you want to go one step further, you can modify the HTML code that is placed in the nginx image. To do so, you can modify `./terraform/sample-app/index.html` in some manner. When the script is ran again, it will detect the change to the file, rebuild the Docker image, and launch a replacement container for the UI.
15. Before committing and pushing the changes, run a devcheck against the code to ensure the formatting and syntax are correct, ignoring any warnings or informational messages:
```bash
make devcheck
```16. Commit and push the changes to the repository:
```bash
git add .
git commit -m "Adding password requirements to registration page"
git push
```17. The push will fire a pipeline that runs the same checks as you did locally. However, since this is a development branch, Terraform deployment will be skipped.
18. Create a pull request in the repository from your branch to `qa`. Creation of the pull request will trigger a pipeline of checks and allow a "reviewer" to validate the changes. Merge the pull request to trigger the deployment workflow against the **qa** environment in your PingOne account. You may choose to confirm the flow exists in your **qa** environment and has your change.
19. Finally, you can create a pull request from `qa` to `prod`. Follow the same review process as for qa. Upon merge of the pull request, the pipeline will validate the changes and deploy the flow to the **prod** environment in your PingOne account.
## Cleanup
When you are finished with the demonstration, you can clean up the resources from the PingOne `dev` environment by running the following commands. The `--destroy` flag will destroy the resources in the development environment:
```bash
source localsecrets
./scripts/local_feature_deploy.sh --destroy
```After the development resources are destroyed:
- delete the branch from the local and remote repositories
- resolve the issue in the Github UI
- delete the feature folder from the `application-state/dev/` in the S3 bucket## Conclusion
This repository demonstrates a simplified example of how a pipeline can be used to manage the development and deployment of an application that relies on services managed by a central IAM platform team. The pipeline is designed to be flexible and extensible, allowing for the addition of new features and services as needed. By following the steps outlined in this repository, you can gain a better understanding of how to implement a similar pipeline in your own environment.