Skip to content

Latest commit

 

History

History
171 lines (118 loc) · 10.2 KB

CONTRIBUTING.md

File metadata and controls

171 lines (118 loc) · 10.2 KB

Contributing

Setting up the development environment

  1. Install Golang environment (check instructions on the official page https://go.dev/doc/install depending on you OS).
  2. Fork this repo and clone it.
  3. Run make dev-setup in the main directory of the cloned repository.
  4. You can clean up the dev setup by running make dev-cleanup.

Repository structure

The notable technical files/directories inside the repository:

  • Makefile - contains instructions to set up the developer's environment, run tests, etc.
  • pkg/provider - definition of the provider
  • pkg/resources, pkg/datasources - definitions and tests (consult section Running the tests locally below) for resources and datasources
  • pkg/acceptance - helpers for acceptance and integration tests
  • pkg/sdk - definitions of the SDK objects (SDK is our client to Snowflake, using gosnowflake driver underneath)
  • pkg/sdk/testint - integration tests for the SDK (consult section Running the tests locally below)

⚠️ Important ⚠️ We are in progress of cleaning up the repository structure, so beware of the changes in the packages/directories.

Running the tests locally

Currently, we have three main types of tests:

  • SDK unit tests (in directory pkg/sdk, files ending with _test.go)
  • SDK integration tests (in directory pkg/sdk/testint, files ending with _integration_test.go)
  • resource/datasource acceptance tests (in directories pkg/resources and pkg/datasources, files ending with _acceptance_test.go)

Both integration and acceptance tests require the connection to Snowflake (some of the tests require multiple accounts).

The preferred way of running particular tests locally is to create a config file ~/.snowflake/config, with the following content.

[default]
account = "<your account in form of organisation-account_name>"
user = "<your user>"
password = "<your password>"
role = "<your role>"
host="<host of your account, e.g. organisation-account_name.snowflakecomputing.com>"

To be able to run all the tests you additionally need two more profiles [secondary_test_account] and [incorrect_test_profile]:

[secondary_test_account]
account = "<your secondary account in form of organisation-account_name2>"
user = "<your user on the secondary account>"
password = "<your password on the secondary account>"
role = "<your role on the secondary account>"
host="<host of your account, e.g. organisation-account_name2.snowflakecomputing.com>"

[incorrect_test_profile]
account = "<your account in form of organisation-account_name>"
user = "<non-existing user>"
password = "<bad password>"
role = "<any role, e.g. ACCOUNTADMIN>"
host="<host of your account, e.g. organisation-account_name.snowflakecomputing.com>"

We are aware that not everyone has access two multiple accounts, so the majority of tests can be run using just one account. The tests setup however, requires both profiles (default and secondary_test_account) to be present. You can use the same details for secondary_test_account as in the default one, if you don't plan to run tests requiring multiple accounts.

⚠️ Important ⚠️ Some of the tests require the privileged role (like ACCOUNTADMIN). Otherwise, the managed objects may not be created. If you want to use lower role, you have to make sure it has all the necessary privileges added.

To run the tests we have three different commands:

  • make test run unit and integration tests
  • make test-acceptance run acceptance tests
  • make test-integration run integration tests

You can run the particular tests form inside your chosen IDE but remember that you have to set TF_ACC=1 environment variable to run any acceptance tests (the above commands set it for you). It is also worth adding the TF_LOG=DEBUG environment variable too, because the output of the execution is much more verbose.

Making a contribution

Discuss a change with us!

It's important to establish the scope of the change before the actual implementation. We want to avoid the situations in which the PR is rejected because it contradicts some other change we are introducing.

Remember to consult our roadmap, maybe we are already working on the issue!

It's best to approach us through the GitHub issues: either by commenting the already existing one or by creating a new one.

Follow the code conventions inside the repository

We believe that code following the same conventions is easier to maintain and extend. When working on the given part of the provider try to follow the local solutions and not introduce too much new ideas.

Introducing a new part of the SDK

To create new objects in our SDK we use quickly created generator that outputs the majority of the files needed. These files should be later edited and filled with the missing parts. We plan to improve the generator later on, but it should be enough for now. Please read more in the generator readme.

Test the change

Every introduced change should be tested. Depending on the type of the change it may require (any or mix of):

  • adding/modifying existing unit tests (e.g. changing the behavior of validation in the SDK)
  • adding/modifying existing integration tests (e.g. adding missing SDK invocations)
  • adding/modifying existing acceptance tests (e.g. fixing the parameter on the resource level)

It's best to discuss with us what checks we expect prior to making the change.

Describe the breaking changes

If the change requires manual actions when bumping the provider version, they should be added to the migration guide.

Before submitting the PR

The documentation for the provider is generated automatically. We follow the few formatting conventions that are automatically checked with every PR. They can fail and delay the resolution of your PR. To make it much less possible, run make pre-push before pushing your changes to GH. It will reformat your code (or suggest reformatting), generate all the missing docs, clean the dependencies, etc.

Naming and describing the PR

We use Conventional Commits for commit message formatting and PR titles. Please try to adhere to the standard.

Refer to the regular expression for PR title validation.

Implemented changes should be described thoroughly (we will prepare PR template for the known use cases soon):

  • reference the issue that is addressed with the given change
  • summary of changes
  • summary of added tests
  • (optional) what parts will be covered in the subsequent PRs

Requesting the review

We check for the new PRs in our repository every day Monday-Friday. We usually need 1-2 days to leave the review comments. However, there are times when you can expect even more than a week response time. In such cases, please be patient, and ping us after a week if we do not post a reason for the delay ourselves. It's possible that we just missed it.

During our review we try to point out the unhandled special cases, missing tests, and deviations from the established conventions. Remember, review comment is like an invitation to dance: you don't have to agree but please provide the substantive reasons.

⚠️ Important ⚠️ Tests and checks are not run automatically after your PR. We run them manually, when we are happy with the state of the change (even if some corrections are still necessary).

Advanced Debugging

If you want to build and test the provider locally (manually, not through acceptance tests), build the binary first using make build-local or install to the proper local directory by invoking make install-tf (to uninstall run make uninstall-tf).

Next, edit your ~/.terraformrc file to include the following:

provider_installation {

  dev_overrides {
      "registry.terraform.io/Snowflake-Labs/snowflake" = "<path_to_binary>"
  }

  direct {}
}

To debug the provider with a debugger:

  1. Launch the provider with the -debug command line argument in your debugger session. Once the provider starts, it will print instructions on setting the TF_REATTACH_PROVIDERS environment variable.

    Provider started. To attach Terraform CLI, set the TF_REATTACH_PROVIDERS environment variable with the following:
    
    Command Prompt:	set "TF_REATTACH_PROVIDERS={"registry.terraform.io/Snowflake-Labs/snowflake":{"Protocol":"grpc","ProtocolVersion":5,"Pid":35140,"Test":true,"Addr": {"Network":"tcp","String":"127.0.0.1:54706"}}}"
    PowerShell:	$env:TF_REATTACH_PROVIDERS='{"registry.terraform.io/Snowflake-Labs/snowflake":{"Protocol":"grpc","ProtocolVersion":5,"Pid":35140,"Test":true,"Addr":{"Network":"tcp","String":"127.0.0.1:54706"}}}'
  2. Open a terminal where you will execute Terraform and set the TF_REATTACH_PROVIDERS environment variable using the command from the first step.

  3. Run Terraform as usual from this terminal. Any breakpoints you set will halt execution, and you can troubleshoot the provider from your debugger.

Note: The TF_REATTACH_PROVIDERS environment variable needs to be set every time you restart your debugger session as some values like the Pid or the TCP port will change with every execution.

For further instructions, please check the official Terraform Plugin Development guide.