Skip to content

Commit

Permalink
Merge branch 'ly-docs-ci-updates' of github.com:dbt-labs/docs.getdbt.…
Browse files Browse the repository at this point in the history
…com into ly-docs-ci-updates
  • Loading branch information
nghi-ly committed Sep 11, 2024
2 parents 6d74223 + 62eb352 commit 128c0a0
Show file tree
Hide file tree
Showing 8 changed files with 24 additions and 17 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ _Users with IT or read-only licenses will be denied access to dbt Cloud via the
App users are able to access all information that's available to the API service token.

## Procurement
The dbt Snowflake Native App is available on the [Snowflake Marketplace](https://app.snowflake.com/marketplace/listing/GZTYZSRT2R3). With the purchase of the listing, users will have access to the Native App and a dbt Cloud account that's on the Enterprise plan.
The dbt Snowflake Native App is available on the [Snowflake Marketplace](https://app.snowflake.com/marketplace/listing/GZTYZSRT2R3). Purchasing it includes access to the Native App and a dbt Cloud account that's on the Enterprise plan. Existing dbt Cloud Enterprise customers can also access it. If interested, contact your Enterprise account manager.

If you're interested, please [contact us](matilto:[email protected]) for more information.

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ pagination_prev: null
---
dbt Cloud can connect with a variety of data platform providers including:
- [AlloyDB](/docs/cloud/connect-data-platform/connect-redshift-postgresql-alloydb)
- [Amazon Athena (Beta)](/docs/cloud/connect-data-platform/connect-amazon-athena)
- [Amazon Athena](/docs/cloud/connect-data-platform/connect-amazon-athena)<Lifecycle status="Preview" />
- [Amazon Redshift](/docs/cloud/connect-data-platform/connect-redshift-postgresql-alloydb)
- [Apache Spark](/docs/cloud/connect-data-platform/connect-apache-spark)
- [Azure Synapse Analytics](/docs/cloud/connect-data-platform/connect-azure-synapse-analytics)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,13 +5,7 @@ description: "Configure the Amazon Athena data platform connection in dbt Cloud.
sidebar_label: "Connect Amazon Athena"
---

# Connect Amazon Athena <Lifecycle status="beta" />

:::note beta

This is a beta feature with limited availability. A public preview will follow shortly, for wider early access. For more information, check out our [product lifecycle](/docs/dbt-versions/product-lifecycles#dbt-cloud) page.

:::
# Connect Amazon Athena <Lifecycle status="preview" />

Your environment(s) must be on ["Versionless"](/docs/dbt-versions/versionless-cloud) to use the Amazon Athena connection.

Expand Down
6 changes: 3 additions & 3 deletions website/docs/docs/core/connect-data-platform/athena-setup.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,9 @@
title: "Athena setup"
description: "Read this guide to learn about the Athena warehouse setup in dbt."
meta:
maintained_by: Community
authors: Community
github_repo: 'dbt-athena/dbt-athena'
maintained_by: dbt Labs
authors: dbt Labs
github_repo: 'dbt-labs/dbt-athena'
pypi_package: 'dbt-athena-community'
min_core_version: 'v1.3.0'
cloud_support: Not Supported
Expand Down
4 changes: 4 additions & 0 deletions website/docs/docs/dbt-versions/release-notes.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,10 @@ Release notes are grouped by month for both multi-tenant and virtual private clo

\* The official release date for this new format of release notes is May 15th, 2024. Historical release notes for prior dates may not reflect all available features released earlier this year or their tenancy availability.

## September 2024

- **New:** The [Amazon Athena warehouse connection](/docs/cloud/connect-data-platform/connect-amazon-athena) is available as a public preview for dbt Cloud accounts that have upgraded to [`versionless`](/docs/dbt-versions/versionless-cloud).

## August 2024
- **Fix:** Fixed an issue in [dbt Explorer](/docs/collaborate/explore-projects) where navigating to a consumer project from a public node resulted in displaying a random public model rather than the original selection.
- **New**: You can now configure metrics at granularities at finer time grains, such as hour, minute, or even by the second. This is particularly useful for more detailed analysis and for datasets where high-resolution time data is required, such as minute-by-minute event tracking. Refer to [dimensions](/docs/build/dimensions) for more information about time granularity.
Expand Down
3 changes: 3 additions & 0 deletions website/docs/docs/deploy/continuous-deployment.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,11 @@ description: "Learn about continuous deployment (CD) workflows "

To help you improve data transformations and ship data products faster, you can run [merge jobs](/docs/deploy/merge-jobs) to implement a continuous deployment (CD) workflow in dbt Cloud. Merge jobs can automatically build modified models whenever a pull request (PR) merges, making sure the latest code changes are in production. You don't have to wait for the next scheduled job to run to get the latest updates.

<Lightbox src="/img/docs/dbt-cloud/using-dbt-cloud/cd-workflow.png" width="90%" title="Workflow of continuous deployment in dbt Cloud"/>

You can also implement continuous integration (CI) in dbt Cloud, which can help further to reduce the time it takes to push changes to production and improve code quality. To learn more, refer to [Continuous integration in dbt Cloud](/docs/deploy/continuous-integration).


## How merge jobs work

When you set up merge jobs, dbt Cloud listens for notifications from your [Git provider](/docs/cloud/git/git-configuration-in-dbt-cloud) indicating that a PR has been merged. When dbt Cloud receives one of these notifications, it enqueues a new run of the merge job.
Expand Down
16 changes: 11 additions & 5 deletions website/docs/guides/redshift-qs.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,6 +43,9 @@ Check out [dbt Fundamentals](https://learn.getdbt.com/courses/dbt-fundamentals)
1. Sign in to your [AWS account](https://signin.aws.amazon.com/console) as a root user or an IAM user depending on your level of access.
2. Use a CloudFormation template to quickly set up a Redshift cluster. A CloudFormation template is a configuration file that automatically spins up the necessary resources in AWS. [Start a CloudFormation stack](https://console.aws.amazon.com/cloudformation/home?region=us-east-1#/stacks/new?stackName=dbt-workshop&templateURL=https://tpch-sample-data.s3.amazonaws.com/create-dbtworkshop-infr) and you can refer to the [create-dbtworkshop-infr JSON file](https://github.com/aws-samples/aws-modernization-with-dbtlabs/blob/main/resources/cloudformation/create-dbtworkshop-infr) for more template details.

:::tip
To avoid connectivity issues with dbt Cloud, make sure to allow inbound traffic on port 5439 from [dbt Cloud's IP addresses](/docs/cloud/about-cloud/access-regions-ip-addresses) in your Redshift security groups and Network Access Control Lists (NACLs) settings.
:::

3. Click **Next** for each page until you reach the **Select acknowledgement** checkbox. Select **I acknowledge that AWS CloudFormation might create IAM resources with custom names** and click **Create Stack**. You should land on the stack page with a CREATE_IN_PROGRESS status.

Expand Down Expand Up @@ -165,6 +168,7 @@ Now we are going to load our sample data into the S3 bucket that our Cloudformat
select * from jaffle_shop.orders;
select * from stripe.payment;
```
## Connect dbt Cloud to Redshift
1. Create a new project in [dbt Cloud](/docs/cloud/about-cloud/access-regions-ip-addresses). From **Account settings** (using the gear menu in the top right corner), click **+ New Project**.
2. Enter a project name and click **Continue**.
Expand All @@ -173,17 +177,19 @@ Now we are going to load our sample data into the S3 bucket that our Cloudformat
- **Hostname** &mdash; Your entire hostname.
- **Port** &mdash; `5439`
- **Database** &mdash; `dbtworkshop`.
<div style={{maxWidth: '400px'}}>
<Lightbox src="/img/redshift_tutorial/images/dbt_cloud_redshift_account_settings.png" title="dbt Cloud - Redshift Cluster Settings" />
</div>
<Lightbox src="/img/redshift_tutorial/images/dbt_cloud_redshift_account_settings.png" width="90%" title="dbt Cloud - Redshift Cluster Settings" />
:::tip
To avoid connectivity issues with dbt Cloud, make sure to allow inbound traffic on port 5439 from [dbt Cloud's IP addresses](/docs/cloud/about-cloud/access-regions-ip-addresses) in your Redshift security groups and Network Access Control Lists (NACLs) settings.
:::

5. Set your development credentials. These credentials will be used by dbt Cloud to connect to Redshift. Those credentials (as provided in your CloudFormation output) will be:
- **Username** &mdash; `dbtadmin`
- **Password** &mdash; This is the autogenerated password that you used earlier in the guide
- **Schema** &mdash; dbt Cloud automatically generates a schema name for you. By convention, this is `dbt_<first-initial><last-name>`. This is the schema connected directly to your development environment, and it's where your models will be built when running dbt within the Cloud IDE.
<div style={{maxWidth: '400px'}}>
<Lightbox src="/img/redshift_tutorial/images/dbt_cloud_redshift_development_credentials.png" title="dbt Cloud - Redshift Development Credentials" />
</div>
6. Click **Test Connection**. This verifies that dbt Cloud can access your Redshift cluster.
7. Click **Next** if the test succeeded. If it failed, you might need to check your Redshift settings and credentials.
Expand Down
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

0 comments on commit 128c0a0

Please sign in to comment.