Connect Onehouse
dbt supports connecting to Onehouse SQL using the Apache Spark Connector with the Thrift method.
note
Connect to a Onehouse SQL Cluster with the dbt-spark adapter.**
Requirements
- For dbt, ensure your Onehouse SQL endpoint is accessible via external DNS/IP, whitelisting dbt IPs.
What works
- All dbt Commands, including:
dbt clean,dbt compile,dbt debug,dbt seed, anddbt run. - dbt materializations:
tableandincremental - Apache Hudi table types of Merge on Read (MoR) and Copy on Write (CoW). It is recommended to use MoR for mutable workloads.
Limitations
- Views are not supported
dbt seedhas row / record limits.dbt seedonly supports Copy on Write tables.
dbt connection
Fill in the following fields when creating an Apache Spark warehouse connection using the Thrift connection method:
| Loading table... |
dbt project
We recommend that you set default configurations on the dbt_project.yml to ensure that the adapter executes with Onehouse compatible sql
| Loading table... |
dbt_project.yml template
+materialized: table | incremental
+file_format: hudi
+location_root: <storage_uri>
+tblproperties:
hoodie.table.type: mor | cow
A dbt_project.yml example if using jaffle shop would be
models:
jaffle_shop:
+file_format: hudi
+location_root: s3://lakehouse/demolake/dbt_ecomm/
+tblproperties:
hoodie.table.type: mor
staging:
+materialized: incremental
marts:
+materialized: table
Was this page helpful?
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
0
