init
This commit is contained in:
		
							parent
							
								
									236666e4d7
								
							
						
					
					
						commit
						022cf0139a
					
				| 
						 | 
					@ -0,0 +1,10 @@
 | 
				
			||||||
 | 
					#!bin/bash
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					push:
 | 
				
			||||||
 | 
						git add . && git commit -a -m "update" && git push
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					run:
 | 
				
			||||||
 | 
						dbt deps --profiles-dir=. --project-dir=.
 | 
				
			||||||
 | 
					    #dbt run --profiles-dir=. --project-dir=. --full-refresh
 | 
				
			||||||
 | 
						dbt run --profiles-dir=. --project-dir=. --full-refresh --select se_orders
 | 
				
			||||||
 | 
					
 | 
				
			||||||
							
								
								
									
										20
									
								
								README.md
								
								
								
								
							
							
						
						
									
										20
									
								
								README.md
								
								
								
								
							| 
						 | 
					@ -1 +1,19 @@
 | 
				
			||||||
# dbt-selly-express
 | 
					## Installing dbt
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					1. Activate your venv and run `pip3 install dbt`
 | 
				
			||||||
 | 
					1. Copy `airbyte-normalization/sample_files/profiles.yml` over to `~/.dbt/profiles.yml`
 | 
				
			||||||
 | 
					1. Edit to configure your profiles accordingly
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Running dbt
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					1. `cd airbyte-normalization`
 | 
				
			||||||
 | 
					1. You can now run dbt commands, to check the setup is fine: `dbt debug`
 | 
				
			||||||
 | 
					1. To build the dbt tables in your warehouse: `dbt run`
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Running dbt from Airbyte generated config
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					1. You can also change directory (`cd /tmp/dev_root/workspace/1/0/normalize` for example) to one of the workspace generated by Airbyte within one of the `normalize` folder.
 | 
				
			||||||
 | 
					1. You should find `profiles.yml` and a bunch of other dbt files/folders created there.
 | 
				
			||||||
 | 
					1. To check everything is setup properly: `dbt debug --profiles-dir=$(pwd) --project-dir=$(pwd)`
 | 
				
			||||||
 | 
					1. You can modify the `.sql` files and run `dbt run --profiles-dir=$(pwd) --project-dir=$(pwd)` too
 | 
				
			||||||
 | 
					1. You can inspect compiled dbt `.sql` files before they are run in the destination engine in `normalize/build/compiled` or `normalize/build/run` folders
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,63 @@
 | 
				
			||||||
 | 
					# This file is necessary to install dbt-utils with dbt deps
 | 
				
			||||||
 | 
					# the content will be overwritten by the transform function
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# Name your package! Package names should contain only lowercase characters
 | 
				
			||||||
 | 
					# and underscores. A good package name should reflect your organization's
 | 
				
			||||||
 | 
					# name or the intended use of these models
 | 
				
			||||||
 | 
					name: 'airbyte_utils'
 | 
				
			||||||
 | 
					version: '1.0'
 | 
				
			||||||
 | 
					config-version: 2
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# This setting configures which "profile" dbt uses for this project. Profiles contain
 | 
				
			||||||
 | 
					# database connection information, and should be configured in the  ~/.dbt/profiles.yml file
 | 
				
			||||||
 | 
					profile: 'normalize'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# These configurations specify where dbt should look for different types of files.
 | 
				
			||||||
 | 
					# The `source-paths` config, for example, states that source models can be found
 | 
				
			||||||
 | 
					# in the "models/" directory. You probably won't need to change these!
 | 
				
			||||||
 | 
					source-paths: ["models"]
 | 
				
			||||||
 | 
					docs-paths: ["docs"]
 | 
				
			||||||
 | 
					analysis-paths: ["analysis"]
 | 
				
			||||||
 | 
					test-paths: ["tests"]
 | 
				
			||||||
 | 
					data-paths: ["data"]
 | 
				
			||||||
 | 
					macro-paths: ["macros"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					target-path: "../build"  # directory which will store compiled SQL files
 | 
				
			||||||
 | 
					log-path: "../logs"  # directory which will store DBT logs
 | 
				
			||||||
 | 
					modules-path: "../dbt_modules"  # directory which will store external DBT dependencies
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					clean-targets:         # directories to be removed by `dbt clean`
 | 
				
			||||||
 | 
					    - "build"
 | 
				
			||||||
 | 
					    - "dbt_modules"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					quoting:
 | 
				
			||||||
 | 
					  database: true
 | 
				
			||||||
 | 
					# Temporarily disabling the behavior of the ExtendedNameTransformer on table/schema names, see (issue #1785)
 | 
				
			||||||
 | 
					# all schemas should be unquoted
 | 
				
			||||||
 | 
					  schema: false
 | 
				
			||||||
 | 
					  identifier: true
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# You can define configurations for models in the `source-paths` directory here.
 | 
				
			||||||
 | 
					# Using these configurations, you can enable or disable models, change how they
 | 
				
			||||||
 | 
					# are materialized, and more!
 | 
				
			||||||
 | 
					models:
 | 
				
			||||||
 | 
					  airbyte_utils:
 | 
				
			||||||
 | 
					    +materialized: table
 | 
				
			||||||
 | 
					    generated:
 | 
				
			||||||
 | 
					      airbyte_ctes:
 | 
				
			||||||
 | 
					        +tags: airbyte_internal_cte
 | 
				
			||||||
 | 
					        +materialized: ephemeral
 | 
				
			||||||
 | 
					      airbyte_incremental:
 | 
				
			||||||
 | 
					        +tags: incremental_tables
 | 
				
			||||||
 | 
					        +materialized: incremental
 | 
				
			||||||
 | 
					        +on_schema_change: sync_all_columns
 | 
				
			||||||
 | 
					      airbyte_tables:
 | 
				
			||||||
 | 
					        +tags: normalized_tables
 | 
				
			||||||
 | 
					        +materialized: table
 | 
				
			||||||
 | 
					      airbyte_views:
 | 
				
			||||||
 | 
					        +tags: airbyte_internal_views
 | 
				
			||||||
 | 
					        +materialized: view
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					dispatch:
 | 
				
			||||||
 | 
					  - macro_namespace: dbt_utils
 | 
				
			||||||
 | 
					    search_order: ['airbyte_utils', 'dbt_utils']
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1 @@
 | 
				
			||||||
 | 
					{"streams":[{"stream":{"name":"se-orders","json_schema":{"type":"object","properties":{"to":{"type":"string"},"_id":{"type":"string"},"cod":{"type":"number"},"code":{"type":"string"},"from":{"type":"string"},"note":{"type":"string"},"value":{"type":"number"},"client":{"type":"string"},"status":{"type":"string"},"volume":{"type":"string"},"weight":{"type":"number"},"courier":{"type":"string"},"distance":{"type":"number"},"createdAt":{"type":"string"},"updatedAt":{"type":"string"},"updatedBy":{"type":"string"},"itemVolume":{"type":"number"},"searchString":{"type":"string"},"extraServices":{"type":"array"}}},"supported_sync_modes":["full_refresh","incremental"],"default_cursor_field":[],"source_defined_primary_key":[],"namespace":"selly-express"},"sync_mode":"full_refresh","cursor_field":[],"destination_sync_mode":"overwrite","primary_key":[]}]}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1 @@
 | 
				
			||||||
 | 
					{"ssl":false,"host":"18.140.112.89","port":5432,"schema":"public","database":"mongo_etl","password":"bvxJaDGW2R55uyDXfODJ2a0Y","username":"selly","tunnel_method":{"tunnel_method":"NO_TUNNEL"}}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,166 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    Adapter Macros for the following functions:
 | 
				
			||||||
 | 
					    - Bigquery: unnest() -> https://cloud.google.com/bigquery/docs/reference/standard-sql/arrays#flattening-arrays-and-repeated-fields
 | 
				
			||||||
 | 
					    - Snowflake: flatten() -> https://docs.snowflake.com/en/sql-reference/functions/flatten.html
 | 
				
			||||||
 | 
					    - Redshift: -> https://blog.getdbt.com/how-to-unnest-arrays-in-redshift/
 | 
				
			||||||
 | 
					    - postgres: unnest() -> https://www.postgresqltutorial.com/postgresql-array/
 | 
				
			||||||
 | 
					    - MSSQL: openjson() –> https://docs.microsoft.com/en-us/sql/relational-databases/json/validate-query-and-change-json-data-with-built-in-functions-sql-server?view=sql-server-ver15
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# cross_join_unnest -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('cross_join_unnest')(stream_name, array_col) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    {% do exceptions.warn("Undefined macro cross_join_unnest for this destination engine") %}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    cross join unnest({{ array_col }}) as {{ array_col }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    {% do exceptions.warn("Normalization does not support unnesting for Oracle yet.") %}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    cross join jsonb_array_elements(
 | 
				
			||||||
 | 
					        case jsonb_typeof({{ array_col }})
 | 
				
			||||||
 | 
					        when 'array' then {{ array_col }}
 | 
				
			||||||
 | 
					        else '[]' end
 | 
				
			||||||
 | 
					    ) as _airbyte_nested_data
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    left join joined on _airbyte_{{ stream_name }}_hashid = joined._airbyte_hashid
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    left join joined on _airbyte_{{ stream_name }}_hashid = joined._airbyte_hashid
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					    cross join table(flatten({{ array_col }})) as {{ array_col }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__cross_join_unnest(stream_name, array_col) -%}
 | 
				
			||||||
 | 
					{# https://docs.microsoft.com/en-us/sql/relational-databases/json/convert-json-data-to-rows-and-columns-with-openjson-sql-server?view=sql-server-ver15#option-1---openjson-with-the-default-output #}
 | 
				
			||||||
 | 
					    CROSS APPLY (
 | 
				
			||||||
 | 
						    SELECT [value] = CASE
 | 
				
			||||||
 | 
								WHEN [type] = 4 THEN (SELECT [value] FROM OPENJSON([value]))
 | 
				
			||||||
 | 
								WHEN [type] = 5 THEN [value]
 | 
				
			||||||
 | 
								END
 | 
				
			||||||
 | 
						    FROM OPENJSON({{ array_col }})
 | 
				
			||||||
 | 
					    ) AS {{ array_col }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# unnested_column_value -- this macro is related to unnest_cte #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('unnested_column_value')(column_col) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    {{ column_col }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    _airbyte_nested_data
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    {{ column_col }}.value
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    _airbyte_nested_data
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    _airbyte_nested_data
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    {{ column_col }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__unnested_column_value(column_col) -%}
 | 
				
			||||||
 | 
					    {# unnested array/sub_array will be located in `value` column afterwards, we need to address to it #}
 | 
				
			||||||
 | 
					    {{ column_col }}.value
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# unnest_cte -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro unnest_cte(from_table, stream_name, column_col) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('unnest_cte')(from_table, stream_name, column_col) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__unnest_cte(from_table, stream_name, column_col) -%}{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# -- based on https://blog.getdbt.com/how-to-unnest-arrays-in-redshift/ #}
 | 
				
			||||||
 | 
					{% macro redshift__unnest_cte(from_table, stream_name, column_col) -%}
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return('') }}
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					    {%- call statement('max_json_array_length', fetch_result=True) -%}
 | 
				
			||||||
 | 
					        with max_value as (
 | 
				
			||||||
 | 
					            select max(json_array_length({{ column_col }}, true)) as max_number_of_items
 | 
				
			||||||
 | 
					            from {{ from_table }}
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					        select
 | 
				
			||||||
 | 
					            case when max_number_of_items is not null and max_number_of_items > 1
 | 
				
			||||||
 | 
					            then max_number_of_items
 | 
				
			||||||
 | 
					            else 1 end as max_number_of_items
 | 
				
			||||||
 | 
					        from max_value
 | 
				
			||||||
 | 
					    {%- endcall -%}
 | 
				
			||||||
 | 
					    {%- set max_length = load_result('max_json_array_length') -%}
 | 
				
			||||||
 | 
					with numbers as (
 | 
				
			||||||
 | 
					    {{dbt_utils.generate_series(max_length["data"][0][0])}}
 | 
				
			||||||
 | 
					),
 | 
				
			||||||
 | 
					joined as (
 | 
				
			||||||
 | 
					    select
 | 
				
			||||||
 | 
					        _airbyte_{{ stream_name }}_hashid as _airbyte_hashid,
 | 
				
			||||||
 | 
					        json_extract_array_element_text({{ column_col }}, numbers.generated_number::int - 1, true) as _airbyte_nested_data
 | 
				
			||||||
 | 
					    from {{ from_table }}
 | 
				
			||||||
 | 
					    cross join numbers
 | 
				
			||||||
 | 
					    -- only generate the number of records in the cross join that corresponds
 | 
				
			||||||
 | 
					    -- to the number of items in {{ from_table }}.{{ column_col }}
 | 
				
			||||||
 | 
					    where numbers.generated_number <= json_array_length({{ column_col }}, true)
 | 
				
			||||||
 | 
					)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__unnest_cte(from_table, stream_name, column_col) -%}
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return('') }}
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- call statement('max_json_array_length', fetch_result=True) -%}
 | 
				
			||||||
 | 
					        with max_value as (
 | 
				
			||||||
 | 
					            select max(json_length({{ column_col }})) as max_number_of_items
 | 
				
			||||||
 | 
					            from {{ from_table }}
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					        select
 | 
				
			||||||
 | 
					            case when max_number_of_items is not null and max_number_of_items > 1
 | 
				
			||||||
 | 
					            then max_number_of_items
 | 
				
			||||||
 | 
					            else 1 end as max_number_of_items
 | 
				
			||||||
 | 
					        from max_value
 | 
				
			||||||
 | 
					    {%- endcall -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- set max_length = load_result('max_json_array_length') -%}
 | 
				
			||||||
 | 
					    with numbers as (
 | 
				
			||||||
 | 
					        {{ dbt_utils.generate_series(max_length["data"][0][0]) }}
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					    joined as (
 | 
				
			||||||
 | 
					        select
 | 
				
			||||||
 | 
					            _airbyte_{{ stream_name }}_hashid as _airbyte_hashid,
 | 
				
			||||||
 | 
					            {# -- json_extract(column_col, '$[i][0]') as _airbyte_nested_data #}
 | 
				
			||||||
 | 
					            json_extract({{ column_col }}, concat("$[", numbers.generated_number - 1, "][0]")) as _airbyte_nested_data
 | 
				
			||||||
 | 
					        from {{ from_table }}
 | 
				
			||||||
 | 
					        cross join numbers
 | 
				
			||||||
 | 
					        -- only generate the number of records in the cross join that corresponds
 | 
				
			||||||
 | 
					        -- to the number of items in {{ from_table }}.{{ column_col }}
 | 
				
			||||||
 | 
					        where numbers.generated_number <= json_length({{ column_col }})
 | 
				
			||||||
 | 
					    )
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,23 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    concat in dbt 0.6.4 used to work fine for bigquery but the new implementaion in 0.7.3 is less scalable (can not handle too many columns)
 | 
				
			||||||
 | 
					    Therefore, we revert the implementation here and add versions for missing destinations
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro concat(fields) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('concat')(fields) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__concat(fields) -%}
 | 
				
			||||||
 | 
					    {#-- concat() in SQL bigquery scales better with number of columns than using the '||' operator --#}
 | 
				
			||||||
 | 
					    concat({{ fields|join(', ') }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__concat(fields) -%}
 | 
				
			||||||
 | 
					    {#-- CONCAT() in SQL SERVER accepts from 2 to 254 arguments, we use batches for the main concat, to overcome the limit. --#}
 | 
				
			||||||
 | 
					    {% set concat_chunks = [] %}
 | 
				
			||||||
 | 
					    {% for chunk in fields|batch(253) -%}
 | 
				
			||||||
 | 
					        {% set _ = concat_chunks.append( "concat(" ~ chunk|join(', ') ~ ",'')" ) %}
 | 
				
			||||||
 | 
					    {% endfor %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    concat({{ concat_chunks|join(', ') }}, '')
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,7 @@
 | 
				
			||||||
 | 
					{% macro mysql__current_timestamp() %}
 | 
				
			||||||
 | 
					    CURRENT_TIMESTAMP
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__current_timestamp() %}
 | 
				
			||||||
 | 
					    CURRENT_TIMESTAMP
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,181 @@
 | 
				
			||||||
 | 
					{# json  -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro type_json() -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('type_json')() }}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__type_json() %}
 | 
				
			||||||
 | 
					    string
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro redshift__type_json() -%}
 | 
				
			||||||
 | 
					    varchar
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__type_json() %}
 | 
				
			||||||
 | 
					    jsonb
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro oracle__type_json() -%}
 | 
				
			||||||
 | 
					    varchar2(4000)
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__type_json() %}
 | 
				
			||||||
 | 
					    variant
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro mysql__type_json() -%}
 | 
				
			||||||
 | 
					    json
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro sqlserver__type_json() -%}
 | 
				
			||||||
 | 
					    VARCHAR(max)
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_json() %}
 | 
				
			||||||
 | 
					    String
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# string ------------------------------------------------- #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro mysql__type_string() -%}
 | 
				
			||||||
 | 
					    char
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro oracle__type_string() -%}
 | 
				
			||||||
 | 
					    varchar2(4000)
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__type_string() %}
 | 
				
			||||||
 | 
					    VARCHAR(max)
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro clickhouse__type_string() -%}
 | 
				
			||||||
 | 
					    String
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# float ------------------------------------------------- #}
 | 
				
			||||||
 | 
					{% macro mysql__type_float() %}
 | 
				
			||||||
 | 
					    float
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__type_float() %}
 | 
				
			||||||
 | 
					    float
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_float() %}
 | 
				
			||||||
 | 
					    Float64
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# int  ------------------------------------------------- #}
 | 
				
			||||||
 | 
					{% macro default__type_int() %}
 | 
				
			||||||
 | 
					    signed
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__type_int() %}
 | 
				
			||||||
 | 
					    int
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_int() %}
 | 
				
			||||||
 | 
					    INT
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# bigint ------------------------------------------------- #}
 | 
				
			||||||
 | 
					{% macro mysql__type_bigint() %}
 | 
				
			||||||
 | 
					    signed
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__type_bigint() %}
 | 
				
			||||||
 | 
					    numeric
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_bigint() %}
 | 
				
			||||||
 | 
					    BIGINT
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# numeric ------------------------------------------------- --#}
 | 
				
			||||||
 | 
					{% macro mysql__type_numeric() %}
 | 
				
			||||||
 | 
					    float
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_numeric() %}
 | 
				
			||||||
 | 
					    Float64
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# timestamp ------------------------------------------------- --#}
 | 
				
			||||||
 | 
					{% macro mysql__type_timestamp() %}
 | 
				
			||||||
 | 
					    time
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro sqlserver__type_timestamp() -%}
 | 
				
			||||||
 | 
					    {#-- in TSQL timestamp is really datetime --#}
 | 
				
			||||||
 | 
					    {#-- https://docs.microsoft.com/en-us/sql/t-sql/functions/date-and-time-data-types-and-functions-transact-sql?view=sql-server-ver15#DateandTimeDataTypes --#}
 | 
				
			||||||
 | 
					    datetime
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_timestamp() %}
 | 
				
			||||||
 | 
					    DateTime64
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# timestamp with time zone  -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro type_timestamp_with_timezone() -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('type_timestamp_with_timezone')() }}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__type_timestamp_with_timezone() %}
 | 
				
			||||||
 | 
					    timestamp with time zone
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__type_timestamp_with_timezone() %}
 | 
				
			||||||
 | 
					    timestamp
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{#-- MySQL doesnt allow cast operation to work with TIMESTAMP so we have to use char --#}
 | 
				
			||||||
 | 
					{%- macro mysql__type_timestamp_with_timezone() -%}
 | 
				
			||||||
 | 
					    char
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__type_timestamp_with_timezone() %}
 | 
				
			||||||
 | 
					    varchar2(4000)
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro sqlserver__type_timestamp_with_timezone() -%}
 | 
				
			||||||
 | 
					    {#-- in TSQL timestamp is really datetime or datetime2 --#}
 | 
				
			||||||
 | 
					    {#-- https://docs.microsoft.com/en-us/sql/t-sql/functions/date-and-time-data-types-and-functions-transact-sql?view=sql-server-ver15#DateandTimeDataTypes --#}
 | 
				
			||||||
 | 
					    datetime
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_timestamp_with_timezone() %}
 | 
				
			||||||
 | 
					    DateTime64
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# date  -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro type_date() -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('type_date')() }}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__type_date() %}
 | 
				
			||||||
 | 
					    date
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__type_date() %}
 | 
				
			||||||
 | 
					    varchar2(4000)
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro sqlserver__type_date() -%}
 | 
				
			||||||
 | 
					    date
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__type_date() %}
 | 
				
			||||||
 | 
					    Date
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,7 @@
 | 
				
			||||||
 | 
					{% macro mysql__except() %}
 | 
				
			||||||
 | 
					    {% do exceptions.warn("MySQL does not support EXCEPT operator") %}
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__except() %}
 | 
				
			||||||
 | 
					    minus
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,5 @@
 | 
				
			||||||
 | 
					{# converting hash in varchar _macro #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__hash(field) -%}
 | 
				
			||||||
 | 
					    convert(varchar(32), HashBytes('md5',  coalesce(cast({{field}} as {{dbt_utils.type_string()}}), '')), 2)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,226 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    Adapter Macros for the following functions:
 | 
				
			||||||
 | 
					    - Bigquery: JSON_EXTRACT(json_string_expr, json_path_format) -> https://cloud.google.com/bigquery/docs/reference/standard-sql/json_functions
 | 
				
			||||||
 | 
					    - Snowflake: JSON_EXTRACT_PATH_TEXT( <column_identifier> , '<path_name>' ) -> https://docs.snowflake.com/en/sql-reference/functions/json_extract_path_text.html
 | 
				
			||||||
 | 
					    - Redshift: json_extract_path_text('json_string', 'path_elem' [,'path_elem'[, ...] ] [, null_if_invalid ] ) -> https://docs.aws.amazon.com/redshift/latest/dg/JSON_EXTRACT_PATH_TEXT.html
 | 
				
			||||||
 | 
					    - Postgres: json_extract_path_text(<from_json>, 'path' [, 'path' [, ...}}) -> https://www.postgresql.org/docs/12/functions-json.html
 | 
				
			||||||
 | 
					    - MySQL: JSON_EXTRACT(json_doc, 'path' [, 'path'] ...) -> https://dev.mysql.com/doc/refman/8.0/en/json-search-functions.html
 | 
				
			||||||
 | 
					    - ClickHouse: JSONExtractString(json_doc, 'path' [, 'path'] ...) -> https://clickhouse.com/docs/en/sql-reference/functions/json-functions/
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# format_json_path --------------------------------------------------     #}
 | 
				
			||||||
 | 
					{% macro format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {{ adapter.dispatch('format_json_path')(json_path_list) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {{ '.' ~ json_path_list|join('.') }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					  {{ '\'$."' ~ json_path_list|join('."') ~ '"\'' }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace('"', '\\"')) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ '"$[\'' ~ str_list|join('\'][\'') ~ '\']"' }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace("'", "''")) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ "'" ~ str_list|join("','") ~ "'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {# -- '$."x"."y"."z"' #}
 | 
				
			||||||
 | 
					    {{ "'$.\"" ~ json_path_list|join(".") ~ "\"'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace("'", "''")) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ "'" ~ str_list|join("','") ~ "'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace("'", "''").replace('"', '""')) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ "'\"" ~ str_list|join('"."') ~ "\"'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {# -- '$."x"."y"."z"' #}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace("'", "''").replace('"', '\\"')) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ "'$.\"" ~ str_list|join(".") ~ "\"'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__format_json_path(json_path_list) -%}
 | 
				
			||||||
 | 
					    {%- set str_list = [] -%}
 | 
				
			||||||
 | 
					    {%- for json_path in json_path_list -%}
 | 
				
			||||||
 | 
					        {%- if str_list.append(json_path.replace("'", "''").replace('"', '\\"')) -%} {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {{ "'" ~ str_list|join("','") ~ "'" }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# json_extract -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {{ adapter.dispatch('json_extract')(from_table, json_column, json_path_list, normalized_json_path) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract({{ from_table}}.{{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_value({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        json_extract({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        json_extract({{ from_table}}.{{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        jsonb_extract_path({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        jsonb_extract_path({{ from_table }}.{{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        json_extract({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        json_extract({{ from_table }}.{{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        case when json_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }}, true) != '' then json_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }}, true) end
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        case when json_extract_path_text({{ from_table }}.{{ json_column }}, {{ format_json_path(json_path_list) }}, true) != '' then json_extract_path_text({{ from_table }}.{{ json_column }}, {{ format_json_path(json_path_list) }}, true) end
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        get_path(parse_json({{ json_column }}), {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        get_path(parse_json({{ from_table }}.{{ json_column }}), {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_query({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__json_extract(from_table, json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {%- if from_table|string() == '' %}
 | 
				
			||||||
 | 
					        JSONExtractRaw({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% else %}
 | 
				
			||||||
 | 
					        JSONExtractRaw({{ from_table }}.{{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					    {% endif -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# json_extract_scalar -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {{ adapter.dispatch('json_extract_scalar')(json_column, json_path_list, normalized_json_path) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract_scalar({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_value({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract_scalar({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    jsonb_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_value({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    case when json_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }}, true) != '' then json_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }}, true) end
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    to_varchar(get_path(parse_json({{ json_column }}), {{ format_json_path(json_path_list) }}))
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_value({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__json_extract_scalar(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    JSONExtractRaw({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# json_extract_array -------------------------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    {{ adapter.dispatch('json_extract_array')(json_column, json_path_list, normalized_json_path) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract_array({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_value({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract_array({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro postgres__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    jsonb_extract_path({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract({{ json_column }}, {{ format_json_path(normalized_json_path) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_extract_path_text({{ json_column }}, {{ format_json_path(json_path_list) }}, true)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro snowflake__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    get_path(parse_json({{ json_column }}), {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    json_query({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__json_extract_array(json_column, json_path_list, normalized_json_path) -%}
 | 
				
			||||||
 | 
					    JSONExtractArrayRaw({{ json_column }}, {{ format_json_path(json_path_list) }})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,16 @@
 | 
				
			||||||
 | 
					{# quote  ----------------------------------     #}
 | 
				
			||||||
 | 
					{% macro quote(column_name) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('quote')(column_name) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__quote(column_name) -%}
 | 
				
			||||||
 | 
					  adapter.quote(column_name)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__quote(column_name) -%}
 | 
				
			||||||
 | 
					  {{ '\"' ~ column_name ~ '\"'}}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__quote(column_name) -%}
 | 
				
			||||||
 | 
					  {{ '\"' ~ column_name ~ '\"'}}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,25 @@
 | 
				
			||||||
 | 
					{# surrogate_key  ----------------------------------     #}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__surrogate_key(field_list) -%}
 | 
				
			||||||
 | 
					    ora_hash(
 | 
				
			||||||
 | 
					        {%- for field in field_list %}
 | 
				
			||||||
 | 
					            {% if not loop.last %}
 | 
				
			||||||
 | 
					                {{ field }} || '~' ||
 | 
				
			||||||
 | 
					            {% else %}
 | 
				
			||||||
 | 
					                {{ field }}
 | 
				
			||||||
 | 
					            {% endif %}
 | 
				
			||||||
 | 
					        {%- endfor %}
 | 
				
			||||||
 | 
					    )
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro clickhouse__surrogate_key(field_list) -%}
 | 
				
			||||||
 | 
					    assumeNotNull(hex(MD5(
 | 
				
			||||||
 | 
					        {%- for field in field_list %}
 | 
				
			||||||
 | 
					            {% if not loop.last %}
 | 
				
			||||||
 | 
					                toString({{ field }}) || '~' ||
 | 
				
			||||||
 | 
					            {% else %}
 | 
				
			||||||
 | 
					                toString({{ field }})
 | 
				
			||||||
 | 
					            {% endif %}
 | 
				
			||||||
 | 
					        {%- endfor %}
 | 
				
			||||||
 | 
					    )))
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,67 @@
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# boolean_to_string -------------------------------------------------     #}
 | 
				
			||||||
 | 
					{% macro boolean_to_string(boolean_column) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('boolean_to_string')(boolean_column) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__boolean_to_string(boolean_column) -%}
 | 
				
			||||||
 | 
					    {{ boolean_column }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro redshift__boolean_to_string(boolean_column) -%}
 | 
				
			||||||
 | 
					    case when {{ boolean_column }} then 'true' else 'false' end
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# array_to_string -------------------------------------------------     #}
 | 
				
			||||||
 | 
					{% macro array_to_string(array_column) -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('array_to_string')(array_column) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__array_to_string(array_column) -%}
 | 
				
			||||||
 | 
					    {{ array_column }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro bigquery__array_to_string(array_column) -%}
 | 
				
			||||||
 | 
					    array_to_string({{ array_column }}, "|", "")
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__array_to_string(array_column) -%}
 | 
				
			||||||
 | 
					    cast({{ array_column }} as varchar2(4000))
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro sqlserver__array_to_string(array_column) -%}
 | 
				
			||||||
 | 
					    cast({{ array_column }} as {{dbt_utils.type_string()}})
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# cast_to_boolean -------------------------------------------------     #}
 | 
				
			||||||
 | 
					{% macro cast_to_boolean(field) -%}
 | 
				
			||||||
 | 
					    {{ adapter.dispatch('cast_to_boolean')(field) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro default__cast_to_boolean(field) -%}
 | 
				
			||||||
 | 
					    cast({{ field }} as boolean)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# -- MySQL does not support cast function converting string directly to boolean (an alias of tinyint(1), https://dev.mysql.com/doc/refman/8.0/en/cast-functions.html#function_cast #}
 | 
				
			||||||
 | 
					{% macro mysql__cast_to_boolean(field) -%}
 | 
				
			||||||
 | 
					    IF(lower({{ field }}) = 'true', true, false)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# -- Redshift does not support converting string directly to boolean, it must go through int first #}
 | 
				
			||||||
 | 
					{% macro redshift__cast_to_boolean(field) -%}
 | 
				
			||||||
 | 
					    cast(decode({{ field }}, 'true', '1', 'false', '0')::integer as boolean)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# -- MS SQL Server does not support converting string directly to boolean, it must be casted as bit #}
 | 
				
			||||||
 | 
					{% macro sqlserver__cast_to_boolean(field) -%}
 | 
				
			||||||
 | 
					    cast({{ field }} as bit)
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# empty_string_to_null -------------------------------------------------     #}
 | 
				
			||||||
 | 
					{% macro empty_string_to_null(field) -%}
 | 
				
			||||||
 | 
					    {{ return(adapter.dispatch('empty_string_to_null')(field)) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro default__empty_string_to_null(field) -%}
 | 
				
			||||||
 | 
					    nullif({{ field }}, '')
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,4 @@
 | 
				
			||||||
 | 
					-- see https://docs.getdbt.com/docs/building-a-dbt-project/building-models/using-custom-schemas/#an-alternative-pattern-for-generating-schema-names
 | 
				
			||||||
 | 
					{% macro generate_schema_name(custom_schema_name, node) -%}
 | 
				
			||||||
 | 
					    {{ generate_schema_name_for_env(custom_schema_name, node) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,51 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    These macros control how incremental models are updated in Airbyte's normalization step
 | 
				
			||||||
 | 
					    - get_max_normalized_cursor retrieve the value of the last normalized data
 | 
				
			||||||
 | 
					    - incremental_clause controls the predicate to filter on new data to process incrementally
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro incremental_clause(col_emitted_at)  -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('incremental_clause')(col_emitted_at) }}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro default__incremental_clause(col_emitted_at) -%}
 | 
				
			||||||
 | 
					{% if is_incremental() %}
 | 
				
			||||||
 | 
					and coalesce(
 | 
				
			||||||
 | 
					    cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }}) >= (select max(cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }})) from {{ this }}),
 | 
				
			||||||
 | 
					    {# -- if {{ col_emitted_at }} is NULL in either table, the previous comparison would evaluate to NULL, #}
 | 
				
			||||||
 | 
					    {# -- so we coalesce and make sure the row is always returned for incremental processing instead #}
 | 
				
			||||||
 | 
					    true)
 | 
				
			||||||
 | 
					{% endif %}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{# -- see https://on-systems.tech/113-beware-dbt-incremental-updates-against-snowflake-external-tables/ #}
 | 
				
			||||||
 | 
					{%- macro snowflake__incremental_clause(col_emitted_at) -%}
 | 
				
			||||||
 | 
					{% if is_incremental() %}
 | 
				
			||||||
 | 
					    {% if get_max_normalized_cursor(col_emitted_at) %}
 | 
				
			||||||
 | 
					and cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }}) >=
 | 
				
			||||||
 | 
					    cast('{{ get_max_normalized_cursor(col_emitted_at) }}' as {{ type_timestamp_with_timezone() }})
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					{% endif %}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro sqlserver__incremental_clause(col_emitted_at) -%}
 | 
				
			||||||
 | 
					{% if is_incremental() %}
 | 
				
			||||||
 | 
					and ((select max(cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }})) from {{ this }}) is null
 | 
				
			||||||
 | 
					  or cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }}) >=
 | 
				
			||||||
 | 
					     (select max(cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }})) from {{ this }}))
 | 
				
			||||||
 | 
					{% endif %}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro get_max_normalized_cursor(col_emitted_at) %}
 | 
				
			||||||
 | 
					{% if execute and is_incremental() %}
 | 
				
			||||||
 | 
					 {% if env_var('INCREMENTAL_CURSOR', 'UNSET') == 'UNSET' %}
 | 
				
			||||||
 | 
					     {% set query %}
 | 
				
			||||||
 | 
					        select max(cast({{ col_emitted_at }} as {{ type_timestamp_with_timezone() }})) from {{ this }}
 | 
				
			||||||
 | 
					     {% endset %}
 | 
				
			||||||
 | 
					     {% set max_cursor = run_query(query).columns[0][0] %}
 | 
				
			||||||
 | 
					     {% do return(max_cursor) %}
 | 
				
			||||||
 | 
					 {% else %}
 | 
				
			||||||
 | 
					    {% do return(env_var('INCREMENTAL_CURSOR')) %}
 | 
				
			||||||
 | 
					 {% endif %}
 | 
				
			||||||
 | 
					{% endif %}
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,34 @@
 | 
				
			||||||
 | 
					{% macro oracle__test_equal_rowcount(model, compare_model) %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{#-- Needs to be set at parse time, before we return '' below --#}
 | 
				
			||||||
 | 
					{{ config(fail_calc = 'coalesce(diff_count, 0)') }}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{#-- Prevent querying of db in parsing mode. This works because this macro does not create any new refs. #}
 | 
				
			||||||
 | 
					{%- if not execute -%}
 | 
				
			||||||
 | 
					    {{ return('') }}
 | 
				
			||||||
 | 
					{% endif %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					with a as (
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    select count(*) as count_a from {{ model }}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					),
 | 
				
			||||||
 | 
					b as (
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    select count(*) as count_b from {{ compare_model }}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					),
 | 
				
			||||||
 | 
					final as (
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    select
 | 
				
			||||||
 | 
					        count_a,
 | 
				
			||||||
 | 
					        count_b,
 | 
				
			||||||
 | 
					        abs(count_a - count_b) as diff_count
 | 
				
			||||||
 | 
					    from a
 | 
				
			||||||
 | 
					    cross join b
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					select diff_count from final
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,107 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					-- Adapted from https://github.com/dbt-labs/dbt-utils/blob/0-19-0-updates/macros/schema_tests/equality.sql
 | 
				
			||||||
 | 
					-- dbt-utils version: 0.6.4
 | 
				
			||||||
 | 
					-- This macro needs to be updated accordingly when dbt-utils is upgraded.
 | 
				
			||||||
 | 
					-- This is needed because MySQL does not support the EXCEPT operator!
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro mysql__test_equality(model, compare_model, compare_columns=None) %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return('') }}
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_relation(model, 'test_equality') -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- if not compare_columns -%}
 | 
				
			||||||
 | 
					        {%- do dbt_utils._is_ephemeral(model, 'test_equality') -%}
 | 
				
			||||||
 | 
					        {%- set compare_columns = adapter.get_columns_in_relation(model) | map(attribute='quoted') -%}
 | 
				
			||||||
 | 
					    {%- endif -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {% set compare_cols_csv = compare_columns | join(', ') %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    with a as (
 | 
				
			||||||
 | 
					        select * from {{ model }}
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    b as (
 | 
				
			||||||
 | 
					        select * from {{ compare_model }}
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    a_minus_b as (
 | 
				
			||||||
 | 
					        select {{ compare_cols_csv }} from a
 | 
				
			||||||
 | 
					        where ({{ compare_cols_csv }}) not in
 | 
				
			||||||
 | 
					            (select {{ compare_cols_csv }} from b)
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    b_minus_a as (
 | 
				
			||||||
 | 
					        select {{ compare_cols_csv }} from b
 | 
				
			||||||
 | 
					        where ({{ compare_cols_csv }}) not in
 | 
				
			||||||
 | 
					            (select {{ compare_cols_csv }} from a)
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    unioned as (
 | 
				
			||||||
 | 
					        select * from a_minus_b
 | 
				
			||||||
 | 
					        union all
 | 
				
			||||||
 | 
					        select * from b_minus_a
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    final as (
 | 
				
			||||||
 | 
					        select (select count(*) from unioned) +
 | 
				
			||||||
 | 
					        (select abs(
 | 
				
			||||||
 | 
					            (select count(*) from a_minus_b) -
 | 
				
			||||||
 | 
					            (select count(*) from b_minus_a)
 | 
				
			||||||
 | 
					            ))
 | 
				
			||||||
 | 
					        as count
 | 
				
			||||||
 | 
					    )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    select count from final
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{% macro oracle__test_equality(model) %}
 | 
				
			||||||
 | 
					    {#-- Prevent querying of db in parsing mode. This works because this macro does not create any new refs. #}
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return('') }}
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    -- setup
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_relation(model, 'test_equality') -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {#-
 | 
				
			||||||
 | 
					    If the compare_cols arg is provided, we can run this test without querying the
 | 
				
			||||||
 | 
					    information schema — this allows the model to be an ephemeral model
 | 
				
			||||||
 | 
					    -#}
 | 
				
			||||||
 | 
					    {%- set compare_columns = kwargs.get('compare_columns', None) -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- if not compare_columns -%}
 | 
				
			||||||
 | 
					        {%- do dbt_utils._is_ephemeral(model, 'test_equality') -%}
 | 
				
			||||||
 | 
					        {%- set compare_columns = adapter.get_columns_in_relation(model) | map(attribute='quoted') -%}
 | 
				
			||||||
 | 
					    {%- endif -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {% set compare_model = kwargs.get('compare_model', kwargs.get('arg')) %}
 | 
				
			||||||
 | 
					    {% set compare_cols_csv = compare_columns | join(', ') %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    with a as (
 | 
				
			||||||
 | 
					        select * from {{ model }}
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					    b as (
 | 
				
			||||||
 | 
					        select * from {{ compare_model }}
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					    a_minus_b as (
 | 
				
			||||||
 | 
					        select {{compare_cols_csv}} from a
 | 
				
			||||||
 | 
					        {{ dbt_utils.except() }}
 | 
				
			||||||
 | 
					        select {{compare_cols_csv}} from b
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					    b_minus_a as (
 | 
				
			||||||
 | 
					        select {{compare_cols_csv}} from b
 | 
				
			||||||
 | 
					        {{ dbt_utils.except() }}
 | 
				
			||||||
 | 
					        select {{compare_cols_csv}} from a
 | 
				
			||||||
 | 
					    ),
 | 
				
			||||||
 | 
					    unioned as (
 | 
				
			||||||
 | 
					        select * from a_minus_b
 | 
				
			||||||
 | 
					        union all
 | 
				
			||||||
 | 
					        select * from b_minus_a
 | 
				
			||||||
 | 
					    )
 | 
				
			||||||
 | 
					    select count(*) from unioned
 | 
				
			||||||
 | 
					{% endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,51 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    This overrides the behavior of the macro `should_full_refresh` so full refresh are triggered if:
 | 
				
			||||||
 | 
					    - the dbt cli is run with --full-refresh flag or the model is configured explicitly to full_refresh
 | 
				
			||||||
 | 
					    - the column _airbyte_ab_id does not exists in the normalized tables and make sure it is well populated.
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro need_full_refresh(col_ab_id, target_table=this) -%}
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return(false) }}
 | 
				
			||||||
 | 
					    {%- endif -%}
 | 
				
			||||||
 | 
					    {%- set found_column = [] %}
 | 
				
			||||||
 | 
					    {%- set cols = adapter.get_columns_in_relation(target_table) -%}
 | 
				
			||||||
 | 
					    {%- for col in cols -%}
 | 
				
			||||||
 | 
					        {%- if col.column == col_ab_id -%}
 | 
				
			||||||
 | 
					            {% do found_column.append(col.column) %}
 | 
				
			||||||
 | 
					        {%- endif -%}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					    {%- if found_column -%}
 | 
				
			||||||
 | 
					        {{ return(false) }}
 | 
				
			||||||
 | 
					    {%- else -%}
 | 
				
			||||||
 | 
					        {{ dbt_utils.log_info(target_table ~ "." ~ col_ab_id ~ " does not exist yet. The table will be created or rebuilt with dbt.full_refresh") }}
 | 
				
			||||||
 | 
					        {{ return(true) }}
 | 
				
			||||||
 | 
					    {%- endif -%}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro should_full_refresh() -%}
 | 
				
			||||||
 | 
					  {% set config_full_refresh = config.get('full_refresh') %}
 | 
				
			||||||
 | 
					  {%- if config_full_refresh is none -%}
 | 
				
			||||||
 | 
					    {% set config_full_refresh = flags.FULL_REFRESH %}
 | 
				
			||||||
 | 
					  {%- endif -%}
 | 
				
			||||||
 | 
					  {%- if not config_full_refresh -%}
 | 
				
			||||||
 | 
					    {% set config_full_refresh = need_full_refresh(get_col_ab_id(), this) %}
 | 
				
			||||||
 | 
					  {%- endif -%}
 | 
				
			||||||
 | 
					  {% do return(config_full_refresh) %}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro get_col_ab_id() -%}
 | 
				
			||||||
 | 
					  {{ adapter.dispatch('get_col_ab_id')() }}
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro default__get_col_ab_id() -%}
 | 
				
			||||||
 | 
					    _airbyte_ab_id
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro oracle__get_col_ab_id() -%}
 | 
				
			||||||
 | 
					    "_AIRBYTE_AB_ID"
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					{%- macro snowflake__get_col_ab_id() -%}
 | 
				
			||||||
 | 
					    _AIRBYTE_AB_ID
 | 
				
			||||||
 | 
					{%- endmacro -%}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,46 @@
 | 
				
			||||||
 | 
					{#
 | 
				
			||||||
 | 
					    Similar to the star macro here: https://github.com/dbt-labs/dbt-utils/blob/main/macros/sql/star.sql
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    This star_intersect macro takes an additional 'intersect' relation as argument.
 | 
				
			||||||
 | 
					    Its behavior is to select columns from both 'intersect' and 'from' relations with the following rules:
 | 
				
			||||||
 | 
					    - if the columns are existing in both 'from' and the 'intersect' relations, then the column from 'intersect' is used
 | 
				
			||||||
 | 
					    - if it's not in the both relation, then only the column in the 'from' relation is used
 | 
				
			||||||
 | 
					#}
 | 
				
			||||||
 | 
					{% macro star_intersect(from, intersect, from_alias=False, intersect_alias=False, except=[]) -%}
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_relation(from, 'star_intersect') -%}
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_ephemeral(from, 'star_intersect') -%}
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_relation(intersect, 'star_intersect') -%}
 | 
				
			||||||
 | 
					    {%- do dbt_utils._is_ephemeral(intersect, 'star_intersect') -%}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {#-- Prevent querying of db in parsing mode. This works because this macro does not create any new refs. #}
 | 
				
			||||||
 | 
					    {%- if not execute -%}
 | 
				
			||||||
 | 
					        {{ return('') }}
 | 
				
			||||||
 | 
					    {% endif %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- set include_cols = [] %}
 | 
				
			||||||
 | 
					    {%- set cols = adapter.get_columns_in_relation(from) -%}
 | 
				
			||||||
 | 
					    {%- set except = except | map("lower") | list %}
 | 
				
			||||||
 | 
					    {%- for col in cols -%}
 | 
				
			||||||
 | 
					        {%- if col.column|lower not in except -%}
 | 
				
			||||||
 | 
					            {% do include_cols.append(col.column) %}
 | 
				
			||||||
 | 
					        {%- endif %}
 | 
				
			||||||
 | 
					    {%- endfor %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- set include_intersect_cols = [] %}
 | 
				
			||||||
 | 
					    {%- set intersect_cols = adapter.get_columns_in_relation(intersect) -%}
 | 
				
			||||||
 | 
					    {%- for col in intersect_cols -%}
 | 
				
			||||||
 | 
					        {%- if col.column|lower not in except -%}
 | 
				
			||||||
 | 
					            {% do include_intersect_cols.append(col.column) %}
 | 
				
			||||||
 | 
					        {%- endif %}
 | 
				
			||||||
 | 
					    {%- endfor %}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    {%- for col in include_cols %}
 | 
				
			||||||
 | 
					        {%- if col in include_intersect_cols -%}
 | 
				
			||||||
 | 
					            {%- if intersect_alias %}{{ intersect_alias }}.{% else %}{%- endif -%}{{ adapter.quote(col)|trim }}
 | 
				
			||||||
 | 
					            {%- if not loop.last %},{{ '\n  ' }}{% endif %}
 | 
				
			||||||
 | 
					        {%- else %}
 | 
				
			||||||
 | 
					            {%- if from_alias %}{{ from_alias }}.{% else %}{{ from }}.{%- endif -%}{{ adapter.quote(col)|trim }} as {{ adapter.quote(col)|trim }}
 | 
				
			||||||
 | 
					            {%- if not loop.last %},{{ '\n  ' }}{% endif %}
 | 
				
			||||||
 | 
					        {%- endif %}
 | 
				
			||||||
 | 
					    {%- endfor -%}
 | 
				
			||||||
 | 
					{%- endmacro %}
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,35 @@
 | 
				
			||||||
 | 
					{{ config(
 | 
				
			||||||
 | 
					    indexes = [{'columns':['_airbyte_emitted_at'],'type':'btree'}],
 | 
				
			||||||
 | 
					    unique_key = '_airbyte_ab_id',
 | 
				
			||||||
 | 
					    schema = "_airbyte_selly_express",
 | 
				
			||||||
 | 
					    tags = [ "top-level-intermediate" ]
 | 
				
			||||||
 | 
					) }}
 | 
				
			||||||
 | 
					-- SQL model to parse JSON blob stored in a single column and extract into separated field columns as described by the JSON Schema
 | 
				
			||||||
 | 
					-- depends_on: {{ source('selly_express', '_airbyte_raw_se_orders') }}
 | 
				
			||||||
 | 
					select
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['to'], ['to']) }} as {{ adapter.quote('to') }},
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['_id'], ['_id']) }} as _id,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['cod'], ['cod']) }} as cod,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['code'], ['code']) }} as code,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['from'], ['from']) }} as {{ adapter.quote('from') }},
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['note'], ['note']) }} as note,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['value'], ['value']) }} as {{ adapter.quote('value') }},
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['client'], ['client']) }} as client,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['status'], ['status']) }} as status,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['volume'], ['volume']) }} as volume,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['weight'], ['weight']) }} as weight,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['courier'], ['courier']) }} as courier,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['distance'], ['distance']) }} as distance,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['createdAt'], ['createdAt']) }} as createdat,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['updatedAt'], ['updatedAt']) }} as updatedat,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['updatedBy'], ['updatedBy']) }} as updatedby,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['itemVolume'], ['itemVolume']) }} as itemvolume,
 | 
				
			||||||
 | 
					    {{ json_extract_scalar('_airbyte_data', ['searchString'], ['searchString']) }} as searchstring,
 | 
				
			||||||
 | 
					    {{ json_extract_array('_airbyte_data', ['extraServices'], ['extraServices']) }} as extraservices,
 | 
				
			||||||
 | 
					    _airbyte_ab_id,
 | 
				
			||||||
 | 
					    _airbyte_emitted_at,
 | 
				
			||||||
 | 
					    {{ current_timestamp() }} as _airbyte_normalized_at
 | 
				
			||||||
 | 
					from {{ source('selly_express', '_airbyte_raw_se_orders') }} as table_alias
 | 
				
			||||||
 | 
					-- se_orders
 | 
				
			||||||
 | 
					where 1 = 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,35 @@
 | 
				
			||||||
 | 
					{{ config(
 | 
				
			||||||
 | 
					    indexes = [{'columns':['_airbyte_emitted_at'],'type':'btree'}],
 | 
				
			||||||
 | 
					    unique_key = '_airbyte_ab_id',
 | 
				
			||||||
 | 
					    schema = "_airbyte_selly_express",
 | 
				
			||||||
 | 
					    tags = [ "top-level-intermediate" ]
 | 
				
			||||||
 | 
					) }}
 | 
				
			||||||
 | 
					-- SQL model to cast each column to its adequate SQL type converted from the JSON schema type
 | 
				
			||||||
 | 
					-- depends_on: {{ ref('se_orders_ab1') }}
 | 
				
			||||||
 | 
					select
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('to') }} as {{ dbt_utils.type_string() }}) as {{ adapter.quote('to') }},
 | 
				
			||||||
 | 
					    cast(_id as {{ dbt_utils.type_string() }}) as _id,
 | 
				
			||||||
 | 
					    cast(cod as {{ dbt_utils.type_float() }}) as cod,
 | 
				
			||||||
 | 
					    cast(code as {{ dbt_utils.type_string() }}) as code,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('from') }} as {{ dbt_utils.type_string() }}) as {{ adapter.quote('from') }},
 | 
				
			||||||
 | 
					    cast(note as {{ dbt_utils.type_string() }}) as note,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('value') }} as {{ dbt_utils.type_float() }}) as {{ adapter.quote('value') }},
 | 
				
			||||||
 | 
					    cast(client as {{ dbt_utils.type_string() }}) as client,
 | 
				
			||||||
 | 
					    cast(status as {{ dbt_utils.type_string() }}) as status,
 | 
				
			||||||
 | 
					    cast(volume as {{ dbt_utils.type_string() }}) as volume,
 | 
				
			||||||
 | 
					    cast(weight as {{ dbt_utils.type_float() }}) as weight,
 | 
				
			||||||
 | 
					    cast(courier as {{ dbt_utils.type_string() }}) as courier,
 | 
				
			||||||
 | 
					    cast(distance as {{ dbt_utils.type_float() }}) as distance,
 | 
				
			||||||
 | 
					    cast(createdat as {{ dbt_utils.type_string() }}) as createdat,
 | 
				
			||||||
 | 
					    cast(updatedat as {{ dbt_utils.type_string() }}) as updatedat,
 | 
				
			||||||
 | 
					    cast(updatedby as {{ dbt_utils.type_string() }}) as updatedby,
 | 
				
			||||||
 | 
					    cast(itemvolume as {{ dbt_utils.type_float() }}) as itemvolume,
 | 
				
			||||||
 | 
					    cast(searchstring as {{ dbt_utils.type_string() }}) as searchstring,
 | 
				
			||||||
 | 
					    extraservices,
 | 
				
			||||||
 | 
					    _airbyte_ab_id,
 | 
				
			||||||
 | 
					    _airbyte_emitted_at,
 | 
				
			||||||
 | 
					    {{ current_timestamp() }} as _airbyte_normalized_at
 | 
				
			||||||
 | 
					from {{ ref('se_orders_ab1') }}
 | 
				
			||||||
 | 
					-- se_orders
 | 
				
			||||||
 | 
					where 1 = 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,35 @@
 | 
				
			||||||
 | 
					{{ config(
 | 
				
			||||||
 | 
					    indexes = [{'columns':['_airbyte_emitted_at'],'type':'btree'}],
 | 
				
			||||||
 | 
					    unique_key = '_airbyte_ab_id',
 | 
				
			||||||
 | 
					    schema = "_airbyte_selly_express",
 | 
				
			||||||
 | 
					    tags = [ "top-level-intermediate" ]
 | 
				
			||||||
 | 
					) }}
 | 
				
			||||||
 | 
					-- SQL model to build a hash column based on the values of this record
 | 
				
			||||||
 | 
					-- depends_on: {{ ref('se_orders_ab2') }}
 | 
				
			||||||
 | 
					select
 | 
				
			||||||
 | 
					    {{ dbt_utils.surrogate_key([
 | 
				
			||||||
 | 
					        adapter.quote('to'),
 | 
				
			||||||
 | 
					        '_id',
 | 
				
			||||||
 | 
					        'cod',
 | 
				
			||||||
 | 
					        'code',
 | 
				
			||||||
 | 
					        adapter.quote('from'),
 | 
				
			||||||
 | 
					        'note',
 | 
				
			||||||
 | 
					        adapter.quote('value'),
 | 
				
			||||||
 | 
					        'client',
 | 
				
			||||||
 | 
					        'status',
 | 
				
			||||||
 | 
					        'volume',
 | 
				
			||||||
 | 
					        'weight',
 | 
				
			||||||
 | 
					        'courier',
 | 
				
			||||||
 | 
					        'distance',
 | 
				
			||||||
 | 
					        'createdat',
 | 
				
			||||||
 | 
					        'updatedat',
 | 
				
			||||||
 | 
					        'updatedby',
 | 
				
			||||||
 | 
					        'itemvolume',
 | 
				
			||||||
 | 
					        'searchstring',
 | 
				
			||||||
 | 
					        array_to_string('extraservices'),
 | 
				
			||||||
 | 
					    ]) }} as _airbyte_se_orders_hashid,
 | 
				
			||||||
 | 
					    tmp.*
 | 
				
			||||||
 | 
					from {{ ref('se_orders_ab2') }} tmp
 | 
				
			||||||
 | 
					-- se_orders
 | 
				
			||||||
 | 
					where 1 = 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,35 @@
 | 
				
			||||||
 | 
					{{ config(
 | 
				
			||||||
 | 
					    indexes = [{'columns':['_airbyte_emitted_at'],'type':'btree'}],
 | 
				
			||||||
 | 
					    unique_key = '_airbyte_ab_id',
 | 
				
			||||||
 | 
					    schema = "selly_express",
 | 
				
			||||||
 | 
					    tags = [ "top-level" ]
 | 
				
			||||||
 | 
					) }}
 | 
				
			||||||
 | 
					-- Final base SQL model
 | 
				
			||||||
 | 
					-- depends_on: {{ ref('se_orders_ab3') }}
 | 
				
			||||||
 | 
					select
 | 
				
			||||||
 | 
					    {{ adapter.quote('to') }},
 | 
				
			||||||
 | 
					    _id,
 | 
				
			||||||
 | 
					    {{ adapter.quote('client') }}::json->>'code' AS client_code,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('client') }}::json->'fee'->>'total' AS numeric) AS client_fee_total,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('client') }}::json->'fee'->>'shipping' AS numeric) AS client_fee_shipping,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('client') }}::json->'fee'->>'other' AS numeric) AS client_fee_other,
 | 
				
			||||||
 | 
					    {{ adapter.quote('courier') }}::json->>'code' AS courier_code,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('courier') }}::json->'fee'->>'total' AS numeric) AS courier_fee_total,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('courier') }}::json->'fee'->>'shipping' AS numeric) AS courier_fee_shipping,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('courier') }}::json->'fee'->>'other' AS numeric) AS courier_fee_other,
 | 
				
			||||||
 | 
					    courier,
 | 
				
			||||||
 | 
					    cast(weight AS numeric) AS weight,
 | 
				
			||||||
 | 
					    volume,
 | 
				
			||||||
 | 
					    code,
 | 
				
			||||||
 | 
					    status,
 | 
				
			||||||
 | 
					    cast({{ adapter.quote('value') }} AS numeric) AS value,
 | 
				
			||||||
 | 
					    createdat::timestamp AS created_at,
 | 
				
			||||||
 | 
					    updatedat::timestamp AS updated_at,
 | 
				
			||||||
 | 
					    _airbyte_ab_id,
 | 
				
			||||||
 | 
					    _airbyte_emitted_at,
 | 
				
			||||||
 | 
					    {{ current_timestamp() }} as _airbyte_normalized_at,
 | 
				
			||||||
 | 
					    _airbyte_se_orders_hashid
 | 
				
			||||||
 | 
					from {{ ref('se_orders_ab3') }}
 | 
				
			||||||
 | 
					-- se_orders from {{ source('selly_express', '_airbyte_raw_se_orders') }}
 | 
				
			||||||
 | 
					where 1 = 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,9 @@
 | 
				
			||||||
 | 
					version: 2
 | 
				
			||||||
 | 
					sources:
 | 
				
			||||||
 | 
					- name: selly_express
 | 
				
			||||||
 | 
					  quoting:
 | 
				
			||||||
 | 
					    database: true
 | 
				
			||||||
 | 
					    schema: false
 | 
				
			||||||
 | 
					    identifier: false
 | 
				
			||||||
 | 
					  tables:
 | 
				
			||||||
 | 
					  - name: _airbyte_raw_se_orders
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,5 @@
 | 
				
			||||||
 | 
					# add dependencies. these will get pulled during the `dbt deps` process.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					packages:
 | 
				
			||||||
 | 
					  - git: "https://github.com/fishtown-analytics/dbt-utils.git"
 | 
				
			||||||
 | 
					    revision: 0.7.4
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,17 @@
 | 
				
			||||||
 | 
					config:
 | 
				
			||||||
 | 
					  partial_parse: true
 | 
				
			||||||
 | 
					  printer_width: 120
 | 
				
			||||||
 | 
					  send_anonymous_usage_stats: false
 | 
				
			||||||
 | 
					  use_colors: true
 | 
				
			||||||
 | 
					normalize:
 | 
				
			||||||
 | 
					  outputs:
 | 
				
			||||||
 | 
					    prod:
 | 
				
			||||||
 | 
					      dbname: mongo_etl
 | 
				
			||||||
 | 
					      host: 18.140.112.89
 | 
				
			||||||
 | 
					      pass: bvxJaDGW2R55uyDXfODJ2a0Y
 | 
				
			||||||
 | 
					      port: 5432
 | 
				
			||||||
 | 
					      schema: public
 | 
				
			||||||
 | 
					      threads: 8
 | 
				
			||||||
 | 
					      type: postgres
 | 
				
			||||||
 | 
					      user: selly
 | 
				
			||||||
 | 
					  target: prod
 | 
				
			||||||
		Loading…
	
		Reference in New Issue