hi guys, I am trying to run a couple of suites on great expectations in Snowflake, but I have an issue with the SQL Alchemy dataset and I am not sure how to inject additional properties in the sql engine. While running a checkpoint I get QueuePool limit of size 5 overflow 10 reached, connection timed out, timeout 30
. I’ve tried adding pool_size
and max_overflow
in the batch_kargs
of the checkpoint but it does not remove the issue. Can anyone give some guidance here? Many thanks!
I am using great expectations 0.13.7 and I am using the command line API, not sure if it’s v2 or v3.
An example of a batch I am attempting to do:
- batch_kwargs:
table: test
schema: t
data_asset_name: t.test
datasource: data
pool_size: 100
max_overflow: 200
expectation_suite_names:
- t.test.warning