hi guys, I am trying to run a couple of suites on great expectations in Snowflake, but I have an issue with the SQL Alchemy dataset and I am not sure how to inject additional properties in the sql engine. While running a checkpoint I get
QueuePool limit of size 5 overflow 10 reached, connection timed out, timeout 30 . I’ve tried adding
max_overflow in the
batch_kargs of the checkpoint but it does not remove the issue. Can anyone give some guidance here? Many thanks!
I am using great expectations 0.13.7 and I am using the command line API, not sure if it’s v2 or v3.
An example of a batch I am attempting to do:
- batch_kwargs: table: test schema: t data_asset_name: t.test datasource: data pool_size: 100 max_overflow: 200 expectation_suite_names: - t.test.warning