The answer is “not directly”. If you are able to load MongoDB/ArangoDB data into a Pandas or Spark dataframe, you can work with the data this way. See e.g. at https://docs.greatexpectations.io/en/latest/tutorials/create_expectations.html?highlight=batch_kwargs#load-a-batch-of-data-to-create-expectations - instead of providing a ‘path’ key, you would provide a ‘dataset’ key with the Pandas or Spark DF.
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
How to load data from S3 for validation with Pandas as a batch | 3 | 647 | June 24, 2020 | |
Configuring S3 as Datastore | 2 | 694 | April 1, 2021 | |
How to connect to data on Databricks Unity Catalog using Spark | 3 | 1071 | November 2, 2023 | |
I am currently working with Great Expectations Core to validate data from two different sources: a CSV file and a MongoDB data source. While I am able to create Expectations and generate local Data Docs, I am encountering the same issue in both cases. S | 1 | 139 | November 7, 2024 | |
Validate foreign keys / load multiple files to a single spark dataframe with batch generator | 1 | 751 | July 8, 2020 |