Skip to main content
Solved

Where are DQE rules applied when executing (DPE or data source)


Forum|alt.badge.img+2

We had a internal discussion, but ended up not being 100% sure. So here's the question.

When we run DQE on a CI, where are the rules evaluated? Does the DPE get the full data (possibly with filters on the MP?) and the run the rules evaluation IN the DPE. Or does the DPE create a SQL-like statement and does the source DB return the results?

As we also run on Hadoop with DPE on Hadoop. Is this the same.

Best answer by anna.spakova

Hello @Marnix Wisselaar , with Hadoop, if you use Spark processing, the DQE is done in Spark to have the best performance. The same if you would have Snowflake pushdown.

In the rest of the cases, the DPE will download the data from the source into temporary files stored on the DPE server, and do the DQE there. Once the job is finished, all temporary files are deleted.

I hope this answers your question.

Kind regards,

Anna

View original
Did this topic help you find an answer to your question?

2 replies

anna.spakova
Ataccamer
Forum|alt.badge.img+3
  • Ataccamer
  • 144 replies
  • Answer
  • December 18, 2023

Hello @Marnix Wisselaar , with Hadoop, if you use Spark processing, the DQE is done in Spark to have the best performance. The same if you would have Snowflake pushdown.

In the rest of the cases, the DPE will download the data from the source into temporary files stored on the DPE server, and do the DQE there. Once the job is finished, all temporary files are deleted.

I hope this answers your question.

Kind regards,

Anna


Cansu
Community Manager
Forum|alt.badge.img+3
  • Community Manager
  • 625 replies
  • December 22, 2023

Hi @Marnix Wisselaar - Iā€™m marking this thread as solved but please do let us know if you have any further questions. Happy holidays!


Reply


Cookie policy

We use cookies to enhance and personalize your experience. If you accept you agree to our full cookie policy. Learn more about our cookies.

 
Cookie settings