Open wajda opened 2 years ago
I am happy to contribute to this.
@vishalag001
Let's start with creating a piece of code that for the given data source, finds a better initial name than just a URI suffix. Take a look at the ExecutionProducerRepositoryImpl.scala:56
There you have a parsed execution plan object with all the information including write operation and datasource URI. From that you need to create a set of unique DataSource
entities that will be stored into the database at the next step. URI is the ID, so it has to stay the same. Also pay attention on which write operation properties are deemed to be optional and which are required. You cannot expect the execution plan to always come from Spark or a Spline agent, so you can only rely on what's defined in the data model or, at the last resort, check the ExecutionPlan agentInfo
and systemInfo
properties to apply your logic on execution plans originated in a Spline Spark Agent, and keep the current logic for any other ones.
Discussed in https://github.com/AbsaOSS/spline/discussions/1093