AbsaOSS / enceladus

Dynamic Conformance Engine
Apache License 2.0
30 stars 14 forks source link

Create methods for managing temporary folders locally and in HDFS #998

Open yruslan opened 4 years ago

yruslan commented 4 years ago

Background

Currently, most of our tests that require access to files (datasets, mapping tables, etc), rely on paths relative to the current folder. Some of data files are generated on fly. That effectively means that we put temporary files inside project folders.

Creating a temporary directory and generating input test files and mapping tables there is not a big deal. getLocalTemporaryDirectory() method can be used for this. But our unit tests should run in distributed mode as well (See #317).

Feature

Need to develop the following methods:

Additional context

yruslan commented 4 years ago

Hadoop configuration provides the base path for temporary files (FS type independent):

spark.sparkContext.hadoopConfiguration.get("hadoop.tmp.dir")