-
Notifications
You must be signed in to change notification settings - Fork 14
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Feature/1556 file access PoC using Hadoop FS API (#1586)
* s3 using hadoop fs api * s3 sdk usage removed (pom, classes, tests) * atum final version 3.1.0 used * readStandardizationInputData(... path: String)(implicit ... fs: FileSystem) -> readStandardizationInputData(input: PathWithFs)
- Loading branch information
Showing
47 changed files
with
423 additions
and
1,103 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -26,21 +26,20 @@ import za.co.absa.enceladus.dao.MenasDAO | |
import za.co.absa.enceladus.dao.auth.MenasKerberosCredentials | ||
import za.co.absa.enceladus.dao.rest.{MenasConnectionStringParser, RestDaoFactory} | ||
import za.co.absa.enceladus.model.Dataset | ||
import za.co.absa.enceladus.utils.fs.HdfsUtils | ||
import za.co.absa.enceladus.utils.testUtils.SparkTestBase | ||
import za.co.absa.enceladus.utils.fs.HadoopFsUtils | ||
import za.co.absa.enceladus.utils.testUtils.{HadoopFsTestBase, SparkTestBase} | ||
|
||
case class XPadTestInputRow(intField: Int, stringField: Option[String]) | ||
case class XPadTestOutputRow(intField: Int, stringField: Option[String], targetField: String) | ||
object XPadTestOutputRow { | ||
def apply(input: XPadTestInputRow, targetField: String): XPadTestOutputRow = XPadTestOutputRow(input.intField, input.stringField, targetField) | ||
} | ||
|
||
class LpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase with MockitoSugar { | ||
class LpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase with MockitoSugar with HadoopFsTestBase { | ||
import spark.implicits._ | ||
|
||
implicit val progArgs: ConformanceConfig = ConformanceConfig() // here we may need to specify some parameters (for certain rules) | ||
implicit val dao: MenasDAO = mock[MenasDAO] // you may have to hard-code your own implementation here (if not working with menas) | ||
implicit val fsUtils: HdfsUtils = new HdfsUtils(spark.sparkContext.hadoopConfiguration) | ||
|
||
val experimentalMR = true | ||
val isCatalystWorkaroundEnabled = true | ||
|
@@ -180,7 +179,7 @@ class LpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase with | |
} | ||
|
||
|
||
class RpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase { | ||
class RpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase with HadoopFsTestBase { | ||
|
||
import spark.implicits._ | ||
|
||
|
@@ -189,7 +188,6 @@ class RpadCustomConformanceRuleSuite extends AnyFunSuite with SparkTestBase { | |
private val meansCredentials = MenasKerberosCredentials("[email protected]", "src/test/resources/user.keytab.example") | ||
implicit val progArgs: ConformanceConfig = ConformanceConfig() // here we may need to specify some parameters (for certain rules) | ||
implicit val dao: MenasDAO = RestDaoFactory.getInstance(meansCredentials, menasBaseUrls) // you may have to hard-code your own implementation here (if not working with menas) | ||
implicit val fsUtils: HdfsUtils = new HdfsUtils(spark.sparkContext.hadoopConfiguration) | ||
|
||
val experimentalMR = true | ||
val isCatalystWorkaroundEnabled = true | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.