When you want to read data which is present in sub directories then Spark prompts below error -
Caused by: java.io.IOException: Not a file: hdfs://sdatalakedev/a/b/c=dinesh/subDir1Solution is to set below property - set mapreduce.input.fileinputformat.input.dir.recursive=true
Comments
Post a Comment