Here is the solution
sc.textFile("hdfs://nn1home:8020/input/war-and-peace.txt")
How did I find out nn1home:8020?
Just search for the file core-site.xml
and look for xml element fs.defaultFS
More Related Contents:
- Begenner at spark Big data programming (spark code)
- How to open/stream .zip files through Spark?
- Spark on yarn concept understanding
- Pyspark: get list of files/directories on HDFS path
- java.lang.NoClassDefFoundError: org/apache/hadoop/fs/StorageStatistics
- How to access s3a:// files from Apache Spark?
- Is it better to have one large parquet file or lots of smaller parquet files?
- Apache Spark: The number of cores vs. the number of executors
- Apache Hadoop Yarn – Underutilization of cores
- Where are logs in Spark on YARN?
- Easiest way to install Python dependencies on Spark executor nodes?
- Spark iterate HDFS directory
- How to turn off INFO logging in Spark?
- How does Hadoop Namenode failover process works?
- How can PySpark be called in debug mode?
- Sqoop import : composite primary key and textual primary key
- When do reduce tasks start in Hadoop?
- Hadoop java.io.IOException: Mkdirs failed to create /some/path
- Set hadoop system user for client embedded in Java webapp
- How can I access S3/S3n from a local Hadoop 2.6 installation?
- OOZIE: properties defined in file referenced in global job-xml not visible in workflow.xml
- How to get the input file name in the mapper in a Hadoop program?
- hadoop map reduce secondary sorting
- How can I force Spark to execute code?
- Parallel Algorithms for Generating Prime Numbers (possibly using Hadoop’s map reduce)
- Default Namenode port of HDFS is 50070.But I have come across at some places 8020 or 9000 [closed]
- Spark Scala list folders in directory
- Hadoop input split size vs block size
- Hadoop WordCount example stuck at map 100% reduce 0%
- hadoop: difference between 0 reducer and identity reducer?