The parameter mapred.max.split.size
which can be set per job individually is what you looking for. Don’t change dfs.block.size
because this is global for HDFS and can lead to problems.
More Related Contents:
- Calling a mapreduce job from a simple java program
- Oozie: Launch Map-Reduce from Oozie action?
- Hadoop DistributedCache is deprecated – what is the preferred API?
- Hadoop Word count: receive the total number of words that start with the letter “c”
- Reading file as single record in hadoop
- MultipleOutputFormat in hadoop
- hadoop java.net.URISyntaxException: Relative path in absolute URI: rsrc:hbase-common-0.98.1-hadoop2.jar
- Hadoop “Unable to load native-hadoop library for your platform” warning
- Hive: Best way to do incremetal updates on a main table
- Read whole text files from a compression in Spark
- hadoop No FileSystem for scheme: file
- Write a file in hdfs with Java
- Should I call ugi.checkTGTAndReloginFromKeytab() before every action on hadoop?
- What determines Kafka consumer offset?
- How to use Sqoop in Java Program?
- Hadoop cluster setup – java.net.ConnectException: Connection refused
- How to convert .txt file to Hadoop’s sequence file format
- Exception in thread “main” java.lang.UnsatisfiedLinkError: org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
- Iterate twice on values (MapReduce)
- Accessing a mapper’s counter from a reducer
- NotSerializableException on anonymous class
- How do I build/run this simple Mahout program without getting exceptions?
- Java8: HashMap to HashMap using Stream / Map-Reduce / Collector
- Is gzip format supported in Spark?
- Change output filename prefix for DataFrame.write()
- out of Memory Error in Hadoop
- How can I force Spark to execute code?
- How to list all files in a directory and its subdirectories in hadoop hdfs
- How to run a jar file in hadoop?
- Exception “: org.apache.hadoop.ipc.RpcException: RPC response exceeds maximum data length” from java