Ok, and your container settings?
On Wed, Jan 31, 2018 at 02:38 nishchay malhotra <[email protected]>
wrote:
> yes my job has about 160,000 maps and my cluster not getting fully
> utilized around 6000 maps ran for 2 hrs and then I killed the job. At any
> point of time only 40 containers are running thats just 11% of my cluster
> capacity.
>
> {
> "classification": "mapred-site",
> "properties": {
> "mapreduce.job.reduce.slowstart.completedmaps":"1",
> "mapreduce.reduce.memory.mb": "3072",
> "mapreduce.map.memory.mb": "2208",
> "mapreduce.map.java.opts":"-Xmx1800m",
> "mapreduce.map.cpu.vcores":"1"
> }
> },
> {
> "classification": "yarn-site",
> "properties": {
> "yarn.scheduler.minimum-allocation-mb": "32”,
> “yarn.scheduler.maximum-allocation-mb”:”253952”,
> “yarn.scheduler.maximum-allocation-vcores: “128”
>
> "yarn.nodemanager.vmem-pmem-ratio":"3",
> "yarn.nodemanager.vmem-check-enabled":"true",
> yarn.nodemanager.resource.cpu-vcores" ; "16”,
> yarn.nodemanager.resource.memory-mb: “23040"
> }
>
> Each node: capacity
> Disk-space=100gb
> memory=28gb
> processors: 8
>
>
> --
William Watson