forked from saltstack-formulas/hadoop-formula
-
Notifications
You must be signed in to change notification settings - Fork 0
/
pillar.example
119 lines (115 loc) · 4.08 KB
/
pillar.example
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
# java_home: /usr/lib/java
hadoop:
version: apache-1.2.1 # ['apache-1.2.1', 'apache-2.2.0', 'hdp-1.3.0', 'hdp-2.2.0', 'cdh-4.5.0', 'cdh-4.5.0-mr1']
targeting_method: grain # [compound, glob] also supported
ssh_key_type: dsa
ssh_key_source: files # pillar looks in hadoop.users.public_keys and hadoop.users.private_keys for <ssh-key_type>-<username>
users:
hadoop: 6000
hdfs: 6001
mapred: 6002
yarn: 6003
config:
directory: /etc/hadoop/conf
log_level: INFO
core-site:
io.native.lib.available:
value: true
io.file.buffer.size:
value: 65536
fs.trash.interval:
value: 60
hdfs:
namenode_target: "roles:hadoop_master" # Specify compound matching string to match all your namenodes
datanode_target: "roles:hadoop_slave" # Specify compound matching string to match all your datanodes e.g. if you were to use pillar I@datanode:true
config:
log_level: INFO
namenode_port: 8020
namenode_http_port: 50070
secondarynamenode_http_port: 50090
# the number of hdfs replicas is normally auto-configured for you in hdfs.settings
# according to the number of available datanodes
# replication: 1
hdfs-site:
dfs.permission:
value: false
dfs.durable.sync:
value: true
dfs.datanode.synconclose:
value: true
mapred:
jobtracker_target: "roles:hadoop_master"
tasktracker_target: "roles:hadoop_slave"
config:
log_level: INFO
jobtracker_port: 9001
jobtracker_http_port: 50030
jobhistory_port: 10020
jobhistory_webapp_port: 19888
history_dir: /mr-history
mapred-site:
mapred.map.memory.mb:
value: 1536
mapred.map.java.opts:
value: -Xmx1024M
mapred.reduce.memory.mb:
value: 3072
mapred.reduce.java.opts:
value: -Xmx1024m
mapred.task.io.sort.mb:
value: 512
mapred.task.io.sort.factor:
value: 100
mapred.reduce.shuffle.parallelcopies:
value: 50
# you only have to configure the capacity-scheduler section to the extent you need it - if omitted the
# resulting file (HADOOP_CONF/capacity-scheduler.xml) will just remain empty (no defaults)
yarn:
resourcemanager_target: "roles:hadoop_master"
nodemanager_target: "roles:hadoop_slave"
config:
log_level: INFO
yarn-site:
yarn.nodemanager.aux-services:
value: mapreduce_shuffle
yarn.nodemanager.aux-services.mapreduce.shuffle.class:
value: org.apache.hadoop.mapred.ShuffleHandler
yarn.resourcemanager.scheduler.class:
value: org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler
capacity-scheduler:
yarn.scheduler.capacity.maximum-applications:
value: 10000
yarn.scheduler.capacity.resource-calculator:
value: org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator
yarn.scheduler.capacity.root.queues:
value: default,customqueue
yarn.scheduler.capacity.root.capacity:
value: 100
yarn.scheduler.capacity.root.default.capacity:
value: 70
yarn.scheduler.capacity.root.default.user-limit-factor:
value: 1
yarn.scheduler.capacity.root.default.maximum-capacity:
value: 100
yarn.scheduler.capacity.root.default.state:
value: RUNNING
yarn.scheduler.capacity.root.default.acl_submit_applications:
value: '*'
yarn.scheduler.capacity.root.default.acl_administer_queue:
value: '*'
yarn.scheduler.capacity.root.customqueue.capacity:
value: 30
yarn.scheduler.capacity.root.customqueue.maximum-am-resource-percent:
value: 1.0
yarn.scheduler.capacity.root.customqueue.user-limit-factor:
value: 1
yarn.scheduler.capacity.root.customqueue.maximum-capacity:
value: 100
yarn.scheduler.capacity.root.customqueue.state:
value: RUNNING
yarn.scheduler.capacity.root.customqueue.acl_submit_applications:
value: '*'
yarn.scheduler.capacity.root.customqueue.acl_administer_queue:
value: '*'
yarn.scheduler.capacity.node-locality-delay:
value: -1