[120] | 1 | #Licensed to the Apache Software Foundation (ASF) under one |
---|
| 2 | #or more contributor license agreements. See the NOTICE file |
---|
| 3 | #distributed with this work for additional information |
---|
| 4 | #regarding copyright ownership. The ASF licenses this file |
---|
| 5 | #to you under the Apache License, Version 2.0 (the |
---|
| 6 | #"License"); you may not use this file except in compliance |
---|
| 7 | #with the License. You may obtain a copy of the License at |
---|
| 8 | |
---|
| 9 | # http://www.apache.org/licenses/LICENSE-2.0 |
---|
| 10 | |
---|
| 11 | #Unless required by applicable law or agreed to in writing, software |
---|
| 12 | #distributed under the License is distributed on an "AS IS" BASIS, |
---|
| 13 | #WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
---|
| 14 | #See the License for the specific language governing permissions and |
---|
| 15 | #limitations under the License. |
---|
| 16 | """define MapReduce as subclass of Service""" |
---|
| 17 | |
---|
| 18 | # -*- python -*- |
---|
| 19 | |
---|
| 20 | import os, copy, time |
---|
| 21 | |
---|
| 22 | from service import * |
---|
| 23 | from hodlib.Hod.nodePool import * |
---|
| 24 | from hodlib.Common.desc import CommandDesc |
---|
| 25 | from hodlib.Common.util import get_exception_string, parseEquals |
---|
| 26 | |
---|
| 27 | class MapReduceExternal(MasterSlave): |
---|
| 28 | """dummy proxy to external MapReduce instance""" |
---|
| 29 | |
---|
| 30 | def __init__(self, serviceDesc, workDirs, version): |
---|
| 31 | MasterSlave.__init__(self, serviceDesc, workDirs,None) |
---|
| 32 | self.launchedMaster = True |
---|
| 33 | self.masterInitialized = True |
---|
| 34 | self.version = version |
---|
| 35 | |
---|
| 36 | def getMasterRequest(self): |
---|
| 37 | return None |
---|
| 38 | |
---|
| 39 | def getMasterCommands(self, serviceDict): |
---|
| 40 | return [] |
---|
| 41 | |
---|
| 42 | def getAdminCommands(self, serviceDict): |
---|
| 43 | return [] |
---|
| 44 | |
---|
| 45 | def getWorkerCommands(self, serviceDict): |
---|
| 46 | return [] |
---|
| 47 | |
---|
| 48 | def getMasterAddrs(self): |
---|
| 49 | attrs = self.serviceDesc.getfinalAttrs() |
---|
| 50 | addr = attrs['mapred.job.tracker'] |
---|
| 51 | return [addr] |
---|
| 52 | |
---|
| 53 | def needsMore(self): |
---|
| 54 | return 0 |
---|
| 55 | |
---|
| 56 | def needsLess(self): |
---|
| 57 | return 0 |
---|
| 58 | |
---|
| 59 | def setMasterParams(self, dict): |
---|
| 60 | self.serviceDesc['final-attrs']['mapred.job.tracker'] = "%s:%s" % (dict['host'], |
---|
| 61 | dict['tracker_port']) |
---|
| 62 | |
---|
| 63 | if self.version < 16: |
---|
| 64 | self.serviceDesc.dict['final-attrs']['mapred.job.tracker.info.port'] = \ |
---|
| 65 | str(self.serviceDesc.dict['info_port']) |
---|
| 66 | else: |
---|
| 67 | # After Hadoop-2185 |
---|
| 68 | self.serviceDesc['final-attrs']['mapred.job.tracker.http.address'] = \ |
---|
| 69 | "%s:%s" %(dict['host'], dict['info_port']) |
---|
| 70 | |
---|
| 71 | def getInfoAddrs(self): |
---|
| 72 | attrs = self.serviceDesc.getfinalAttrs() |
---|
| 73 | if self.version < 16: |
---|
| 74 | addr = attrs['mapred.job.tracker'] |
---|
| 75 | k,v = addr.split( ":") |
---|
| 76 | infoaddr = k + ':' + attrs['mapred.job.tracker.info.port'] |
---|
| 77 | else: |
---|
| 78 | # After Hadoop-2185 |
---|
| 79 | # Note: earlier,we never respected mapred.job.tracker.http.address |
---|
| 80 | infoaddr = attrs['mapred.job.tracker.http.address'] |
---|
| 81 | return [infoaddr] |
---|
| 82 | |
---|
| 83 | class MapReduce(MasterSlave): |
---|
| 84 | |
---|
| 85 | def __init__(self, serviceDesc, workDirs,required_node, version, |
---|
| 86 | workers_per_ring = 1): |
---|
| 87 | MasterSlave.__init__(self, serviceDesc, workDirs,required_node) |
---|
| 88 | |
---|
| 89 | self.masterNode = None |
---|
| 90 | self.masterAddr = None |
---|
| 91 | self.infoAddr = None |
---|
| 92 | self.workers = [] |
---|
| 93 | self.required_node = required_node |
---|
| 94 | self.version = version |
---|
| 95 | self.workers_per_ring = workers_per_ring |
---|
| 96 | |
---|
| 97 | def isLaunchable(self, serviceDict): |
---|
| 98 | hdfs = serviceDict['hdfs'] |
---|
| 99 | if (hdfs.isMasterInitialized()): |
---|
| 100 | return True |
---|
| 101 | return False |
---|
| 102 | |
---|
| 103 | def getMasterRequest(self): |
---|
| 104 | req = NodeRequest(1, [], False) |
---|
| 105 | return req |
---|
| 106 | |
---|
| 107 | def getMasterCommands(self, serviceDict): |
---|
| 108 | |
---|
| 109 | hdfs = serviceDict['hdfs'] |
---|
| 110 | |
---|
| 111 | cmdDesc = self._getJobTrackerCommand(hdfs) |
---|
| 112 | return [cmdDesc] |
---|
| 113 | |
---|
| 114 | def getAdminCommands(self, serviceDict): |
---|
| 115 | return [] |
---|
| 116 | |
---|
| 117 | def getWorkerCommands(self, serviceDict): |
---|
| 118 | |
---|
| 119 | hdfs = serviceDict['hdfs'] |
---|
| 120 | |
---|
| 121 | workerCmds = [] |
---|
| 122 | for id in range(1, self.workers_per_ring + 1): |
---|
| 123 | workerCmds.append(self._getTaskTrackerCommand(str(id), hdfs)) |
---|
| 124 | |
---|
| 125 | return workerCmds |
---|
| 126 | |
---|
| 127 | def setMasterNodes(self, list): |
---|
| 128 | node = list[0] |
---|
| 129 | self.masterNode = node |
---|
| 130 | |
---|
| 131 | def getMasterAddrs(self): |
---|
| 132 | return [self.masterAddr] |
---|
| 133 | |
---|
| 134 | def getInfoAddrs(self): |
---|
| 135 | return [self.infoAddr] |
---|
| 136 | |
---|
| 137 | def getWorkers(self): |
---|
| 138 | return self.workers |
---|
| 139 | |
---|
| 140 | def requiredNode(self): |
---|
| 141 | return self.required_host |
---|
| 142 | |
---|
| 143 | def setMasterParams(self, list): |
---|
| 144 | dict = self._parseEquals(list) |
---|
| 145 | self.masterAddr = dict['mapred.job.tracker'] |
---|
| 146 | k,v = self.masterAddr.split(":") |
---|
| 147 | self.masterNode = k |
---|
| 148 | if self.version < 16: |
---|
| 149 | self.infoAddr = self.masterNode + ':' + dict['mapred.job.tracker.info.port'] |
---|
| 150 | else: |
---|
| 151 | # After Hadoop-2185 |
---|
| 152 | self.infoAddr = dict['mapred.job.tracker.http.address'] |
---|
| 153 | |
---|
| 154 | def _parseEquals(self, list): |
---|
| 155 | return parseEquals(list) |
---|
| 156 | |
---|
| 157 | def _setWorkDirs(self, workDirs, envs, attrs, parentDirs, subDir): |
---|
| 158 | local = [] |
---|
| 159 | system = None |
---|
| 160 | temp = None |
---|
| 161 | hadooptmpdir = None |
---|
| 162 | dfsclient = [] |
---|
| 163 | |
---|
| 164 | for p in parentDirs: |
---|
| 165 | workDirs.append(p) |
---|
| 166 | workDirs.append(os.path.join(p, subDir)) |
---|
| 167 | dir = os.path.join(p, subDir, 'mapred-local') |
---|
| 168 | local.append(dir) |
---|
| 169 | if not system: |
---|
| 170 | system = os.path.join(p, subDir, 'mapred-system') |
---|
| 171 | if not temp: |
---|
| 172 | temp = os.path.join(p, subDir, 'mapred-temp') |
---|
| 173 | if not hadooptmpdir: |
---|
| 174 | # Not used currently, generating hadooptmpdir just in case |
---|
| 175 | hadooptmpdir = os.path.join(p, subDir, 'hadoop-tmp') |
---|
| 176 | dfsclientdir = os.path.join(p, subDir, 'dfs-client') |
---|
| 177 | dfsclient.append(dfsclientdir) |
---|
| 178 | workDirs.append(dfsclientdir) |
---|
| 179 | # FIXME!! use csv |
---|
| 180 | attrs['mapred.local.dir'] = ','.join(local) |
---|
| 181 | attrs['mapred.system.dir'] = 'fillindir' |
---|
| 182 | attrs['mapred.temp.dir'] = temp |
---|
| 183 | attrs['hadoop.tmp.dir'] = hadooptmpdir |
---|
| 184 | |
---|
| 185 | |
---|
| 186 | envs['HADOOP_ROOT_LOGGER'] = "INFO,DRFA" |
---|
| 187 | |
---|
| 188 | |
---|
| 189 | def _getJobTrackerCommand(self, hdfs): |
---|
| 190 | sd = self.serviceDesc |
---|
| 191 | |
---|
| 192 | parentDirs = self.workDirs |
---|
| 193 | workDirs = [] |
---|
| 194 | attrs = sd.getfinalAttrs().copy() |
---|
| 195 | envs = sd.getEnvs().copy() |
---|
| 196 | |
---|
| 197 | if 'mapred.job.tracker' not in attrs: |
---|
| 198 | attrs['mapred.job.tracker'] = 'fillinhostport' |
---|
| 199 | |
---|
| 200 | if self.version < 16: |
---|
| 201 | if 'mapred.job.tracker.info.port' not in attrs: |
---|
| 202 | attrs['mapred.job.tracker.info.port'] = 'fillinport' |
---|
| 203 | else: |
---|
| 204 | # Addressing Hadoop-2185, |
---|
| 205 | if 'mapred.job.tracker.http.address' not in attrs: |
---|
| 206 | attrs['mapred.job.tracker.http.address'] = 'fillinhostport' |
---|
| 207 | |
---|
| 208 | attrs['fs.default.name'] = hdfs.getMasterAddrs()[0] |
---|
| 209 | |
---|
| 210 | self._setWorkDirs(workDirs, envs, attrs, parentDirs, 'mapred-jt') |
---|
| 211 | |
---|
| 212 | dict = { 'name' : 'jobtracker' } |
---|
| 213 | dict['version'] = self.version |
---|
| 214 | dict['program'] = os.path.join('bin', 'hadoop') |
---|
| 215 | dict['argv'] = ['jobtracker'] |
---|
| 216 | dict['envs'] = envs |
---|
| 217 | dict['pkgdirs'] = sd.getPkgDirs() |
---|
| 218 | dict['workdirs'] = workDirs |
---|
| 219 | dict['final-attrs'] = attrs |
---|
| 220 | dict['attrs'] = sd.getAttrs() |
---|
| 221 | cmd = CommandDesc(dict) |
---|
| 222 | return cmd |
---|
| 223 | |
---|
| 224 | def _getTaskTrackerCommand(self, id, hdfs): |
---|
| 225 | |
---|
| 226 | sd = self.serviceDesc |
---|
| 227 | |
---|
| 228 | parentDirs = self.workDirs |
---|
| 229 | workDirs = [] |
---|
| 230 | attrs = sd.getfinalAttrs().copy() |
---|
| 231 | envs = sd.getEnvs().copy() |
---|
| 232 | jt = self.masterAddr |
---|
| 233 | |
---|
| 234 | if jt == None: |
---|
| 235 | raise ValueError, "Can't get job tracker address" |
---|
| 236 | |
---|
| 237 | attrs['mapred.job.tracker'] = jt |
---|
| 238 | attrs['fs.default.name'] = hdfs.getMasterAddrs()[0] |
---|
| 239 | |
---|
| 240 | if self.version < 16: |
---|
| 241 | if 'tasktracker.http.port' not in attrs: |
---|
| 242 | attrs['tasktracker.http.port'] = 'fillinport' |
---|
| 243 | # earlier to 16, tasktrackers always took ephemeral port 0 for |
---|
| 244 | # tasktracker.report.bindAddress |
---|
| 245 | else: |
---|
| 246 | # Adding the following. Hadoop-2185 |
---|
| 247 | if 'mapred.task.tracker.report.address' not in attrs: |
---|
| 248 | attrs['mapred.task.tracker.report.address'] = 'fillinhostport' |
---|
| 249 | if 'mapred.task.tracker.http.address' not in attrs: |
---|
| 250 | attrs['mapred.task.tracker.http.address'] = 'fillinhostport' |
---|
| 251 | |
---|
| 252 | # unique parentDirs in case of multiple tasktrackers per hodring |
---|
| 253 | pd = [] |
---|
| 254 | for dir in parentDirs: |
---|
| 255 | dir = dir + "-" + id |
---|
| 256 | pd.append(dir) |
---|
| 257 | parentDirs = pd |
---|
| 258 | # end of unique workdirs |
---|
| 259 | |
---|
| 260 | self._setWorkDirs(workDirs, envs, attrs, parentDirs, 'mapred-tt') |
---|
| 261 | |
---|
| 262 | dict = { 'name' : 'tasktracker' } |
---|
| 263 | dict['program'] = os.path.join('bin', 'hadoop') |
---|
| 264 | dict['argv'] = ['tasktracker'] |
---|
| 265 | dict['envs'] = envs |
---|
| 266 | dict['pkgdirs'] = sd.getPkgDirs() |
---|
| 267 | dict['workdirs'] = workDirs |
---|
| 268 | dict['final-attrs'] = attrs |
---|
| 269 | dict['attrs'] = sd.getAttrs() |
---|
| 270 | cmd = CommandDesc(dict) |
---|
| 271 | return cmd |
---|
| 272 | |
---|