001/* 002 * Licensed to the Apache Software Foundation (ASF) under one 003 * or more contributor license agreements. See the NOTICE file 004 * distributed with this work for additional information 005 * regarding copyright ownership. The ASF licenses this file 006 * to you under the Apache License, Version 2.0 (the 007 * "License"); you may not use this file except in compliance 008 * with the License. You may obtain a copy of the License at 009 * 010 * http://www.apache.org/licenses/LICENSE-2.0 011 * 012 * Unless required by applicable law or agreed to in writing, software 013 * distributed under the License is distributed on an "AS IS" BASIS, 014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ 018package org.apache.hadoop.hbase.coordination; 019 020import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.CHECK; 021import static org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective.FORCE; 022import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.DELETED; 023import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.FAILURE; 024import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.IN_PROGRESS; 025import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.SUCCESS; 026import static org.apache.hadoop.hbase.util.ConcurrentMapUtils.computeIfAbsent; 027 028import java.io.IOException; 029import java.util.List; 030import org.apache.hadoop.conf.Configuration; 031import org.apache.hadoop.hbase.HConstants; 032import org.apache.hadoop.hbase.ServerName; 033import org.apache.hadoop.hbase.SplitLogCounters; 034import org.apache.hadoop.hbase.SplitLogTask; 035import org.apache.hadoop.hbase.coordination.ZKSplitLogManagerCoordination.TaskFinisher.Status; 036import org.apache.hadoop.hbase.exceptions.DeserializationException; 037import org.apache.hadoop.hbase.log.HBaseMarkers; 038import org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective; 039import org.apache.hadoop.hbase.master.SplitLogManager.Task; 040import org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus; 041import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; 042import org.apache.hadoop.hbase.wal.WALSplitUtil; 043import org.apache.hadoop.hbase.zookeeper.ZKListener; 044import org.apache.hadoop.hbase.zookeeper.ZKMetadata; 045import org.apache.hadoop.hbase.zookeeper.ZKSplitLog; 046import org.apache.hadoop.hbase.zookeeper.ZKUtil; 047import org.apache.hadoop.hbase.zookeeper.ZKWatcher; 048import org.apache.hadoop.hbase.zookeeper.ZNodePaths; 049import org.apache.hadoop.util.StringUtils; 050import org.apache.yetus.audience.InterfaceAudience; 051import org.apache.zookeeper.AsyncCallback; 052import org.apache.zookeeper.CreateMode; 053import org.apache.zookeeper.KeeperException; 054import org.apache.zookeeper.KeeperException.NoNodeException; 055import org.apache.zookeeper.ZooDefs.Ids; 056import org.apache.zookeeper.data.Stat; 057import org.slf4j.Logger; 058import org.slf4j.LoggerFactory; 059 060/** 061 * ZooKeeper based implementation of {@link SplitLogManagerCoordination} 062 */ 063@InterfaceAudience.Private 064public class ZKSplitLogManagerCoordination extends ZKListener 065 implements SplitLogManagerCoordination { 066 067 public static final int DEFAULT_TIMEOUT = 120000; 068 public static final int DEFAULT_ZK_RETRIES = 3; 069 public static final int DEFAULT_MAX_RESUBMIT = 3; 070 071 private static final Logger LOG = LoggerFactory.getLogger(SplitLogManagerCoordination.class); 072 073 private final TaskFinisher taskFinisher; 074 private final Configuration conf; 075 076 private long zkretries; 077 private long resubmitThreshold; 078 private long timeout; 079 080 SplitLogManagerDetails details; 081 082 public boolean ignoreZKDeleteForTesting = false; 083 084 public ZKSplitLogManagerCoordination(Configuration conf, ZKWatcher watcher) { 085 super(watcher); 086 this.conf = conf; 087 taskFinisher = new TaskFinisher() { 088 @Override 089 public Status finish(ServerName workerName, String logfile) { 090 try { 091 WALSplitUtil.finishSplitLogFile(logfile, conf); 092 } catch (IOException e) { 093 LOG.warn("Could not finish splitting of log file " + logfile, e); 094 return Status.ERR; 095 } 096 return Status.DONE; 097 } 098 }; 099 } 100 101 @Override 102 public void init() throws IOException { 103 this.zkretries = conf.getLong("hbase.splitlog.zk.retries", DEFAULT_ZK_RETRIES); 104 this.resubmitThreshold = conf.getLong("hbase.splitlog.max.resubmit", DEFAULT_MAX_RESUBMIT); 105 this.timeout = conf.getInt(HConstants.HBASE_SPLITLOG_MANAGER_TIMEOUT, DEFAULT_TIMEOUT); 106 if (this.watcher != null) { 107 this.watcher.registerListener(this); 108 lookForOrphans(); 109 } 110 } 111 112 @Override 113 public String prepareTask(String taskname) { 114 return ZKSplitLog.getEncodedNodeName(watcher, taskname); 115 } 116 117 @Override 118 public int remainingTasksInCoordination() { 119 int count = 0; 120 try { 121 List<String> tasks = 122 ZKUtil.listChildrenNoWatch(watcher, watcher.getZNodePaths().splitLogZNode); 123 if (tasks != null) { 124 int listSize = tasks.size(); 125 for (int i = 0; i < listSize; i++) { 126 if (!ZKSplitLog.isRescanNode(tasks.get(i))) { 127 count++; 128 } 129 } 130 } 131 } catch (KeeperException ke) { 132 LOG.warn("Failed to check remaining tasks", ke); 133 count = -1; 134 } 135 return count; 136 } 137 138 /** 139 * It is possible for a task to stay in UNASSIGNED state indefinitely - say SplitLogManager wants 140 * to resubmit a task. It forces the task to UNASSIGNED state but it dies before it could create 141 * the RESCAN task node to signal the SplitLogWorkers to pick up the task. To prevent this 142 * scenario the SplitLogManager resubmits all orphan and UNASSIGNED tasks at startup. n 143 */ 144 private void handleUnassignedTask(String path) { 145 if (ZKSplitLog.isRescanNode(watcher, path)) { 146 return; 147 } 148 Task task = findOrCreateOrphanTask(path); 149 if (task.isOrphan() && (task.incarnation.get() == 0)) { 150 LOG.info("Resubmitting unassigned orphan task " + path); 151 // ignore failure to resubmit. The timeout-monitor will handle it later 152 // albeit in a more crude fashion 153 resubmitTask(path, task, FORCE); 154 } 155 } 156 157 @Override 158 public void deleteTask(String path) { 159 deleteNode(path, zkretries); 160 } 161 162 @Override 163 public boolean resubmitTask(String path, Task task, ResubmitDirective directive) { 164 // its ok if this thread misses the update to task.deleted. It will fail later 165 if (task.status != IN_PROGRESS) { 166 return false; 167 } 168 int version; 169 if (directive != FORCE) { 170 // We're going to resubmit: 171 // 1) immediately if the worker server is now marked as dead 172 // 2) after a configurable timeout if the server is not marked as dead but has still not 173 // finished the task. This allows to continue if the worker cannot actually handle it, 174 // for any reason. 175 final long time = EnvironmentEdgeManager.currentTime() - task.last_update; 176 final boolean alive = details.getMaster().getServerManager() != null 177 ? details.getMaster().getServerManager().isServerOnline(task.cur_worker_name) 178 : true; 179 if (alive && time < timeout) { 180 LOG.trace("Skipping the resubmit of " + task.toString() + " because the server " 181 + task.cur_worker_name + " is not marked as dead, we waited for " + time 182 + " while the timeout is " + timeout); 183 return false; 184 } 185 186 if (task.unforcedResubmits.get() >= resubmitThreshold) { 187 if (!task.resubmitThresholdReached) { 188 task.resubmitThresholdReached = true; 189 SplitLogCounters.tot_mgr_resubmit_threshold_reached.increment(); 190 LOG.info("Skipping resubmissions of task " + path + " because threshold " 191 + resubmitThreshold + " reached"); 192 } 193 return false; 194 } 195 // race with heartbeat() that might be changing last_version 196 version = task.last_version; 197 } else { 198 SplitLogCounters.tot_mgr_resubmit_force.increment(); 199 version = -1; 200 } 201 LOG.info("Resubmitting task " + path); 202 task.incarnation.incrementAndGet(); 203 boolean result = resubmit(path, version); 204 if (!result) { 205 task.heartbeatNoDetails(EnvironmentEdgeManager.currentTime()); 206 return false; 207 } 208 // don't count forced resubmits 209 if (directive != FORCE) { 210 task.unforcedResubmits.incrementAndGet(); 211 } 212 task.setUnassigned(); 213 rescan(Long.MAX_VALUE); 214 SplitLogCounters.tot_mgr_resubmit.increment(); 215 return true; 216 } 217 218 @Override 219 public void checkTasks() { 220 rescan(Long.MAX_VALUE); 221 }; 222 223 /** 224 * signal the workers that a task was resubmitted by creating the RESCAN node. 225 */ 226 private void rescan(long retries) { 227 // The RESCAN node will be deleted almost immediately by the 228 // SplitLogManager as soon as it is created because it is being 229 // created in the DONE state. This behavior prevents a buildup 230 // of RESCAN nodes. But there is also a chance that a SplitLogWorker 231 // might miss the watch-trigger that creation of RESCAN node provides. 232 // Since the TimeoutMonitor will keep resubmitting UNASSIGNED tasks 233 // therefore this behavior is safe. 234 SplitLogTask slt = new SplitLogTask.Done(this.details.getServerName()); 235 this.watcher.getRecoverableZooKeeper().getZooKeeper().create(ZKSplitLog.getRescanNode(watcher), 236 slt.toByteArray(), Ids.OPEN_ACL_UNSAFE, CreateMode.EPHEMERAL_SEQUENTIAL, 237 new CreateRescanAsyncCallback(), Long.valueOf(retries)); 238 } 239 240 @Override 241 public void submitTask(String path) { 242 createNode(path, zkretries); 243 } 244 245 @Override 246 public void checkTaskStillAvailable(String path) { 247 // A negative retry count will lead to ignoring all error processing. 248 this.watcher.getRecoverableZooKeeper().getZooKeeper().getData(path, this.watcher, 249 new GetDataAsyncCallback(), Long.valueOf(-1) /* retry count */); 250 SplitLogCounters.tot_mgr_get_data_queued.increment(); 251 } 252 253 private void deleteNode(String path, Long retries) { 254 SplitLogCounters.tot_mgr_node_delete_queued.increment(); 255 // Once a task znode is ready for delete, that is it is in the TASK_DONE 256 // state, then no one should be writing to it anymore. That is no one 257 // will be updating the znode version any more. 258 this.watcher.getRecoverableZooKeeper().getZooKeeper().delete(path, -1, 259 new DeleteAsyncCallback(), retries); 260 } 261 262 private void deleteNodeSuccess(String path) { 263 if (ignoreZKDeleteForTesting) { 264 return; 265 } 266 Task task; 267 task = details.getTasks().remove(path); 268 if (task == null) { 269 if (ZKSplitLog.isRescanNode(watcher, path)) { 270 SplitLogCounters.tot_mgr_rescan_deleted.increment(); 271 } 272 SplitLogCounters.tot_mgr_missing_state_in_delete.increment(); 273 LOG.debug("Deleted task without in memory state " + path); 274 return; 275 } 276 synchronized (task) { 277 task.status = DELETED; 278 task.notify(); 279 } 280 SplitLogCounters.tot_mgr_task_deleted.increment(); 281 } 282 283 private void deleteNodeFailure(String path) { 284 LOG.info("Failed to delete node " + path + " and will retry soon."); 285 return; 286 } 287 288 private void createRescanSuccess(String path) { 289 SplitLogCounters.tot_mgr_rescan.increment(); 290 getDataSetWatch(path, zkretries); 291 } 292 293 private void createRescanFailure() { 294 LOG.error(HBaseMarkers.FATAL, "logic failure, rescan failure must not happen"); 295 } 296 297 /** 298 * Helper function to check whether to abandon retries in ZooKeeper AsyncCallback functions 299 * @param statusCode integer value of a ZooKeeper exception code 300 * @param action description message about the retried action 301 * @return true when need to abandon retries otherwise false 302 */ 303 private boolean needAbandonRetries(int statusCode, String action) { 304 if (statusCode == KeeperException.Code.SESSIONEXPIRED.intValue()) { 305 LOG.error("ZK session expired. Master is expected to shut down. Abandoning retries for " 306 + "action=" + action); 307 return true; 308 } 309 return false; 310 } 311 312 private void createNode(String path, Long retry_count) { 313 SplitLogTask slt = new SplitLogTask.Unassigned(details.getServerName()); 314 ZKUtil.asyncCreate(this.watcher, path, slt.toByteArray(), new CreateAsyncCallback(), 315 retry_count); 316 SplitLogCounters.tot_mgr_node_create_queued.increment(); 317 return; 318 } 319 320 private void createNodeSuccess(String path) { 321 LOG.debug("Put up splitlog task at znode " + path); 322 getDataSetWatch(path, zkretries); 323 } 324 325 private void createNodeFailure(String path) { 326 // TODO the Manager should split the log locally instead of giving up 327 LOG.warn("Failed to create task node " + path); 328 setDone(path, FAILURE); 329 } 330 331 private void getDataSetWatch(String path, Long retry_count) { 332 this.watcher.getRecoverableZooKeeper().getZooKeeper().getData(path, this.watcher, 333 new GetDataAsyncCallback(), retry_count); 334 SplitLogCounters.tot_mgr_get_data_queued.increment(); 335 } 336 337 private void getDataSetWatchSuccess(String path, byte[] data, int version) 338 throws DeserializationException { 339 if (data == null) { 340 if (version == Integer.MIN_VALUE) { 341 // assume all done. The task znode suddenly disappeared. 342 setDone(path, SUCCESS); 343 return; 344 } 345 SplitLogCounters.tot_mgr_null_data.increment(); 346 LOG.error(HBaseMarkers.FATAL, "logic error - got null data " + path); 347 setDone(path, FAILURE); 348 return; 349 } 350 data = ZKMetadata.removeMetaData(data); 351 SplitLogTask slt = SplitLogTask.parseFrom(data); 352 if (slt.isUnassigned()) { 353 LOG.debug("Task not yet acquired " + path + ", ver=" + version); 354 handleUnassignedTask(path); 355 } else if (slt.isOwned()) { 356 heartbeat(path, version, slt.getServerName()); 357 } else if (slt.isResigned()) { 358 LOG.info("Task " + path + " entered state=" + slt.toString()); 359 resubmitOrFail(path, FORCE); 360 } else if (slt.isDone()) { 361 LOG.info("Task " + path + " entered state=" + slt.toString()); 362 if (taskFinisher != null && !ZKSplitLog.isRescanNode(watcher, path)) { 363 if (taskFinisher.finish(slt.getServerName(), ZKSplitLog.getFileName(path)) == Status.DONE) { 364 setDone(path, SUCCESS); 365 } else { 366 resubmitOrFail(path, CHECK); 367 } 368 } else { 369 setDone(path, SUCCESS); 370 } 371 } else if (slt.isErr()) { 372 LOG.info("Task " + path + " entered state=" + slt.toString()); 373 resubmitOrFail(path, CHECK); 374 } else { 375 LOG.error(HBaseMarkers.FATAL, 376 "logic error - unexpected zk state for path = " + path + " data = " + slt.toString()); 377 setDone(path, FAILURE); 378 } 379 } 380 381 private void resubmitOrFail(String path, ResubmitDirective directive) { 382 if (resubmitTask(path, findOrCreateOrphanTask(path), directive) == false) { 383 setDone(path, FAILURE); 384 } 385 } 386 387 private void getDataSetWatchFailure(String path) { 388 LOG.warn("Failed to set data watch " + path); 389 setDone(path, FAILURE); 390 } 391 392 private void setDone(String path, TerminationStatus status) { 393 Task task = details.getTasks().get(path); 394 if (task == null) { 395 if (!ZKSplitLog.isRescanNode(watcher, path)) { 396 SplitLogCounters.tot_mgr_unacquired_orphan_done.increment(); 397 LOG.debug("Unacquired orphan task is done " + path); 398 } 399 } else { 400 synchronized (task) { 401 if (task.status == IN_PROGRESS) { 402 if (status == SUCCESS) { 403 SplitLogCounters.tot_mgr_log_split_success.increment(); 404 LOG.info("Done splitting " + path); 405 } else { 406 SplitLogCounters.tot_mgr_log_split_err.increment(); 407 LOG.warn("Error splitting " + path); 408 } 409 task.status = status; 410 if (task.batch != null) { 411 synchronized (task.batch) { 412 if (status == SUCCESS) { 413 task.batch.done++; 414 } else { 415 task.batch.error++; 416 } 417 task.batch.notify(); 418 } 419 } 420 } 421 } 422 } 423 // delete the task node in zk. It's an async 424 // call and no one is blocked waiting for this node to be deleted. All 425 // task names are unique (log.<timestamp>) there is no risk of deleting 426 // a future task. 427 // if a deletion fails, TimeoutMonitor will retry the same deletion later 428 deleteNode(path, zkretries); 429 return; 430 } 431 432 private Task findOrCreateOrphanTask(String path) { 433 return computeIfAbsent(details.getTasks(), path, Task::new, () -> { 434 LOG.info("Creating orphan task " + path); 435 SplitLogCounters.tot_mgr_orphan_task_acquired.increment(); 436 }); 437 } 438 439 private void heartbeat(String path, int new_version, ServerName workerName) { 440 Task task = findOrCreateOrphanTask(path); 441 if (new_version != task.last_version) { 442 if (task.isUnassigned()) { 443 LOG.info("Task " + path + " acquired by " + workerName); 444 } 445 task.heartbeat(EnvironmentEdgeManager.currentTime(), new_version, workerName); 446 SplitLogCounters.tot_mgr_heartbeat.increment(); 447 } else { 448 // duplicate heartbeats - heartbeats w/o zk node version 449 // changing - are possible. The timeout thread does 450 // getDataSetWatch() just to check whether a node still 451 // exists or not 452 } 453 return; 454 } 455 456 private void lookForOrphans() { 457 List<String> orphans; 458 try { 459 orphans = 460 ZKUtil.listChildrenNoWatch(this.watcher, this.watcher.getZNodePaths().splitLogZNode); 461 if (orphans == null) { 462 LOG.warn("Could not get children of " + this.watcher.getZNodePaths().splitLogZNode); 463 return; 464 } 465 } catch (KeeperException e) { 466 LOG.warn("Could not get children of " + this.watcher.getZNodePaths().splitLogZNode + " " 467 + StringUtils.stringifyException(e)); 468 return; 469 } 470 int rescan_nodes = 0; 471 int listSize = orphans.size(); 472 for (int i = 0; i < listSize; i++) { 473 String path = orphans.get(i); 474 String nodepath = ZNodePaths.joinZNode(watcher.getZNodePaths().splitLogZNode, path); 475 if (ZKSplitLog.isRescanNode(watcher, nodepath)) { 476 rescan_nodes++; 477 LOG.debug("Found orphan rescan node " + path); 478 } else { 479 LOG.info("Found orphan task " + path); 480 } 481 getDataSetWatch(nodepath, zkretries); 482 } 483 LOG.info("Found " + (orphans.size() - rescan_nodes) + " orphan tasks and " + rescan_nodes 484 + " rescan nodes"); 485 } 486 487 @Override 488 public void nodeDataChanged(String path) { 489 Task task; 490 task = details.getTasks().get(path); 491 if (task != null || ZKSplitLog.isRescanNode(watcher, path)) { 492 if (task != null) { 493 task.heartbeatNoDetails(EnvironmentEdgeManager.currentTime()); 494 } 495 getDataSetWatch(path, zkretries); 496 } 497 } 498 499 private boolean resubmit(String path, int version) { 500 try { 501 // blocking zk call but this is done from the timeout thread 502 SplitLogTask slt = new SplitLogTask.Unassigned(this.details.getServerName()); 503 if (ZKUtil.setData(this.watcher, path, slt.toByteArray(), version) == false) { 504 LOG.debug("Failed to resubmit task " + path + " version changed"); 505 return false; 506 } 507 } catch (NoNodeException e) { 508 LOG.warn("Failed to resubmit because znode doesn't exist " + path 509 + " task done (or forced done by removing the znode)"); 510 try { 511 getDataSetWatchSuccess(path, null, Integer.MIN_VALUE); 512 } catch (DeserializationException e1) { 513 LOG.debug("Failed to re-resubmit task " + path + " because of deserialization issue", e1); 514 return false; 515 } 516 return false; 517 } catch (KeeperException.BadVersionException e) { 518 LOG.debug("Failed to resubmit task " + path + " version changed"); 519 return false; 520 } catch (KeeperException e) { 521 SplitLogCounters.tot_mgr_resubmit_failed.increment(); 522 LOG.warn("Failed to resubmit " + path, e); 523 return false; 524 } 525 return true; 526 } 527 528 /** 529 * {@link org.apache.hadoop.hbase.master.SplitLogManager} can use objects implementing this 530 * interface to finish off a partially done task by 531 * {@link org.apache.hadoop.hbase.regionserver.SplitLogWorker}. This provides a serialization 532 * point at the end of the task processing. Must be restartable and idempotent. 533 */ 534 public interface TaskFinisher { 535 /** 536 * status that can be returned finish() 537 */ 538 enum Status { 539 /** 540 * task completed successfully 541 */ 542 DONE(), 543 /** 544 * task completed with error 545 */ 546 ERR(); 547 } 548 549 /** 550 * finish the partially done task. workername provides clue to where the partial results of the 551 * partially done tasks are present. taskname is the name of the task that was put up in 552 * zookeeper. 553 * <p> 554 * nn * @return DONE if task completed successfully, ERR otherwise 555 */ 556 Status finish(ServerName workerName, String taskname); 557 } 558 559 /** 560 * Asynchronous handler for zk create node results. Retries on failures. 561 */ 562 public class CreateAsyncCallback implements AsyncCallback.StringCallback { 563 private final Logger LOG = LoggerFactory.getLogger(CreateAsyncCallback.class); 564 565 @Override 566 public void processResult(int rc, String path, Object ctx, String name) { 567 SplitLogCounters.tot_mgr_node_create_result.increment(); 568 if (rc != 0) { 569 if (needAbandonRetries(rc, "Create znode " + path)) { 570 createNodeFailure(path); 571 return; 572 } 573 if (rc == KeeperException.Code.NODEEXISTS.intValue()) { 574 // What if there is a delete pending against this pre-existing 575 // znode? Then this soon-to-be-deleted task znode must be in TASK_DONE 576 // state. Only operations that will be carried out on this node by 577 // this manager are get-znode-data, task-finisher and delete-znode. 578 // And all code pieces correctly handle the case of suddenly 579 // disappearing task-znode. 580 LOG.debug("Found pre-existing znode " + path); 581 SplitLogCounters.tot_mgr_node_already_exists.increment(); 582 } else { 583 Long retry_count = (Long) ctx; 584 LOG.warn("Create rc=" + KeeperException.Code.get(rc) + " for " + path 585 + " remaining retries=" + retry_count); 586 if (retry_count == 0) { 587 SplitLogCounters.tot_mgr_node_create_err.increment(); 588 createNodeFailure(path); 589 } else { 590 SplitLogCounters.tot_mgr_node_create_retry.increment(); 591 createNode(path, retry_count - 1); 592 } 593 return; 594 } 595 } 596 createNodeSuccess(path); 597 } 598 } 599 600 /** 601 * Asynchronous handler for zk get-data-set-watch on node results. Retries on failures. 602 */ 603 public class GetDataAsyncCallback implements AsyncCallback.DataCallback { 604 private final Logger LOG = LoggerFactory.getLogger(GetDataAsyncCallback.class); 605 606 @Override 607 public void processResult(int rc, String path, Object ctx, byte[] data, Stat stat) { 608 SplitLogCounters.tot_mgr_get_data_result.increment(); 609 if (rc != 0) { 610 if (needAbandonRetries(rc, "GetData from znode " + path)) { 611 return; 612 } 613 if (rc == KeeperException.Code.NONODE.intValue()) { 614 SplitLogCounters.tot_mgr_get_data_nonode.increment(); 615 LOG.warn("Task znode " + path + " vanished or not created yet."); 616 // ignore since we should not end up in a case where there is in-memory task, 617 // but no znode. The only case is between the time task is created in-memory 618 // and the znode is created. See HBASE-11217. 619 return; 620 } 621 Long retry_count = (Long) ctx; 622 623 if (retry_count < 0) { 624 LOG.warn("Getdata rc=" + KeeperException.Code.get(rc) + " " + path 625 + ". Ignoring error. No error handling. No retrying."); 626 return; 627 } 628 LOG.warn("Getdata rc=" + KeeperException.Code.get(rc) + " " + path + " remaining retries=" 629 + retry_count); 630 if (retry_count == 0) { 631 SplitLogCounters.tot_mgr_get_data_err.increment(); 632 getDataSetWatchFailure(path); 633 } else { 634 SplitLogCounters.tot_mgr_get_data_retry.increment(); 635 getDataSetWatch(path, retry_count - 1); 636 } 637 return; 638 } 639 try { 640 getDataSetWatchSuccess(path, data, stat.getVersion()); 641 } catch (DeserializationException e) { 642 LOG.warn("Deserialization problem", e); 643 } 644 return; 645 } 646 } 647 648 /** 649 * Asynchronous handler for zk delete node results. Retries on failures. 650 */ 651 public class DeleteAsyncCallback implements AsyncCallback.VoidCallback { 652 private final Logger LOG = LoggerFactory.getLogger(DeleteAsyncCallback.class); 653 654 @Override 655 public void processResult(int rc, String path, Object ctx) { 656 SplitLogCounters.tot_mgr_node_delete_result.increment(); 657 if (rc != 0) { 658 if (needAbandonRetries(rc, "Delete znode " + path)) { 659 details.getFailedDeletions().add(path); 660 return; 661 } 662 if (rc != KeeperException.Code.NONODE.intValue()) { 663 SplitLogCounters.tot_mgr_node_delete_err.increment(); 664 Long retry_count = (Long) ctx; 665 LOG.warn("Delete rc=" + KeeperException.Code.get(rc) + " for " + path 666 + " remaining retries=" + retry_count); 667 if (retry_count == 0) { 668 LOG.warn("Delete failed " + path); 669 details.getFailedDeletions().add(path); 670 deleteNodeFailure(path); 671 } else { 672 deleteNode(path, retry_count - 1); 673 } 674 return; 675 } else { 676 LOG.info(path + " does not exist. Either was created but deleted behind our" 677 + " back by another pending delete OR was deleted" 678 + " in earlier retry rounds. zkretries = " + ctx); 679 } 680 } else { 681 LOG.debug("Deleted " + path); 682 } 683 deleteNodeSuccess(path); 684 } 685 } 686 687 /** 688 * Asynchronous handler for zk create RESCAN-node results. Retries on failures. 689 * <p> 690 * A RESCAN node is created using PERSISTENT_SEQUENTIAL flag. It is a signal for all the 691 * {@link org.apache.hadoop.hbase.regionserver.SplitLogWorker}s to rescan for new tasks. 692 */ 693 public class CreateRescanAsyncCallback implements AsyncCallback.StringCallback { 694 private final Logger LOG = LoggerFactory.getLogger(CreateRescanAsyncCallback.class); 695 696 @Override 697 public void processResult(int rc, String path, Object ctx, String name) { 698 if (rc != 0) { 699 if (needAbandonRetries(rc, "CreateRescan znode " + path)) { 700 return; 701 } 702 Long retry_count = (Long) ctx; 703 LOG.warn("rc=" + KeeperException.Code.get(rc) + " for " + path + " remaining retries=" 704 + retry_count); 705 if (retry_count == 0) { 706 createRescanFailure(); 707 } else { 708 rescan(retry_count - 1); 709 } 710 return; 711 } 712 // path is the original arg, name is the actual name that was created 713 createRescanSuccess(name); 714 } 715 } 716 717 @Override 718 public void setDetails(SplitLogManagerDetails details) { 719 this.details = details; 720 } 721 722 @Override 723 public SplitLogManagerDetails getDetails() { 724 return details; 725 } 726 727 /** 728 * Temporary function that is used by unit tests only 729 */ 730 public void setIgnoreDeleteForTesting(boolean b) { 731 ignoreZKDeleteForTesting = b; 732 } 733}