org.slf4j.Logger.isDebugEnabled()方法的使用及代码示例

x33g5p2x  于2022-01-16 转载在 其他  
字(9.9k)|赞(0)|评价(0)|浏览(849)

本文整理了Java中org.slf4j.Logger.isDebugEnabled()方法的一些代码示例,展示了Logger.isDebugEnabled()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。Logger.isDebugEnabled()方法的具体详情如下:
包路径:org.slf4j.Logger
类名称:Logger
方法名:isDebugEnabled

Logger.isDebugEnabled介绍

[英]Is the logger instance enabled for the DEBUG level?
[中]是否为调试级别启用了记录器实例?

代码示例

代码示例来源:origin: spring-projects/spring-framework

  1. public void debug(Object message) {
  2. if (message instanceof String || this.logger.isDebugEnabled()) {
  3. this.logger.debug(String.valueOf(message));
  4. }
  5. }

代码示例来源:origin: skylot/jadx

  1. public void printMissingClasses() {
  2. int count = missingClasses.size();
  3. if (count == 0) {
  4. return;
  5. }
  6. LOG.warn("Found {} references to unknown classes", count);
  7. if (LOG.isDebugEnabled()) {
  8. List<String> clsNames = new ArrayList<>(missingClasses);
  9. Collections.sort(clsNames);
  10. for (String cls : clsNames) {
  11. LOG.debug(" {}", cls);
  12. }
  13. }
  14. }
  15. }

代码示例来源:origin: apache/hive

  1. static void logException(String msg, Exception e) {
  2. if (LOG.isDebugEnabled()) {
  3. LOG.debug(msg, e);
  4. } else {
  5. LOG.info(msg + ": " + e.getMessage());
  6. }
  7. }

代码示例来源:origin: apache/hive

  1. private void addCredentials(ReduceWork reduceWork, DAG dag) {
  2. Set<URI> fileSinkUris = new HashSet<URI>();
  3. List<Node> topNodes = new ArrayList<Node>();
  4. topNodes.add(reduceWork.getReducer());
  5. collectFileSinkUris(topNodes, fileSinkUris);
  6. if (LOG.isDebugEnabled()) {
  7. for (URI fileSinkUri: fileSinkUris) {
  8. LOG.debug("Marking ReduceWork output URI as needing credentials: " + fileSinkUri);
  9. }
  10. }
  11. dag.addURIsForCredentials(fileSinkUris);
  12. }

代码示例来源:origin: apache/hbase

  1. private void stopUsingCurrentWriter() {
  2. if (currentWriter != null) {
  3. if (LOG.isDebugEnabled()) {
  4. LOG.debug("Stopping to use a writer after [" + Bytes.toString(currentWriterEndKey)
  5. + "] row; wrote out " + cellsInCurrentWriter + " kvs");
  6. }
  7. cellsInCurrentWriter = 0;
  8. }
  9. currentWriter = null;
  10. currentWriterEndKey =
  11. (existingWriters.size() + 1 == boundaries.size()) ? null : boundaries.get(existingWriters
  12. .size() + 1);
  13. }
  14. }

代码示例来源:origin: alibaba/fescar

  1. private void printMergeMessageLog(MergedWarpMessage mergeMessage) {
  2. if (LOGGER.isDebugEnabled()) {
  3. LOGGER.debug("merge msg size:" + mergeMessage.msgIds.size());
  4. for (AbstractMessage cm : mergeMessage.msgs) { LOGGER.debug(cm.toString()); }
  5. StringBuffer sb = new StringBuffer();
  6. for (long l : mergeMessage.msgIds) { sb.append(MSG_ID_PREFIX).append(l).append(SINGLE_LOG_POSTFIX); }
  7. sb.append("\n");
  8. for (long l : futures.keySet()) { sb.append(FUTURES_PREFIX).append(l).append(SINGLE_LOG_POSTFIX); }
  9. LOGGER.debug(sb.toString());
  10. }
  11. }
  12. }

代码示例来源:origin: alibaba/fescar

  1. @Override
  2. public void onCheckMessage(long msgId, ChannelHandlerContext ctx, ServerMessageSender sender) {
  3. try {
  4. sender.sendResponse(msgId, ctx.channel(), HeartbeatMessage.PONG);
  5. } catch (Throwable throwable) {
  6. LOGGER.error("", "send response error", throwable);
  7. }
  8. if (LOGGER.isDebugEnabled()) {
  9. LOGGER.debug("received PING from " + ctx.channel().remoteAddress());
  10. }
  11. }

代码示例来源:origin: apache/flink

  1. @Override
  2. public void initializeState(StateInitializationContext context) throws Exception {
  3. super.initializeState(context);
  4. checkState(checkpointedState == null, "The reader state has already been initialized.");
  5. checkpointedState = context.getOperatorStateStore().getSerializableListState("splits");
  6. int subtaskIdx = getRuntimeContext().getIndexOfThisSubtask();
  7. if (context.isRestored()) {
  8. LOG.info("Restoring state for the {} (taskIdx={}).", getClass().getSimpleName(), subtaskIdx);
  9. // this may not be null in case we migrate from a previous Flink version.
  10. if (restoredReaderState == null) {
  11. restoredReaderState = new ArrayList<>();
  12. for (TimestampedFileInputSplit split : checkpointedState.get()) {
  13. restoredReaderState.add(split);
  14. }
  15. if (LOG.isDebugEnabled()) {
  16. LOG.debug("{} (taskIdx={}) restored {}.", getClass().getSimpleName(), subtaskIdx, restoredReaderState);
  17. }
  18. }
  19. } else {
  20. LOG.info("No state to restore for the {} (taskIdx={}).", getClass().getSimpleName(), subtaskIdx);
  21. }
  22. }

代码示例来源:origin: apache/pulsar

  1. @Override
  2. protected void handleGetTopicsOfNamespaceSuccess(CommandGetTopicsOfNamespaceResponse success) {
  3. checkArgument(state == State.Ready);
  4. long requestId = success.getRequestId();
  5. List<String> topics = success.getTopicsList();
  6. if (log.isDebugEnabled()) {
  7. log.debug("{} Received get topics of namespace success response from server: {} - topics.size: {}",
  8. ctx.channel(), success.getRequestId(), topics.size());
  9. }
  10. CompletableFuture<List<String>> requestFuture = pendingGetTopicsRequests.remove(requestId);
  11. if (requestFuture != null) {
  12. requestFuture.complete(topics);
  13. } else {
  14. log.warn("{} Received unknown request id from server: {}", ctx.channel(), success.getRequestId());
  15. }
  16. }

代码示例来源:origin: apache/flink

  1. @Override
  2. public void initializeState(FunctionInitializationContext context) throws Exception {
  3. Preconditions.checkArgument(this.restoredBucketStates == null,
  4. "The " + getClass().getSimpleName() + " has already been initialized.");
  5. try {
  6. initFileSystem();
  7. } catch (IOException e) {
  8. LOG.error("Error while creating FileSystem when initializing the state of the RollingSink.", e);
  9. throw new RuntimeException("Error while creating FileSystem when initializing the state of the RollingSink.", e);
  10. }
  11. if (this.refTruncate == null) {
  12. this.refTruncate = reflectTruncate(fs);
  13. }
  14. OperatorStateStore stateStore = context.getOperatorStateStore();
  15. restoredBucketStates = stateStore.getSerializableListState("rolling-states");
  16. int subtaskIndex = getRuntimeContext().getIndexOfThisSubtask();
  17. if (context.isRestored()) {
  18. LOG.info("Restoring state for the {} (taskIdx={}).", getClass().getSimpleName(), subtaskIndex);
  19. for (BucketState bucketState : restoredBucketStates.get()) {
  20. handleRestoredBucketState(bucketState);
  21. }
  22. if (LOG.isDebugEnabled()) {
  23. LOG.debug("{} (taskIdx= {}) restored {}", getClass().getSimpleName(), subtaskIndex, bucketState);
  24. }
  25. } else {
  26. LOG.info("No state to restore for the {} (taskIdx= {}).", getClass().getSimpleName(), subtaskIndex);
  27. }
  28. }

代码示例来源:origin: apache/geode

  1. @Override
  2. protected void rebalanceCache() {
  3. try {
  4. getLogger().info("Rebalancing: " + this.cache);
  5. RebalanceResults results = RegionHelper.rebalanceCache(this.cache);
  6. if (getLogger().isDebugEnabled()) {
  7. getLogger().debug("Done rebalancing: " + this.cache);
  8. getLogger().debug(RegionHelper.getRebalanceResultsMessage(results));
  9. }
  10. } catch (Exception e) {
  11. getLogger().warn("Rebalance failed because of the following exception:", e);
  12. }
  13. }

代码示例来源:origin: resilience4j/resilience4j

  1. @Override
  2. public void onSubscribe(Subscription s) {
  3. if (LOG.isDebugEnabled()) {
  4. LOG.info("onSubscribe");
  5. }
  6. sa.setSubscription(s);
  7. }

代码示例来源:origin: apache/nifi

  1. void expire(final FlowFile flowFile, final String details) {
  2. try {
  3. final ProvenanceEventRecord record = build(flowFile, ProvenanceEventType.EXPIRE).setDetails(details).build();
  4. events.add(record);
  5. } catch (final Exception e) {
  6. logger.error("Failed to generate Provenance Event due to " + e);
  7. if (logger.isDebugEnabled()) {
  8. logger.error("", e);
  9. }
  10. }
  11. }

代码示例来源:origin: apache/kafka

  1. /**
  2. * Queue a call for sending.
  3. *
  4. * If the AdminClient thread has exited, this will fail. Otherwise, it will succeed (even
  5. * if the AdminClient is shutting down). This function should called when retrying an
  6. * existing call.
  7. *
  8. * @param call The new call object.
  9. * @param now The current time in milliseconds.
  10. */
  11. void enqueue(Call call, long now) {
  12. if (log.isDebugEnabled()) {
  13. log.debug("Queueing {} with a timeout {} ms from now.", call, call.deadlineMs - now);
  14. }
  15. boolean accepted = false;
  16. synchronized (this) {
  17. if (newCalls != null) {
  18. newCalls.add(call);
  19. accepted = true;
  20. }
  21. }
  22. if (accepted) {
  23. client.wakeup(); // wake the thread if it is in poll()
  24. } else {
  25. log.debug("The AdminClient thread has exited. Timing out {}.", call);
  26. call.fail(Long.MAX_VALUE, new TimeoutException("The AdminClient thread has exited."));
  27. }
  28. }

代码示例来源:origin: apache/hive

  1. private static boolean[] pickStripesInternal(SearchArgument sarg, int[] filterColumns,
  2. List<StripeStatistics> stripeStats, int stripeCount, Path filePath,
  3. final SchemaEvolution evolution) {
  4. boolean[] includeStripe = new boolean[stripeCount];
  5. for (int i = 0; i < includeStripe.length; ++i) {
  6. includeStripe[i] = (i >= stripeStats.size()) ||
  7. isStripeSatisfyPredicate(stripeStats.get(i), sarg, filterColumns, evolution);
  8. if (LOG.isDebugEnabled() && !includeStripe[i]) {
  9. LOG.debug("Eliminating ORC stripe-" + i + " of file '" + filePath
  10. + "' as it did not satisfy predicate condition.");
  11. }
  12. }
  13. return includeStripe;
  14. }

代码示例来源:origin: apache/flink

  1. @Override
  2. public InputSplit getNextInputSplit(String host, int taskId) {
  3. InputSplit next = null;
  4. // keep the synchronized part short
  5. synchronized (this.splits) {
  6. if (this.splits.size() > 0) {
  7. next = this.splits.remove(this.splits.size() - 1);
  8. }
  9. }
  10. if (LOG.isDebugEnabled()) {
  11. if (next == null) {
  12. LOG.debug("No more input splits available");
  13. } else {
  14. LOG.debug("Assigning split " + next + " to " + host);
  15. }
  16. }
  17. return next;
  18. }
  19. }

代码示例来源:origin: micronaut-projects/micronaut-core

  1. private void logException(Throwable cause) {
  2. //handling connection reset by peer exceptions
  3. if (cause instanceof IOException && IGNORABLE_ERROR_MESSAGE.matcher(cause.getMessage()).matches()) {
  4. if (LOG.isDebugEnabled()) {
  5. LOG.debug("Swallowed an IOException caused by client connectivity: " + cause.getMessage(), cause);
  6. }
  7. } else {
  8. if (LOG.isErrorEnabled()) {
  9. LOG.error("Unexpected error occurred: " + cause.getMessage(), cause);
  10. }
  11. }
  12. }

代码示例来源:origin: apache/nifi

  1. @Override
  2. public Response toResponse(NoConnectedNodesException ex) {
  3. // log the error
  4. logger.info(String.format("Cluster failed processing request: %s. Returning %s response.", ex, Response.Status.INTERNAL_SERVER_ERROR));
  5. if (logger.isDebugEnabled()) {
  6. logger.debug(StringUtils.EMPTY, ex);
  7. }
  8. return Response.status(Response.Status.INTERNAL_SERVER_ERROR).entity("Action was performed, but no nodes are connected.").type("text/plain").build();
  9. }

代码示例来源:origin: perwendel/spark

  1. /**
  2. * Trigger a browser redirect
  3. *
  4. * @param location Where to redirect
  5. */
  6. public void redirect(String location) {
  7. if (LOG.isDebugEnabled()) {
  8. LOG.debug("Redirecting ({} {} to {}", "Found", HttpServletResponse.SC_FOUND, location);
  9. }
  10. try {
  11. response.sendRedirect(location);
  12. } catch (IOException ioException) {
  13. LOG.warn("Redirect failure", ioException);
  14. }
  15. }

代码示例来源:origin: spring-projects/spring-framework

  1. public void debug(Object message, Throwable exception) {
  2. if (message instanceof String || this.logger.isDebugEnabled()) {
  3. this.logger.debug(String.valueOf(message), exception);
  4. }
  5. }

相关文章