diff --git a/dev-support/pmd/pmd-ruleset.xml b/dev-support/pmd/pmd-ruleset.xml index d03a463def62..e3afd8e3081f 100644 --- a/dev-support/pmd/pmd-ruleset.xml +++ b/dev-support/pmd/pmd-ruleset.xml @@ -39,6 +39,7 @@ + diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java index 7ee2a5160540..7141a65306dd 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java @@ -93,8 +93,8 @@ public String toString() { public void appendTo(StringBuilder sb) { containerBlockID.appendTo(sb); - sb.append(" bcsId: ").append(blockCommitSequenceId); - sb.append(" replicaIndex: ").append(replicaIndex); + sb.append(" bcsId: ").append(blockCommitSequenceId) + .append(" replicaIndex: ").append(replicaIndex); } @JsonIgnore diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java index cff7d73dcd13..d168fdc5a4d7 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java @@ -538,20 +538,19 @@ public int hashCode() { @Override public String toString() { final StringBuilder b = - new StringBuilder(getClass().getSimpleName()).append('{'); - b.append(" Id: ").append(id.getId()); - b.append(", Nodes: ["); + new StringBuilder(getClass().getSimpleName()).append('{') + .append(" Id: ").append(id.getId()) + .append(", Nodes: ["); for (DatanodeDetails datanodeDetails : nodeStatus.keySet()) { - b.append(" {").append(datanodeDetails); - b.append(", ReplicaIndex: ").append(this.getReplicaIndex(datanodeDetails)).append("},"); - } - b.append(']'); - b.append(", ReplicationConfig: ").append(replicationConfig); - b.append(", State:").append(getPipelineState()); - b.append(", leaderId:").append(leaderId != null ? leaderId.toString() : ""); - b.append(", CreationTimestamp").append(getCreationTimestamp() - .atZone(ZoneId.systemDefault())); - b.append('}'); + b.append(" {").append(datanodeDetails) + .append(", ReplicaIndex: ").append(this.getReplicaIndex(datanodeDetails)).append("},"); + } + b.append(']') + .append(", ReplicationConfig: ").append(replicationConfig) + .append(", State:").append(getPipelineState()) + .append(", leaderId:").append(leaderId != null ? leaderId.toString() : "") + .append(", CreationTimestamp").append(getCreationTimestamp().atZone(ZoneId.systemDefault())) + .append('}'); return b.toString(); } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java index 7b0fda00e130..8e94ba01eca8 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java @@ -299,8 +299,8 @@ public String toString() { public void appendTo(StringBuilder sb) { sb.append("[blockId="); blockID.appendTo(sb); - sb.append(", size=").append(size); - sb.append(']'); + sb.append(", size=").append(size) + .append(']'); } public long getBlockGroupLength() { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java index 39525eb08938..490772e366c4 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java @@ -1330,13 +1330,13 @@ public static String smProtoToString(RaftGroupId gid, if (containerController != null) { String location = containerController.getContainerLocation(contId); - builder.append(", container path="); - builder.append(location); + builder.append(", container path=") + .append(location); } } catch (Exception t) { LOG.info("smProtoToString failed", t); - builder.append("smProtoToString failed with "); - builder.append(t.getMessage()); + builder.append("smProtoToString failed with ") + .append(t.getMessage()); } return builder.toString(); } diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java index 8574909b386e..bc8040b24bfc 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java @@ -157,13 +157,13 @@ public ReplicationCommandPriority getPriority() { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append(getType()); - sb.append(": cmdID: ").append(getId()) + sb.append(getType()) + .append(": cmdID: ").append(getId()) .append(", encodedToken: \"").append(getEncodedToken()).append('"') .append(", term: ").append(getTerm()) - .append(", deadlineMsSinceEpoch: ").append(getDeadline()); - sb.append(", containerId=").append(getContainerID()); - sb.append(", replicaIndex=").append(getReplicaIndex()); + .append(", deadlineMsSinceEpoch: ").append(getDeadline()) + .append(", containerId=").append(getContainerID()) + .append(", replicaIndex=").append(getReplicaIndex()); if (targetDatanode != null) { sb.append(", targetNode=").append(targetDatanode); } else { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java index 5a26a7b569e9..c4bfdf2c5b9a 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java @@ -218,8 +218,8 @@ private String generateVolumeLocation(String base, int volumeCount) { StringBuilder sb = new StringBuilder(); for (int i = 0; i < volumeCount; i++) { - sb.append(base).append("/vol").append(i); - sb.append(','); + sb.append(base).append("/vol").append(i) + .append(','); } return sb.substring(0, sb.length() - 1); } diff --git a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java index a6f6dba51753..056503abb743 100644 --- a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java +++ b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java @@ -480,19 +480,16 @@ protected byte[][] toArrays(ECChunk[] chunks) { protected void dumpSetting() { if (allowDump) { StringBuilder sb = new StringBuilder("Erasure coder test settings:\n"); - sb.append(" numDataUnits=").append(numDataUnits); - sb.append(" numParityUnits=").append(numParityUnits); - sb.append(" chunkSize=").append(chunkSize).append('\n'); - - sb.append(" erasedDataIndexes="). - append(Arrays.toString(erasedDataIndexes)); - sb.append(" erasedParityIndexes="). - append(Arrays.toString(erasedParityIndexes)); - sb.append(" usingDirectBuffer=").append(usingDirectBuffer); - sb.append(" allowVerboseDump=").append(allowDump); - sb.append('\n'); - - System.out.println(sb.toString()); + sb.append(" numDataUnits=").append(numDataUnits) + .append(" numParityUnits=").append(numParityUnits) + .append(" chunkSize=").append(chunkSize).append('\n') + .append(" erasedDataIndexes=").append(Arrays.toString(erasedDataIndexes)) + .append(" erasedParityIndexes=").append(Arrays.toString(erasedParityIndexes)) + .append(" usingDirectBuffer=").append(usingDirectBuffer) + .append(" allowVerboseDump=").append(allowDump) + .append('\n'); + + System.out.println(sb); } } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java index 658bc41825c1..f83b8fae1203 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java @@ -891,20 +891,20 @@ private int getAvailableNodesCount(String scope, List excludedScopes, public String toString() { // print max level StringBuilder tree = new StringBuilder(); - tree.append("Level: "); - tree.append(maxLevel); - tree.append('\n'); + tree.append("Level: ") + .append(maxLevel) + .append('\n'); netlock.readLock().lock(); try { // print the number of leaves int numOfLeaves = clusterTree.getNumOfLeaves(); - tree.append("Number of leaves:"); - tree.append(numOfLeaves); - tree.append('\n'); + tree.append("Number of leaves:") + .append(numOfLeaves) + .append('\n'); // print all nodes for (int i = 0; i < numOfLeaves; i++) { - tree.append(clusterTree.getLeaf(i).getNetworkFullPath()); - tree.append('\n'); + tree.append(clusterTree.getLeaf(i).getNetworkFullPath()) + .append('\n'); } } finally { netlock.readLock().unlock(); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java index b611541578ea..6316bfe95f91 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java @@ -49,10 +49,10 @@ public long getReplicatedSize() { @Override public String toString() { StringBuilder sb = new StringBuilder(64); - sb.append(" localID: ").append(blockID.getContainerBlockID().getLocalID()); - sb.append(" containerID: ").append(blockID.getContainerBlockID().getContainerID()); - sb.append(" size: ").append(size); - sb.append(" replicatedSize: ").append(replicatedSize); + sb.append(" localID: ").append(blockID.getContainerBlockID().getLocalID()) + .append(" containerID: ").append(blockID.getContainerBlockID().getContainerID()) + .append(" size: ").append(size) + .append(" replicatedSize: ").append(replicatedSize); return sb.toString(); } } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java index e1cd6a9a85ab..d0725cdd91f4 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java @@ -169,8 +169,8 @@ public static String smProtoToString(StateMachineLogEntryProto proto) { SCMRatisRequestProto.parseFrom(proto.getLogData().asReadOnlyByteBuffer()))); } catch (Throwable ex) { LOG.error("smProtoToString failed", ex); - builder.append("smProtoToString failed with"); - builder.append(ex.getMessage()); + builder.append("smProtoToString failed with") + .append(ex.getMessage()); } return builder.toString(); } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java index 49b01ef44847..2ee85d9d0d93 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java @@ -497,11 +497,11 @@ public Map> getContainersPendingReplication(DatanodeDe private String replicaDetails(Collection replicas) { StringBuilder sb = new StringBuilder(); - sb.append("Replicas{"); - sb.append(replicas.stream() - .map(Object::toString) - .collect(Collectors.joining(","))); - sb.append('}'); + sb.append("Replicas{") + .append(replicas.stream() + .map(Object::toString) + .collect(Collectors.joining(","))) + .append('}'); return sb.toString(); } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java index b161e0e84d76..0725cb7b71b9 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java @@ -388,8 +388,8 @@ public List getContainerWithPipelineBatch( try { ContainerWithPipeline cp = getContainerWithPipelineCommon(containerID); cpList.add(cp); - strContainerIDs.append(ContainerID.valueOf(containerID).toString()); - strContainerIDs.append(','); + strContainerIDs.append(ContainerID.valueOf(containerID).toString()) + .append(','); } catch (IOException ex) { AUDIT.logReadFailure(buildAuditMessageForFailure( SCMAction.GET_CONTAINER_WITH_PIPELINE_BATCH, diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java index 51df2647b444..dd403f46a182 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java @@ -280,12 +280,12 @@ private String constructCommandAuditMap(List cmds) { auditMap.append('['); for (SCMCommandProto cmd : cmds) { if (cmd.getCommandType().equals(deleteBlocksCommand)) { - auditMap.append("commandType: ").append(cmd.getCommandType()); - auditMap.append(" deleteTransactionsCount: ") - .append(cmd.getDeleteBlocksCommandProto().getDeletedBlocksTransactionsCount()); - auditMap.append(" cmdID: ").append(cmd.getDeleteBlocksCommandProto().getCmdId()); - auditMap.append(" encodedToken: \"").append(cmd.getEncodedToken()).append('"'); - auditMap.append(" deadlineMsSinceEpoch: ").append(cmd.getDeadlineMsSinceEpoch()); + auditMap.append("commandType: ").append(cmd.getCommandType()) + .append(" deleteTransactionsCount: ") + .append(cmd.getDeleteBlocksCommandProto().getDeletedBlocksTransactionsCount()) + .append(" cmdID: ").append(cmd.getDeleteBlocksCommandProto().getCmdId()) + .append(" encodedToken: \"").append(cmd.getEncodedToken()).append('"') + .append(" deadlineMsSinceEpoch: ").append(cmd.getDeadlineMsSinceEpoch()); } else { auditMap.append(TextFormat.shortDebugString(cmd)); } diff --git a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java index 2e0117269f09..014daaa8dcf6 100644 --- a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java +++ b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java @@ -286,9 +286,9 @@ public int getInvocationCount() { public String toString() { final StringBuilder sb = new StringBuilder( "FixedRetryInterval{"); - sb.append("interval=").append(intervalMillis); - sb.append(", invocationCount=").append(invocationCount); - sb.append('}'); + sb.append("interval=").append(intervalMillis) + .append(", invocationCount=").append(invocationCount) + .append('}'); return sb.toString(); } } diff --git a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java index c7a8873c695b..e5b7555870b4 100644 --- a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java +++ b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java @@ -93,8 +93,8 @@ private static String buildThreadDump() { Thread.State.WAITING.equals(thread.getState()) ? "WAITING (on object monitor)" : thread.getState())); for (StackTraceElement stackTraceElement : e.getValue()) { - dump.append("\n at "); - dump.append(stackTraceElement); + dump.append("\n at ") + .append(stackTraceElement); } dump.append('\n'); } diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java index 62f38da74d01..4db959ebf7d9 100644 --- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java +++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java @@ -100,9 +100,9 @@ private String generateReport(List protos) { String dn = DiskBalancerSubCommandUtil.getDatanodeHostAndIp(p.getNode()); StringBuilder header = new StringBuilder(); - header.append("Datanode: ").append(dn).append('\n'); - header.append("Aggregate VolumeDataDensity: "). - append(p.getCurrentVolumeDensitySum()).append('\n'); + header.append("Datanode: ").append(dn).append('\n') + .append("Aggregate VolumeDataDensity: ").append(p.getCurrentVolumeDensitySum()) + .append('\n'); if (p.hasIdealUsage() && p.hasDiskBalancerConf() && p.getDiskBalancerConf().hasThreshold()) { @@ -110,11 +110,11 @@ private String generateReport(List protos) { double threshold = p.getDiskBalancerConf().getThreshold(); double lt = idealUsage - threshold / 100.0; double ut = idealUsage + threshold / 100.0; - header.append("IdealUsage: ").append(String.format("%.8f", idealUsage)); - header.append(" | Threshold: ").append(threshold).append('%'); - header.append(" | ThresholdRange: (").append(String.format("%.8f", lt)); - header.append(", ").append(String.format("%.8f", ut)).append(')').append('\n').append('\n'); - header.append("Volume Details:").append('\n'); + header.append("IdealUsage: ").append(String.format("%.8f", idealUsage)) + .append(" | Threshold: ").append(threshold).append('%') + .append(" | ThresholdRange: (").append(String.format("%.8f", lt)) + .append(", ").append(String.format("%.8f", ut)).append(')').append('\n').append('\n') + .append("Volume Details:").append('\n'); } formatBuilder.append("%s%n"); contentList.add(header.toString()); @@ -150,21 +150,20 @@ private String generateReport(List protos) { } } - formatBuilder.append("%nNote:%n"); - formatBuilder.append(" - Aggregate VolumeDataDensity: Sum of per-volume density" + - " (deviation from ideal); higher means more imbalance.%n"); - formatBuilder.append(" - IdealUsage: Target utilization ratio (0-1) when volumes" + - " are evenly balanced.%n"); - formatBuilder.append(" - ThresholdRange: Acceptable deviation (percent); volumes within" + - " IdealUsage +/- Threshold are considered balanced.%n"); - formatBuilder.append(" - VolumeDensity: Deviation of a particular volume's utilization from IdealUsage.%n"); - formatBuilder.append(" - Utilization: Ratio of actual used space to capacity (0-1) for a particular volume.%n"); - formatBuilder.append(" - TotalCapacity: Total volume capacity.%n"); - formatBuilder.append(" - UsedSpace: Ozone used space.%n"); - formatBuilder.append(" - Container Pre-AllocatedSpace: Space reserved for containers not yet written to disk.%n"); - formatBuilder.append(" - EffectiveUsedSpace: This is the actual used space of volume which is visible" + - " to the diskBalancer : (ozoneCapacity minus ozoneAvailable) + containerPreAllocatedSpace + " + - "move delta for source volume.%n"); + formatBuilder.append("%nNote:%n") + .append(" - Aggregate VolumeDataDensity: Sum of per-volume density (deviation from ideal);") + .append(" higher means more imbalance.%n") + .append(" - IdealUsage: Target utilization ratio (0-1) when volumes are evenly balanced.%n") + .append(" - ThresholdRange: Acceptable deviation (percent); volumes within") + .append(" IdealUsage +/- Threshold are considered balanced.%n") + .append(" - VolumeDensity: Deviation of a particular volume's utilization from IdealUsage.%n") + .append(" - Utilization: Ratio of actual used space to capacity (0-1) for a particular volume.%n") + .append(" - TotalCapacity: Total volume capacity.%n") + .append(" - UsedSpace: Ozone used space.%n") + .append(" - Container Pre-AllocatedSpace: Space reserved for containers not yet written to disk.%n") + .append(" - EffectiveUsedSpace: This is the actual used space of volume which is visible") + .append(" to the diskBalancer : (ozoneCapacity minus ozoneAvailable) + containerPreAllocatedSpace + ") + .append("move delta for source volume.%n"); return String.format(formatBuilder.toString(), contentList.toArray(new String[0])); } diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java index f84a5b9a78e2..bd98767f9972 100644 --- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java +++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java @@ -131,13 +131,11 @@ private String generateStatus(List protos) { contentList.add(estimatedTimeLeft >= 0 ? String.valueOf(estimatedTimeLeft) : "N/A"); } - formatBuilder.append("%nNote:%n"); - formatBuilder.append(" - EstBytesToMove is calculated based on the target disk even state" + - " with the configured threshold.%n"); - formatBuilder.append(" - EstTimeLeft is calculated based on EstimatedBytesToMove and configured" + - " disk bandwidth.%n"); - formatBuilder.append(" - Both EstimatedBytes and EstTimeLeft could be non-zero while no containers" + - " can be moved, especially when the configured threshold or disk capacity is too small."); + formatBuilder.append("%nNote:%n") + .append(" - EstBytesToMove is calculated based on the target disk even state with the configured threshold.%n") + .append(" - EstTimeLeft is calculated based on EstimatedBytesToMove and configured disk bandwidth.%n") + .append(" - Both EstimatedBytes and EstTimeLeft could be non-zero while no containers" + + " can be moved, especially when the configured threshold or disk capacity is too small."); return String.format(formatBuilder.toString(), contentList.toArray(new String[0])); diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java index c5477e0cb0ec..951624213ce1 100644 --- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java +++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java @@ -197,8 +197,8 @@ private void printOpenKeysList(ListOpenFilesResult res) { private String getMessageString(ListOpenFilesResult res, List openFileList) { StringBuilder sb = new StringBuilder(); sb.append(res.getTotalOpenKeyCount()) - .append(" total open files. Showing "); - sb.append(openFileList.size()) + .append(" total open files. Showing ") + .append(openFileList.size()) .append(" open files (limit ") .append(limit) .append(") under path prefix:\n ") diff --git a/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java b/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java index f598f40b9b73..f508ec349500 100644 --- a/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java +++ b/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java @@ -210,8 +210,8 @@ private static String executeStatement(String dbName, String sql) int cols = rsm.getColumnCount(); while (rs.next()) { for (int index = 1; index <= cols; index++) { - result.append(rs.getObject(index)); - result.append('\t'); + result.append(rs.getObject(index)) + .append('\t'); } result.append('\n'); } diff --git a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java index 83f6c4d55267..f1b1a568e1e7 100644 --- a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java +++ b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java @@ -89,31 +89,23 @@ public void success(long rowCount) { @Override public String toString() { final StringBuilder stringBuilder = new StringBuilder(); - stringBuilder.append("Result:{"); - stringBuilder.append("containerID="); - stringBuilder.append(originContainerData.getContainerID()); - stringBuilder.append(", originContainerSchemaVersion="); - stringBuilder.append( - ((KeyValueContainerData) originContainerData).getSchemaVersion()); + stringBuilder.append("Result:{") + .append("containerID=") + .append(originContainerData.getContainerID()) + .append(", originContainerSchemaVersion=") + .append(((KeyValueContainerData) originContainerData).getSchemaVersion()); if (newContainerData != null) { - stringBuilder.append(", schemaV2ContainerFileBackupPath="); - stringBuilder.append(backupContainerFilePath); - - stringBuilder.append(", newContainerSchemaVersion="); - stringBuilder.append( - ((KeyValueContainerData) newContainerData).getSchemaVersion()); - - stringBuilder.append(", schemaV3ContainerFilePath="); - stringBuilder.append(newContainerFilePath); + stringBuilder + .append(", schemaV2ContainerFileBackupPath=").append(backupContainerFilePath) + .append(", newContainerSchemaVersion=").append(((KeyValueContainerData) newContainerData).getSchemaVersion()) + .append(", schemaV3ContainerFilePath=").append(newContainerFilePath); } - stringBuilder.append(", totalRow="); - stringBuilder.append(totalRow); - stringBuilder.append(", costMs="); - stringBuilder.append(getCostMs()); - stringBuilder.append(", status="); - stringBuilder.append(status); - stringBuilder.append('}'); + stringBuilder + .append(", totalRow=").append(totalRow) + .append(", costMs=").append(getCostMs()) + .append(", status=").append(status) + .append('}'); return stringBuilder.toString(); } diff --git a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java index 067de25b3ba0..529233b6a90c 100644 --- a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java +++ b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java @@ -86,27 +86,27 @@ public void fail(Exception exception) { @Override public String toString() { final StringBuilder stringBuilder = new StringBuilder(); - stringBuilder.append("Result:{"); - stringBuilder.append("hddsRootDir="); - stringBuilder.append(getHddsVolume().getHddsRootDir()); - stringBuilder.append(", resultList="); + stringBuilder.append("Result:{") + .append("hddsRootDir=") + .append(getHddsVolume().getHddsRootDir()) + .append(", resultList="); AtomicLong total = new AtomicLong(0L); if (resultMap != null) { resultMap.forEach((k, r) -> { - stringBuilder.append(r.toString()); - stringBuilder.append('\n'); + stringBuilder.append(r.toString()) + .append('\n'); total.addAndGet(r.getTotalRow()); }); } - stringBuilder.append(", totalRow="); - stringBuilder.append(total.get()); - stringBuilder.append(", costMs="); - stringBuilder.append(getCost()); - stringBuilder.append(", status="); - stringBuilder.append(status); + stringBuilder.append(", totalRow=") + .append(total.get()) + .append(", costMs=") + .append(getCost()) + .append(", status=") + .append(status); if (e != null) { - stringBuilder.append(", Exception="); - stringBuilder.append(e); + stringBuilder.append(", Exception=") + .append(e); } stringBuilder.append('}'); return stringBuilder.toString(); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java index aee73248db70..fc96045eba3e 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java @@ -188,15 +188,15 @@ public String toString() { Objects.requireNonNull(authority, "authority == null"); StringBuilder sb = new StringBuilder(); if (!isMount()) { - sb.append(volumeName); - sb.append(OZONE_URI_DELIMITER); + sb.append(volumeName) + .append(OZONE_URI_DELIMITER); if (!bucketName.isEmpty()) { - sb.append(bucketName); - sb.append(OZONE_URI_DELIMITER); + sb.append(bucketName) + .append(OZONE_URI_DELIMITER); } } else { - sb.append(mountName); - sb.append(OZONE_URI_DELIMITER); + sb.append(mountName) + .append(OZONE_URI_DELIMITER); } if (!keyName.isEmpty()) { sb.append(keyName); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java index ddf62a45ed8f..e2477a4cef10 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java @@ -185,16 +185,16 @@ void addAll(long versionToAdd, List locationInfoList) { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append("version:").append(version).append(' '); - sb.append("isMultipartKey:").append(isMultipartKey).append(' '); + sb.append("version:").append(version).append(' ') + .append("isMultipartKey:").append(isMultipartKey).append(' '); for (List kliList : locationVersionMap.values()) { for (OmKeyLocationInfo kli: kliList) { - sb.append("conID ").append(kli.getContainerID()); - sb.append(' '); - sb.append("locID ").append(kli.getLocalID()); - sb.append(' '); - sb.append("bcsID ").append(kli.getBlockCommitSequenceId()); - sb.append(" || "); + sb.append("conID ").append(kli.getContainerID()) + .append(' ') + .append("locID ").append(kli.getLocalID()) + .append(' ') + .append("bcsID ").append(kli.getBlockCommitSequenceId()) + .append(" || "); } } return sb.toString(); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java index 9893f3c5c031..de2403e26b6c 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java @@ -308,8 +308,8 @@ public static String getParentDir(@Nonnull String keyName) { public static String appendFileNameToKeyPath(String keyName, String fileName) { StringBuilder newToKeyName = new StringBuilder(keyName); - newToKeyName.append(OZONE_URI_DELIMITER); - newToKeyName.append(fileName); + newToKeyName.append(OZONE_URI_DELIMITER) + .append(fileName); return newToKeyName.toString(); } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java index 99ea12b1cd67..e0d7ebe37b63 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java @@ -133,8 +133,8 @@ public int hashCode() { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append(getClass().getSimpleName()); - sb.append('{'); + sb.append(getClass().getSimpleName()) + .append('{'); if (keyInfo == null) { sb.append(""); } else { diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java index 5d07f331086b..3b4b68d6705e 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java @@ -131,8 +131,8 @@ public int hashCode() { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append(getClass().getSimpleName()); - sb.append('{'); + sb.append(getClass().getSimpleName()) + .append('{'); if (keyInfo == null) { sb.append(""); } else { diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java index 89aba42511da..f17603d7ec1f 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java @@ -178,8 +178,8 @@ public List> getLongestPrefixPath(String path) { public static String radixPathToString(List> path) { StringBuilder sb = new StringBuilder(); for (RadixNode n : path) { - sb.append(n.getName()); - sb.append(n.getName().equals(PATH_DELIMITER) ? "" : PATH_DELIMITER); + sb.append(n.getName()) + .append(n.getName().equals(PATH_DELIMITER) ? "" : PATH_DELIMITER); } return sb.toString(); } diff --git a/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java b/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java index 4fddb9054be7..a1d1e79731b9 100644 --- a/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java +++ b/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java @@ -184,15 +184,15 @@ private void printProgressBar(PrintStream stream, long value) { for (int j = 0; j < (100 - percent) / shrinkTimes; j++) { sb.append(' '); } - sb.append("| "); - sb.append(value).append('/').append(maxValue); + sb.append("| ") + .append(value).append('/').append(maxValue); long timeInSec = TimeUnit.SECONDS.convert( System.nanoTime() - startTime, TimeUnit.NANOSECONDS); String timeToPrint = String.format("%d:%02d:%02d", timeInSec / 3600, (timeInSec % 3600) / 60, timeInSec % 60); - sb.append(" Time: ").append(timeToPrint); - sb.append("| "); - sb.append(realTimeMessage); + sb.append(" Time: ").append(timeToPrint) + .append("| ") + .append(realTimeMessage); stream.print(sb.toString()); } } diff --git a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java index d8fbd1f1ce39..730cd4ac2e92 100644 --- a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java +++ b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java @@ -1534,10 +1534,10 @@ private void completeMPU(String keyName, String uploadId, List complet StringBuilder completionXml = new StringBuilder(); completionXml.append("\n"); for (PartETag part : completedParts) { - completionXml.append(" \n"); - completionXml.append(" ").append(part.getPartNumber()).append("\n"); - completionXml.append(" ").append(stripQuotes(part.getETag())).append("\n"); - completionXml.append(" \n"); + completionXml.append(" \n") + .append(" ").append(part.getPartNumber()).append("\n") + .append(" ").append(stripQuotes(part.getETag())).append("\n") + .append(" \n"); } completionXml.append(""); diff --git a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java index 35a53dd328bb..4b165175f226 100644 --- a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java +++ b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java @@ -1280,10 +1280,10 @@ private String buildCompleteMultipartUploadXml(List parts) { StringBuilder xml = new StringBuilder(); xml.append("\n"); for (CompletedPart part : parts) { - xml.append(" \n"); - xml.append(" ").append(part.partNumber()).append("\n"); - xml.append(" ").append(stripQuotes(part.eTag())).append("\n"); - xml.append(" \n"); + xml.append(" \n") + .append(" ").append(part.partNumber()).append("\n") + .append(" ").append(stripQuotes(part.eTag())).append("\n") + .append(" \n"); } xml.append(""); return xml.toString(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java index 198874fac8da..d87838287e42 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java @@ -434,13 +434,13 @@ public void testCompareConfigurationClassAgainstXml() { LOG.info("({} member variables)\n", configurationMemberVariables.size()); StringBuilder xmlErrorMsg = new StringBuilder(); for (Class c : configurationClasses) { - xmlErrorMsg.append(c); - xmlErrorMsg.append(' '); + xmlErrorMsg.append(c) + .append(' '); } - xmlErrorMsg.append("has "); - xmlErrorMsg.append(missingXmlSize); - xmlErrorMsg.append(" variables missing in "); - xmlErrorMsg.append(xmlFilename); + xmlErrorMsg.append("has ") + .append(missingXmlSize) + .append(" variables missing in ") + .append(xmlFilename); LOG.error(xmlErrorMsg.toString()); if (missingXmlSize == 0) { LOG.info(" (None)"); @@ -482,10 +482,10 @@ public void testCompareXmlAgainstConfigurationClass() { LOG.info("File {} ({} properties)", xmlFilename, xmlKeyValueMap.size()); StringBuilder configErrorMsg = new StringBuilder(); - configErrorMsg.append(xmlFilename); - configErrorMsg.append(" has "); - configErrorMsg.append(missingConfigSize); - configErrorMsg.append(" properties missing in"); + configErrorMsg.append(xmlFilename) + .append(" has ") + .append(missingConfigSize) + .append(" properties missing in"); Arrays.stream(configurationClasses) .forEach(c -> configErrorMsg.append(" ").append(c)); LOG.info(configErrorMsg.toString()); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java index 31d00b7228ae..cb8045e0e6e5 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java @@ -252,9 +252,9 @@ private void setupPostMethod(Collection toExcludeList) String endBoundary = boundary + "--" + crNl; StringBuilder sb = new StringBuilder(); toExcludeList.forEach(sfn -> { - sb.append(boundary).append(crNl); - sb.append(contentDisposition); - sb.append(sfn).append(crNl); + sb.append(boundary).append(crNl) + .append(contentDisposition) + .append(sfn).append(crNl); }); sb.append(endBoundary); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java index ec1d522b721e..fa801804dd82 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java @@ -4505,9 +4505,9 @@ private void doMultipartUpload(OzoneBucket bucket, String keyName, byte val, // Combine all parts data, and check is it matching with get key data. String part1 = new String(data, UTF_8); String part2 = new String(data, UTF_8); - sb.append(part1); - sb.append(part2); - sb.append(part3); + sb.append(part1) + .append(part2) + .append(part3); assertEquals(sb.toString(), new String(fileContent, UTF_8)); OmKeyArgs keyArgs = new OmKeyArgs.Builder() diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java index ac12189b1f62..8acc63de9aaf 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java @@ -642,9 +642,9 @@ private void setupPostMethod(Collection toExcludeList) String endBoundary = boundary + "--" + crNl; StringBuilder sb = new StringBuilder(); toExcludeList.forEach(sfn -> { - sb.append(boundary).append(crNl); - sb.append(contentDisposition); - sb.append(sfn).append(crNl); + sb.append(boundary).append(crNl) + .append(contentDisposition) + .append(sfn).append(crNl); }); sb.append(endBoundary); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java index ca6a8fd51797..520d9d6ef9e4 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java @@ -1237,16 +1237,16 @@ private String getPartName(PartKeyInfo partKeyInfo, String volName, String partFileName = OzoneFSUtils.getFileName(partKeyInfo.getPartName()); StringBuilder fullKeyPartName = new StringBuilder(); - fullKeyPartName.append(OZONE_URI_DELIMITER); - fullKeyPartName.append(volName); - fullKeyPartName.append(OZONE_URI_DELIMITER); - fullKeyPartName.append(buckName); + fullKeyPartName.append(OZONE_URI_DELIMITER) + .append(volName) + .append(OZONE_URI_DELIMITER) + .append(buckName); if (StringUtils.isNotEmpty(parentDir)) { - fullKeyPartName.append(OZONE_URI_DELIMITER); - fullKeyPartName.append(parentDir); + fullKeyPartName.append(OZONE_URI_DELIMITER) + .append(parentDir); } - fullKeyPartName.append(OZONE_URI_DELIMITER); - fullKeyPartName.append(partFileName); + fullKeyPartName.append(OZONE_URI_DELIMITER) + .append(partFileName); return fullKeyPartName.toString(); } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java index 9dc01e54f997..74349feee3dc 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java @@ -604,9 +604,8 @@ public VolumeBucketId getVolumeBucketIdPairFSO(String fsoKey) throws IOException @Override public String getOzoneKey(String volume, String bucket, String key) { StringBuilder builder = new StringBuilder() - .append(OM_KEY_PREFIX).append(volume); - // TODO : Throw if the Bucket is null? - builder.append(OM_KEY_PREFIX).append(bucket); + .append(OM_KEY_PREFIX).append(volume) + .append(OM_KEY_PREFIX).append(bucket); // TODO : Throw if the Bucket is null? if (StringUtils.isNotBlank(key)) { builder.append(OM_KEY_PREFIX); if (!key.equals(OM_KEY_PREFIX)) { @@ -1769,11 +1768,11 @@ public String getOpenFileName(long volumeId, long bucketId, long parentID, String fileName, String clientId) { StringBuilder openKey = new StringBuilder(); - openKey.append(OM_KEY_PREFIX).append(volumeId); - openKey.append(OM_KEY_PREFIX).append(bucketId); - openKey.append(OM_KEY_PREFIX).append(parentID); - openKey.append(OM_KEY_PREFIX).append(fileName); - openKey.append(OM_KEY_PREFIX).append(clientId); + openKey.append(OM_KEY_PREFIX).append(volumeId) + .append(OM_KEY_PREFIX).append(bucketId) + .append(OM_KEY_PREFIX).append(parentID) + .append(OM_KEY_PREFIX).append(fileName) + .append(OM_KEY_PREFIX).append(clientId); return openKey.toString(); } @@ -1781,9 +1780,9 @@ public String getOpenFileName(long volumeId, long bucketId, public String getRenameKey(String volumeName, String bucketName, long objectID) { StringBuilder renameKey = new StringBuilder(); - renameKey.append(OM_KEY_PREFIX).append(volumeName); - renameKey.append(OM_KEY_PREFIX).append(bucketName); - renameKey.append(OM_KEY_PREFIX).append(objectID); + renameKey.append(OM_KEY_PREFIX).append(volumeName) + .append(OM_KEY_PREFIX).append(bucketName) + .append(OM_KEY_PREFIX).append(objectID); return renameKey.toString(); } @@ -1798,11 +1797,11 @@ public String getMultipartKey(long volumeId, long bucketId, long parentID, String fileName, String uploadId) { StringBuilder openKey = new StringBuilder(); - openKey.append(OM_KEY_PREFIX).append(volumeId); - openKey.append(OM_KEY_PREFIX).append(bucketId); - openKey.append(OM_KEY_PREFIX).append(parentID); - openKey.append(OM_KEY_PREFIX).append(fileName); - openKey.append(OM_KEY_PREFIX).append(uploadId); + openKey.append(OM_KEY_PREFIX).append(volumeId) + .append(OM_KEY_PREFIX).append(bucketId) + .append(OM_KEY_PREFIX).append(parentID) + .append(OM_KEY_PREFIX).append(fileName) + .append(OM_KEY_PREFIX).append(uploadId); return openKey.toString(); } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java index 84722ede6294..aa1402052f3a 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java @@ -197,8 +197,8 @@ public static OMPathInfoWithFSO verifyDirectoryKeysInPath( StringBuilder fullKeyPath = new StringBuilder(bucketKey); while (elements.hasNext()) { String fileName = elements.next().toString(); - fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX); - fullKeyPath.append(fileName); + fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX) + .append(fileName); if (!missing.isEmpty()) { // Add all the sub-dirs to the missing list except the leaf element. // For example, /vol1/buck1/a/b/c/d/e/f/file1.txt. diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java index 2f8562642982..e8a17d2e74fe 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java @@ -362,9 +362,9 @@ protected static void addDeletedKeys(Map auditMap, StringBuilder keys = new StringBuilder(); for (int i = 0; i < deletedKeyInfos.size(); i++) { OmKeyInfo key = deletedKeyInfos.get(i); - keys.append('{').append(KEY).append('=').append(key.getKeyName()).append(", "); - keys.append(DATA_SIZE).append('=').append(key.getDataSize()).append(", "); - keys.append(REPLICATION_CONFIG).append('=').append(key.getReplicationConfig()).append('}'); + keys.append('{').append(KEY).append('=').append(key.getKeyName()).append(", ") + .append(DATA_SIZE).append('=').append(key.getDataSize()).append(", ") + .append(REPLICATION_CONFIG).append('=').append(key.getReplicationConfig()).append('}'); if (i < deletedKeyInfos.size() - 1) { keys.append(", "); } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java index 32e567dd44b8..2fb0f56ae890 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java @@ -119,8 +119,8 @@ public void testWriteFormDataWithSstFile() throws IOException { OmRatisSnapshotProvider.writeFormData(connection, sstFiles); - sb.append(fileName).append(CR_NL); - sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append("--").append(CR_NL); + sb.append(fileName).append(CR_NL) + .append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append("--").append(CR_NL); assertEquals(sb.toString(), new String(outputStream.toByteArray(), StandardCharsets.UTF_8)); } @@ -141,8 +141,8 @@ public void testWriteFormDataWithoutSstFile() throws IOException { private static StringBuilder getStringBuilder() { StringBuilder sb = new StringBuilder(); - sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append(CR_NL); - sb.append(CONTENT_DISPOSITION); + sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append(CR_NL) + .append(CONTENT_DISPOSITION); return sb; } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java index 4abeaf25d0c8..164ccf661cce 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java @@ -790,8 +790,8 @@ private String createDirKey(List dirs, int depth) { for (int i = 0; i < depth; i++) { String dirName = RandomStringUtils.secure().nextAlphabetic(5); dirs.add(dirName); - buf.append(OzoneConsts.OM_KEY_PREFIX); - buf.append(dirName); + buf.append(OzoneConsts.OM_KEY_PREFIX) + .append(dirName); } return buf.toString(); } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java index 93622ef70ba4..236da45a64bb 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java @@ -162,8 +162,8 @@ protected long checkIntermediatePaths(Path keyPath) throws Exception { StringBuilder fullKeyPath = new StringBuilder(bucketKey); while (elements.hasNext()) { String fileName = elements.next().toString(); - fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX); - fullKeyPath.append(fileName); + fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX) + .append(fileName); String dbNodeName = omMetadataManager.getOzonePathKey(volumeId, omBucketInfo.getObjectID(), lastKnownParentId, fileName); OmDirectoryInfo omDirInfo = omMetadataManager.getDirectoryTable(). diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java index 677a722bfe17..adf3184d7edb 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java @@ -504,9 +504,9 @@ private & Versioned> String annotationOf( Class annotationClass, V applyBeforeVersion) { StringBuilder annotation = new StringBuilder(); - annotation.append('@').append(annotationClass.getName()).append('('); - annotation.append("processingPhase = ").append(phase); - annotation.append(", requestType = ").append(reqType.name()); + annotation.append('@').append(annotationClass.getName()).append('(') + .append("processingPhase = ").append(phase) + .append(", requestType = ").append(reqType.name()); if (applyBeforeVersion != null) { annotation.append(", applyBefore = ").append(applyBeforeVersion.name()); } @@ -548,11 +548,11 @@ private StringBuilder buildMethodSignature( for (String modifier : modifiers) { signature.append(modifier).append(' '); } - signature.append(returnType).append(' '); - signature.append("validatorMethod("); - signature.append(createParameterList(paramspecs)); - signature.append(") "); - signature.append(createThrowsClause(exceptions)); + signature.append(returnType).append(' ') + .append("validatorMethod(") + .append(createParameterList(paramspecs)) + .append(") ") + .append(createThrowsClause(exceptions)); return signature.append(" {"); } diff --git a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java index 4acf198c687c..329e5e24c247 100644 --- a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java +++ b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java @@ -30,17 +30,13 @@ public class TableNamingStrategy extends DefaultGeneratorStrategy { @Override public String getJavaClassName(Definition definition, Mode mode) { if (definition instanceof TableDefinition && mode == Mode.DEFAULT) { - StringBuilder result = new StringBuilder(); - - result.append(StringUtils.toCamelCase( + return new StringBuilder().append(StringUtils.toCamelCase( definition.getOutputName() .replace(' ', '_') .replace('-', '_') - .replace('.', '_') - )); - - result.append("Table"); - return result.toString(); + .replace('.', '_'))) + .append("Table") + .toString(); } else { return super.getJavaClassName(definition, mode); } diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java index 233a001400ed..9c989c9f4320 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java @@ -116,8 +116,8 @@ public static String createSignatureBase( LOG.error("DateTime Header not found."); throw S3_AUTHINFO_CREATION_ERROR; } - strToSign.append(signatureInfo.getDateTime()).append(NEWLINE); - strToSign.append(credentialScope).append(NEWLINE); + strToSign.append(signatureInfo.getDateTime()).append(NEWLINE) + .append(credentialScope).append(NEWLINE); String canonicalRequest = buildCanonicalRequest( scheme, @@ -175,12 +175,12 @@ public static String buildCanonicalRequest( StringBuilder canonicalHeaders = new StringBuilder(); for (String header : StringUtils.split(signedHeaders, ';')) { - canonicalHeaders.append(header.toLowerCase()); - canonicalHeaders.append(':'); + canonicalHeaders.append(header.toLowerCase()) + .append(':'); if (headers.containsKey(header)) { String headerValue = headers.get(header); - canonicalHeaders.append(headerValue); - canonicalHeaders.append(NEWLINE); + canonicalHeaders.append(headerValue) + .append(NEWLINE); // Set for testing purpose only to skip date and host validation. try { @@ -306,10 +306,9 @@ private static String getQueryParamString( if (result.length() > 0) { result.append('&'); } - result.append(urlEncode(p)); - result.append('='); - - result.append(urlEncode(queryMap.get(p))); + result.append(urlEncode(p)) + .append('=') + .append(urlEncode(queryMap.get(p))); } } return result.toString();