diff --git a/dev-support/pmd/pmd-ruleset.xml b/dev-support/pmd/pmd-ruleset.xml
index d03a463def62..e3afd8e3081f 100644
--- a/dev-support/pmd/pmd-ruleset.xml
+++ b/dev-support/pmd/pmd-ruleset.xml
@@ -39,6 +39,7 @@
+
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java
index 7ee2a5160540..7141a65306dd 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/BlockID.java
@@ -93,8 +93,8 @@ public String toString() {
public void appendTo(StringBuilder sb) {
containerBlockID.appendTo(sb);
- sb.append(" bcsId: ").append(blockCommitSequenceId);
- sb.append(" replicaIndex: ").append(replicaIndex);
+ sb.append(" bcsId: ").append(blockCommitSequenceId)
+ .append(" replicaIndex: ").append(replicaIndex);
}
@JsonIgnore
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java
index cff7d73dcd13..d168fdc5a4d7 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java
@@ -538,20 +538,19 @@ public int hashCode() {
@Override
public String toString() {
final StringBuilder b =
- new StringBuilder(getClass().getSimpleName()).append('{');
- b.append(" Id: ").append(id.getId());
- b.append(", Nodes: [");
+ new StringBuilder(getClass().getSimpleName()).append('{')
+ .append(" Id: ").append(id.getId())
+ .append(", Nodes: [");
for (DatanodeDetails datanodeDetails : nodeStatus.keySet()) {
- b.append(" {").append(datanodeDetails);
- b.append(", ReplicaIndex: ").append(this.getReplicaIndex(datanodeDetails)).append("},");
- }
- b.append(']');
- b.append(", ReplicationConfig: ").append(replicationConfig);
- b.append(", State:").append(getPipelineState());
- b.append(", leaderId:").append(leaderId != null ? leaderId.toString() : "");
- b.append(", CreationTimestamp").append(getCreationTimestamp()
- .atZone(ZoneId.systemDefault()));
- b.append('}');
+ b.append(" {").append(datanodeDetails)
+ .append(", ReplicaIndex: ").append(this.getReplicaIndex(datanodeDetails)).append("},");
+ }
+ b.append(']')
+ .append(", ReplicationConfig: ").append(replicationConfig)
+ .append(", State:").append(getPipelineState())
+ .append(", leaderId:").append(leaderId != null ? leaderId.toString() : "")
+ .append(", CreationTimestamp").append(getCreationTimestamp().atZone(ZoneId.systemDefault()))
+ .append('}');
return b.toString();
}
diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java
index 7b0fda00e130..8e94ba01eca8 100644
--- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java
+++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/container/common/helpers/BlockData.java
@@ -299,8 +299,8 @@ public String toString() {
public void appendTo(StringBuilder sb) {
sb.append("[blockId=");
blockID.appendTo(sb);
- sb.append(", size=").append(size);
- sb.append(']');
+ sb.append(", size=").append(size)
+ .append(']');
}
public long getBlockGroupLength() {
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java
index 39525eb08938..490772e366c4 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java
@@ -1330,13 +1330,13 @@ public static String smProtoToString(RaftGroupId gid,
if (containerController != null) {
String location = containerController.getContainerLocation(contId);
- builder.append(", container path=");
- builder.append(location);
+ builder.append(", container path=")
+ .append(location);
}
} catch (Exception t) {
LOG.info("smProtoToString failed", t);
- builder.append("smProtoToString failed with ");
- builder.append(t.getMessage());
+ builder.append("smProtoToString failed with ")
+ .append(t.getMessage());
}
return builder.toString();
}
diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java
index 8574909b386e..bc8040b24bfc 100644
--- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java
+++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReplicateContainerCommand.java
@@ -157,13 +157,13 @@ public ReplicationCommandPriority getPriority() {
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
- sb.append(getType());
- sb.append(": cmdID: ").append(getId())
+ sb.append(getType())
+ .append(": cmdID: ").append(getId())
.append(", encodedToken: \"").append(getEncodedToken()).append('"')
.append(", term: ").append(getTerm())
- .append(", deadlineMsSinceEpoch: ").append(getDeadline());
- sb.append(", containerId=").append(getContainerID());
- sb.append(", replicaIndex=").append(getReplicaIndex());
+ .append(", deadlineMsSinceEpoch: ").append(getDeadline())
+ .append(", containerId=").append(getContainerID())
+ .append(", replicaIndex=").append(getReplicaIndex());
if (targetDatanode != null) {
sb.append(", targetNode=").append(targetDatanode);
} else {
diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java
index 5a26a7b569e9..c4bfdf2c5b9a 100644
--- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java
+++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/diskbalancer/TestDiskBalancerService.java
@@ -218,8 +218,8 @@ private String generateVolumeLocation(String base, int volumeCount) {
StringBuilder sb = new StringBuilder();
for (int i = 0; i < volumeCount; i++) {
- sb.append(base).append("/vol").append(i);
- sb.append(',');
+ sb.append(base).append("/vol").append(i)
+ .append(',');
}
return sb.substring(0, sb.length() - 1);
}
diff --git a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java
index a6f6dba51753..056503abb743 100644
--- a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java
+++ b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java
@@ -480,19 +480,16 @@ protected byte[][] toArrays(ECChunk[] chunks) {
protected void dumpSetting() {
if (allowDump) {
StringBuilder sb = new StringBuilder("Erasure coder test settings:\n");
- sb.append(" numDataUnits=").append(numDataUnits);
- sb.append(" numParityUnits=").append(numParityUnits);
- sb.append(" chunkSize=").append(chunkSize).append('\n');
-
- sb.append(" erasedDataIndexes=").
- append(Arrays.toString(erasedDataIndexes));
- sb.append(" erasedParityIndexes=").
- append(Arrays.toString(erasedParityIndexes));
- sb.append(" usingDirectBuffer=").append(usingDirectBuffer);
- sb.append(" allowVerboseDump=").append(allowDump);
- sb.append('\n');
-
- System.out.println(sb.toString());
+ sb.append(" numDataUnits=").append(numDataUnits)
+ .append(" numParityUnits=").append(numParityUnits)
+ .append(" chunkSize=").append(chunkSize).append('\n')
+ .append(" erasedDataIndexes=").append(Arrays.toString(erasedDataIndexes))
+ .append(" erasedParityIndexes=").append(Arrays.toString(erasedParityIndexes))
+ .append(" usingDirectBuffer=").append(usingDirectBuffer)
+ .append(" allowVerboseDump=").append(allowDump)
+ .append('\n');
+
+ System.out.println(sb);
}
}
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java
index 658bc41825c1..f83b8fae1203 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopologyImpl.java
@@ -891,20 +891,20 @@ private int getAvailableNodesCount(String scope, List excludedScopes,
public String toString() {
// print max level
StringBuilder tree = new StringBuilder();
- tree.append("Level: ");
- tree.append(maxLevel);
- tree.append('\n');
+ tree.append("Level: ")
+ .append(maxLevel)
+ .append('\n');
netlock.readLock().lock();
try {
// print the number of leaves
int numOfLeaves = clusterTree.getNumOfLeaves();
- tree.append("Number of leaves:");
- tree.append(numOfLeaves);
- tree.append('\n');
+ tree.append("Number of leaves:")
+ .append(numOfLeaves)
+ .append('\n');
// print all nodes
for (int i = 0; i < numOfLeaves; i++) {
- tree.append(clusterTree.getLeaf(i).getNetworkFullPath());
- tree.append('\n');
+ tree.append(clusterTree.getLeaf(i).getNetworkFullPath())
+ .append('\n');
}
} finally {
netlock.readLock().unlock();
diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java
index b611541578ea..6316bfe95f91 100644
--- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java
+++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeletedBlock.java
@@ -49,10 +49,10 @@ public long getReplicatedSize() {
@Override
public String toString() {
StringBuilder sb = new StringBuilder(64);
- sb.append(" localID: ").append(blockID.getContainerBlockID().getLocalID());
- sb.append(" containerID: ").append(blockID.getContainerBlockID().getContainerID());
- sb.append(" size: ").append(size);
- sb.append(" replicatedSize: ").append(replicatedSize);
+ sb.append(" localID: ").append(blockID.getContainerBlockID().getLocalID())
+ .append(" containerID: ").append(blockID.getContainerBlockID().getContainerID())
+ .append(" size: ").append(size)
+ .append(" replicatedSize: ").append(replicatedSize);
return sb.toString();
}
}
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java
index e1cd6a9a85ab..d0725cdd91f4 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java
@@ -169,8 +169,8 @@ public static String smProtoToString(StateMachineLogEntryProto proto) {
SCMRatisRequestProto.parseFrom(proto.getLogData().asReadOnlyByteBuffer())));
} catch (Throwable ex) {
LOG.error("smProtoToString failed", ex);
- builder.append("smProtoToString failed with");
- builder.append(ex.getMessage());
+ builder.append("smProtoToString failed with")
+ .append(ex.getMessage());
}
return builder.toString();
}
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java
index 49b01ef44847..2ee85d9d0d93 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitorImpl.java
@@ -497,11 +497,11 @@ public Map> getContainersPendingReplication(DatanodeDe
private String replicaDetails(Collection replicas) {
StringBuilder sb = new StringBuilder();
- sb.append("Replicas{");
- sb.append(replicas.stream()
- .map(Object::toString)
- .collect(Collectors.joining(",")));
- sb.append('}');
+ sb.append("Replicas{")
+ .append(replicas.stream()
+ .map(Object::toString)
+ .collect(Collectors.joining(",")))
+ .append('}');
return sb.toString();
}
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java
index b161e0e84d76..0725cb7b71b9 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java
@@ -388,8 +388,8 @@ public List getContainerWithPipelineBatch(
try {
ContainerWithPipeline cp = getContainerWithPipelineCommon(containerID);
cpList.add(cp);
- strContainerIDs.append(ContainerID.valueOf(containerID).toString());
- strContainerIDs.append(',');
+ strContainerIDs.append(ContainerID.valueOf(containerID).toString())
+ .append(',');
} catch (IOException ex) {
AUDIT.logReadFailure(buildAuditMessageForFailure(
SCMAction.GET_CONTAINER_WITH_PIPELINE_BATCH,
diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java
index 51df2647b444..dd403f46a182 100644
--- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java
+++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java
@@ -280,12 +280,12 @@ private String constructCommandAuditMap(List cmds) {
auditMap.append('[');
for (SCMCommandProto cmd : cmds) {
if (cmd.getCommandType().equals(deleteBlocksCommand)) {
- auditMap.append("commandType: ").append(cmd.getCommandType());
- auditMap.append(" deleteTransactionsCount: ")
- .append(cmd.getDeleteBlocksCommandProto().getDeletedBlocksTransactionsCount());
- auditMap.append(" cmdID: ").append(cmd.getDeleteBlocksCommandProto().getCmdId());
- auditMap.append(" encodedToken: \"").append(cmd.getEncodedToken()).append('"');
- auditMap.append(" deadlineMsSinceEpoch: ").append(cmd.getDeadlineMsSinceEpoch());
+ auditMap.append("commandType: ").append(cmd.getCommandType())
+ .append(" deleteTransactionsCount: ")
+ .append(cmd.getDeleteBlocksCommandProto().getDeletedBlocksTransactionsCount())
+ .append(" cmdID: ").append(cmd.getDeleteBlocksCommandProto().getCmdId())
+ .append(" encodedToken: \"").append(cmd.getEncodedToken()).append('"')
+ .append(" deadlineMsSinceEpoch: ").append(cmd.getDeadlineMsSinceEpoch());
} else {
auditMap.append(TextFormat.shortDebugString(cmd));
}
diff --git a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java
index 2e0117269f09..014daaa8dcf6 100644
--- a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java
+++ b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/LambdaTestUtils.java
@@ -286,9 +286,9 @@ public int getInvocationCount() {
public String toString() {
final StringBuilder sb = new StringBuilder(
"FixedRetryInterval{");
- sb.append("interval=").append(intervalMillis);
- sb.append(", invocationCount=").append(invocationCount);
- sb.append('}');
+ sb.append("interval=").append(intervalMillis)
+ .append(", invocationCount=").append(invocationCount)
+ .append('}');
return sb.toString();
}
}
diff --git a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java
index c7a8873c695b..e5b7555870b4 100644
--- a/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java
+++ b/hadoop-hdds/test-utils/src/main/java/org/apache/ozone/test/TimedOutTestsListener.java
@@ -93,8 +93,8 @@ private static String buildThreadDump() {
Thread.State.WAITING.equals(thread.getState()) ?
"WAITING (on object monitor)" : thread.getState()));
for (StackTraceElement stackTraceElement : e.getValue()) {
- dump.append("\n at ");
- dump.append(stackTraceElement);
+ dump.append("\n at ")
+ .append(stackTraceElement);
}
dump.append('\n');
}
diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java
index 62f38da74d01..4db959ebf7d9 100644
--- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java
+++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerReportSubcommand.java
@@ -100,9 +100,9 @@ private String generateReport(List protos) {
String dn = DiskBalancerSubCommandUtil.getDatanodeHostAndIp(p.getNode());
StringBuilder header = new StringBuilder();
- header.append("Datanode: ").append(dn).append('\n');
- header.append("Aggregate VolumeDataDensity: ").
- append(p.getCurrentVolumeDensitySum()).append('\n');
+ header.append("Datanode: ").append(dn).append('\n')
+ .append("Aggregate VolumeDataDensity: ").append(p.getCurrentVolumeDensitySum())
+ .append('\n');
if (p.hasIdealUsage() && p.hasDiskBalancerConf()
&& p.getDiskBalancerConf().hasThreshold()) {
@@ -110,11 +110,11 @@ private String generateReport(List protos) {
double threshold = p.getDiskBalancerConf().getThreshold();
double lt = idealUsage - threshold / 100.0;
double ut = idealUsage + threshold / 100.0;
- header.append("IdealUsage: ").append(String.format("%.8f", idealUsage));
- header.append(" | Threshold: ").append(threshold).append('%');
- header.append(" | ThresholdRange: (").append(String.format("%.8f", lt));
- header.append(", ").append(String.format("%.8f", ut)).append(')').append('\n').append('\n');
- header.append("Volume Details:").append('\n');
+ header.append("IdealUsage: ").append(String.format("%.8f", idealUsage))
+ .append(" | Threshold: ").append(threshold).append('%')
+ .append(" | ThresholdRange: (").append(String.format("%.8f", lt))
+ .append(", ").append(String.format("%.8f", ut)).append(')').append('\n').append('\n')
+ .append("Volume Details:").append('\n');
}
formatBuilder.append("%s%n");
contentList.add(header.toString());
@@ -150,21 +150,20 @@ private String generateReport(List protos) {
}
}
- formatBuilder.append("%nNote:%n");
- formatBuilder.append(" - Aggregate VolumeDataDensity: Sum of per-volume density" +
- " (deviation from ideal); higher means more imbalance.%n");
- formatBuilder.append(" - IdealUsage: Target utilization ratio (0-1) when volumes" +
- " are evenly balanced.%n");
- formatBuilder.append(" - ThresholdRange: Acceptable deviation (percent); volumes within" +
- " IdealUsage +/- Threshold are considered balanced.%n");
- formatBuilder.append(" - VolumeDensity: Deviation of a particular volume's utilization from IdealUsage.%n");
- formatBuilder.append(" - Utilization: Ratio of actual used space to capacity (0-1) for a particular volume.%n");
- formatBuilder.append(" - TotalCapacity: Total volume capacity.%n");
- formatBuilder.append(" - UsedSpace: Ozone used space.%n");
- formatBuilder.append(" - Container Pre-AllocatedSpace: Space reserved for containers not yet written to disk.%n");
- formatBuilder.append(" - EffectiveUsedSpace: This is the actual used space of volume which is visible" +
- " to the diskBalancer : (ozoneCapacity minus ozoneAvailable) + containerPreAllocatedSpace + " +
- "move delta for source volume.%n");
+ formatBuilder.append("%nNote:%n")
+ .append(" - Aggregate VolumeDataDensity: Sum of per-volume density (deviation from ideal);")
+ .append(" higher means more imbalance.%n")
+ .append(" - IdealUsage: Target utilization ratio (0-1) when volumes are evenly balanced.%n")
+ .append(" - ThresholdRange: Acceptable deviation (percent); volumes within")
+ .append(" IdealUsage +/- Threshold are considered balanced.%n")
+ .append(" - VolumeDensity: Deviation of a particular volume's utilization from IdealUsage.%n")
+ .append(" - Utilization: Ratio of actual used space to capacity (0-1) for a particular volume.%n")
+ .append(" - TotalCapacity: Total volume capacity.%n")
+ .append(" - UsedSpace: Ozone used space.%n")
+ .append(" - Container Pre-AllocatedSpace: Space reserved for containers not yet written to disk.%n")
+ .append(" - EffectiveUsedSpace: This is the actual used space of volume which is visible")
+ .append(" to the diskBalancer : (ozoneCapacity minus ozoneAvailable) + containerPreAllocatedSpace + ")
+ .append("move delta for source volume.%n");
return String.format(formatBuilder.toString(), contentList.toArray(new String[0]));
}
diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java
index f84a5b9a78e2..bd98767f9972 100644
--- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java
+++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/DiskBalancerStatusSubcommand.java
@@ -131,13 +131,11 @@ private String generateStatus(List protos) {
contentList.add(estimatedTimeLeft >= 0 ? String.valueOf(estimatedTimeLeft) : "N/A");
}
- formatBuilder.append("%nNote:%n");
- formatBuilder.append(" - EstBytesToMove is calculated based on the target disk even state" +
- " with the configured threshold.%n");
- formatBuilder.append(" - EstTimeLeft is calculated based on EstimatedBytesToMove and configured" +
- " disk bandwidth.%n");
- formatBuilder.append(" - Both EstimatedBytes and EstTimeLeft could be non-zero while no containers" +
- " can be moved, especially when the configured threshold or disk capacity is too small.");
+ formatBuilder.append("%nNote:%n")
+ .append(" - EstBytesToMove is calculated based on the target disk even state with the configured threshold.%n")
+ .append(" - EstTimeLeft is calculated based on EstimatedBytesToMove and configured disk bandwidth.%n")
+ .append(" - Both EstimatedBytes and EstTimeLeft could be non-zero while no containers" +
+ " can be moved, especially when the configured threshold or disk capacity is too small.");
return String.format(formatBuilder.toString(),
contentList.toArray(new String[0]));
diff --git a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java
index c5477e0cb0ec..951624213ce1 100644
--- a/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java
+++ b/hadoop-ozone/cli-admin/src/main/java/org/apache/hadoop/ozone/admin/om/ListOpenFilesSubCommand.java
@@ -197,8 +197,8 @@ private void printOpenKeysList(ListOpenFilesResult res) {
private String getMessageString(ListOpenFilesResult res, List openFileList) {
StringBuilder sb = new StringBuilder();
sb.append(res.getTotalOpenKeyCount())
- .append(" total open files. Showing ");
- sb.append(openFileList.size())
+ .append(" total open files. Showing ")
+ .append(openFileList.size())
.append(" open files (limit ")
.append(limit)
.append(") under path prefix:\n ")
diff --git a/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java b/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java
index f598f40b9b73..f508ec349500 100644
--- a/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java
+++ b/hadoop-ozone/cli-debug/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java
@@ -210,8 +210,8 @@ private static String executeStatement(String dbName, String sql)
int cols = rsm.getColumnCount();
while (rs.next()) {
for (int index = 1; index <= cols; index++) {
- result.append(rs.getObject(index));
- result.append('\t');
+ result.append(rs.getObject(index))
+ .append('\t');
}
result.append('\n');
}
diff --git a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java
index 83f6c4d55267..f1b1a568e1e7 100644
--- a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java
+++ b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/ContainerUpgradeResult.java
@@ -89,31 +89,23 @@ public void success(long rowCount) {
@Override
public String toString() {
final StringBuilder stringBuilder = new StringBuilder();
- stringBuilder.append("Result:{");
- stringBuilder.append("containerID=");
- stringBuilder.append(originContainerData.getContainerID());
- stringBuilder.append(", originContainerSchemaVersion=");
- stringBuilder.append(
- ((KeyValueContainerData) originContainerData).getSchemaVersion());
+ stringBuilder.append("Result:{")
+ .append("containerID=")
+ .append(originContainerData.getContainerID())
+ .append(", originContainerSchemaVersion=")
+ .append(((KeyValueContainerData) originContainerData).getSchemaVersion());
if (newContainerData != null) {
- stringBuilder.append(", schemaV2ContainerFileBackupPath=");
- stringBuilder.append(backupContainerFilePath);
-
- stringBuilder.append(", newContainerSchemaVersion=");
- stringBuilder.append(
- ((KeyValueContainerData) newContainerData).getSchemaVersion());
-
- stringBuilder.append(", schemaV3ContainerFilePath=");
- stringBuilder.append(newContainerFilePath);
+ stringBuilder
+ .append(", schemaV2ContainerFileBackupPath=").append(backupContainerFilePath)
+ .append(", newContainerSchemaVersion=").append(((KeyValueContainerData) newContainerData).getSchemaVersion())
+ .append(", schemaV3ContainerFilePath=").append(newContainerFilePath);
}
- stringBuilder.append(", totalRow=");
- stringBuilder.append(totalRow);
- stringBuilder.append(", costMs=");
- stringBuilder.append(getCostMs());
- stringBuilder.append(", status=");
- stringBuilder.append(status);
- stringBuilder.append('}');
+ stringBuilder
+ .append(", totalRow=").append(totalRow)
+ .append(", costMs=").append(getCostMs())
+ .append(", status=").append(status)
+ .append('}');
return stringBuilder.toString();
}
diff --git a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java
index 067de25b3ba0..529233b6a90c 100644
--- a/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java
+++ b/hadoop-ozone/cli-repair/src/main/java/org/apache/hadoop/ozone/repair/datanode/schemaupgrade/VolumeUpgradeResult.java
@@ -86,27 +86,27 @@ public void fail(Exception exception) {
@Override
public String toString() {
final StringBuilder stringBuilder = new StringBuilder();
- stringBuilder.append("Result:{");
- stringBuilder.append("hddsRootDir=");
- stringBuilder.append(getHddsVolume().getHddsRootDir());
- stringBuilder.append(", resultList=");
+ stringBuilder.append("Result:{")
+ .append("hddsRootDir=")
+ .append(getHddsVolume().getHddsRootDir())
+ .append(", resultList=");
AtomicLong total = new AtomicLong(0L);
if (resultMap != null) {
resultMap.forEach((k, r) -> {
- stringBuilder.append(r.toString());
- stringBuilder.append('\n');
+ stringBuilder.append(r.toString())
+ .append('\n');
total.addAndGet(r.getTotalRow());
});
}
- stringBuilder.append(", totalRow=");
- stringBuilder.append(total.get());
- stringBuilder.append(", costMs=");
- stringBuilder.append(getCost());
- stringBuilder.append(", status=");
- stringBuilder.append(status);
+ stringBuilder.append(", totalRow=")
+ .append(total.get())
+ .append(", costMs=")
+ .append(getCost())
+ .append(", status=")
+ .append(status);
if (e != null) {
- stringBuilder.append(", Exception=");
- stringBuilder.append(e);
+ stringBuilder.append(", Exception=")
+ .append(e);
}
stringBuilder.append('}');
return stringBuilder.toString();
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java
index aee73248db70..fc96045eba3e 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OFSPath.java
@@ -188,15 +188,15 @@ public String toString() {
Objects.requireNonNull(authority, "authority == null");
StringBuilder sb = new StringBuilder();
if (!isMount()) {
- sb.append(volumeName);
- sb.append(OZONE_URI_DELIMITER);
+ sb.append(volumeName)
+ .append(OZONE_URI_DELIMITER);
if (!bucketName.isEmpty()) {
- sb.append(bucketName);
- sb.append(OZONE_URI_DELIMITER);
+ sb.append(bucketName)
+ .append(OZONE_URI_DELIMITER);
}
} else {
- sb.append(mountName);
- sb.append(OZONE_URI_DELIMITER);
+ sb.append(mountName)
+ .append(OZONE_URI_DELIMITER);
}
if (!keyName.isEmpty()) {
sb.append(keyName);
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java
index ddf62a45ed8f..e2477a4cef10 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyLocationInfoGroup.java
@@ -185,16 +185,16 @@ void addAll(long versionToAdd, List locationInfoList) {
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
- sb.append("version:").append(version).append(' ');
- sb.append("isMultipartKey:").append(isMultipartKey).append(' ');
+ sb.append("version:").append(version).append(' ')
+ .append("isMultipartKey:").append(isMultipartKey).append(' ');
for (List kliList : locationVersionMap.values()) {
for (OmKeyLocationInfo kli: kliList) {
- sb.append("conID ").append(kli.getContainerID());
- sb.append(' ');
- sb.append("locID ").append(kli.getLocalID());
- sb.append(' ');
- sb.append("bcsID ").append(kli.getBlockCommitSequenceId());
- sb.append(" || ");
+ sb.append("conID ").append(kli.getContainerID())
+ .append(' ')
+ .append("locID ").append(kli.getLocalID())
+ .append(' ')
+ .append("bcsID ").append(kli.getBlockCommitSequenceId())
+ .append(" || ");
}
}
return sb.toString();
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java
index 9893f3c5c031..de2403e26b6c 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java
@@ -308,8 +308,8 @@ public static String getParentDir(@Nonnull String keyName) {
public static String appendFileNameToKeyPath(String keyName,
String fileName) {
StringBuilder newToKeyName = new StringBuilder(keyName);
- newToKeyName.append(OZONE_URI_DELIMITER);
- newToKeyName.append(fileName);
+ newToKeyName.append(OZONE_URI_DELIMITER)
+ .append(fileName);
return newToKeyName.toString();
}
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java
index 99ea12b1cd67..e0d7ebe37b63 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatus.java
@@ -133,8 +133,8 @@ public int hashCode() {
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
- sb.append(getClass().getSimpleName());
- sb.append('{');
+ sb.append(getClass().getSimpleName())
+ .append('{');
if (keyInfo == null) {
sb.append("");
} else {
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java
index 5d07f331086b..3b4b68d6705e 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFileStatusLight.java
@@ -131,8 +131,8 @@ public int hashCode() {
@Override
public String toString() {
StringBuilder sb = new StringBuilder();
- sb.append(getClass().getSimpleName());
- sb.append('{');
+ sb.append(getClass().getSimpleName())
+ .append('{');
if (keyInfo == null) {
sb.append("");
} else {
diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
index 89aba42511da..f17603d7ec1f 100644
--- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
+++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/RadixTree.java
@@ -178,8 +178,8 @@ public List> getLongestPrefixPath(String path) {
public static String radixPathToString(List> path) {
StringBuilder sb = new StringBuilder();
for (RadixNode n : path) {
- sb.append(n.getName());
- sb.append(n.getName().equals(PATH_DELIMITER) ? "" : PATH_DELIMITER);
+ sb.append(n.getName())
+ .append(n.getName().equals(PATH_DELIMITER) ? "" : PATH_DELIMITER);
}
return sb.toString();
}
diff --git a/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java b/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java
index 4fddb9054be7..a1d1e79731b9 100644
--- a/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java
+++ b/hadoop-ozone/freon/src/main/java/org/apache/hadoop/ozone/freon/ProgressBar.java
@@ -184,15 +184,15 @@ private void printProgressBar(PrintStream stream, long value) {
for (int j = 0; j < (100 - percent) / shrinkTimes; j++) {
sb.append(' ');
}
- sb.append("| ");
- sb.append(value).append('/').append(maxValue);
+ sb.append("| ")
+ .append(value).append('/').append(maxValue);
long timeInSec = TimeUnit.SECONDS.convert(
System.nanoTime() - startTime, TimeUnit.NANOSECONDS);
String timeToPrint = String.format("%d:%02d:%02d", timeInSec / 3600,
(timeInSec % 3600) / 60, timeInSec % 60);
- sb.append(" Time: ").append(timeToPrint);
- sb.append("| ");
- sb.append(realTimeMessage);
+ sb.append(" Time: ").append(timeToPrint)
+ .append("| ")
+ .append(realTimeMessage);
stream.print(sb.toString());
}
}
diff --git a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java
index d8fbd1f1ce39..730cd4ac2e92 100644
--- a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java
+++ b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v1/AbstractS3SDKV1Tests.java
@@ -1534,10 +1534,10 @@ private void completeMPU(String keyName, String uploadId, List complet
StringBuilder completionXml = new StringBuilder();
completionXml.append("\n");
for (PartETag part : completedParts) {
- completionXml.append(" \n");
- completionXml.append(" ").append(part.getPartNumber()).append("\n");
- completionXml.append(" ").append(stripQuotes(part.getETag())).append("\n");
- completionXml.append(" \n");
+ completionXml.append(" \n")
+ .append(" ").append(part.getPartNumber()).append("\n")
+ .append(" ").append(stripQuotes(part.getETag())).append("\n")
+ .append(" \n");
}
completionXml.append("");
diff --git a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java
index 35a53dd328bb..4b165175f226 100644
--- a/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java
+++ b/hadoop-ozone/integration-test-s3/src/test/java/org/apache/hadoop/ozone/s3/awssdk/v2/AbstractS3SDKV2Tests.java
@@ -1280,10 +1280,10 @@ private String buildCompleteMultipartUploadXml(List parts) {
StringBuilder xml = new StringBuilder();
xml.append("\n");
for (CompletedPart part : parts) {
- xml.append(" \n");
- xml.append(" ").append(part.partNumber()).append("\n");
- xml.append(" ").append(stripQuotes(part.eTag())).append("\n");
- xml.append(" \n");
+ xml.append(" \n")
+ .append(" ").append(part.partNumber()).append("\n")
+ .append(" ").append(stripQuotes(part.eTag())).append("\n")
+ .append(" \n");
}
xml.append("");
return xml.toString();
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java
index 198874fac8da..d87838287e42 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/conf/TestConfigurationFieldsBase.java
@@ -434,13 +434,13 @@ public void testCompareConfigurationClassAgainstXml() {
LOG.info("({} member variables)\n", configurationMemberVariables.size());
StringBuilder xmlErrorMsg = new StringBuilder();
for (Class c : configurationClasses) {
- xmlErrorMsg.append(c);
- xmlErrorMsg.append(' ');
+ xmlErrorMsg.append(c)
+ .append(' ');
}
- xmlErrorMsg.append("has ");
- xmlErrorMsg.append(missingXmlSize);
- xmlErrorMsg.append(" variables missing in ");
- xmlErrorMsg.append(xmlFilename);
+ xmlErrorMsg.append("has ")
+ .append(missingXmlSize)
+ .append(" variables missing in ")
+ .append(xmlFilename);
LOG.error(xmlErrorMsg.toString());
if (missingXmlSize == 0) {
LOG.info(" (None)");
@@ -482,10 +482,10 @@ public void testCompareXmlAgainstConfigurationClass() {
LOG.info("File {} ({} properties)", xmlFilename, xmlKeyValueMap.size());
StringBuilder configErrorMsg = new StringBuilder();
- configErrorMsg.append(xmlFilename);
- configErrorMsg.append(" has ");
- configErrorMsg.append(missingConfigSize);
- configErrorMsg.append(" properties missing in");
+ configErrorMsg.append(xmlFilename)
+ .append(" has ")
+ .append(missingConfigSize)
+ .append(" properties missing in");
Arrays.stream(configurationClasses)
.forEach(c -> configErrorMsg.append(" ").append(c));
LOG.info(configErrorMsg.toString());
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java
index 31d00b7228ae..cb8045e0e6e5 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestSCMDbCheckpointServlet.java
@@ -252,9 +252,9 @@ private void setupPostMethod(Collection toExcludeList)
String endBoundary = boundary + "--" + crNl;
StringBuilder sb = new StringBuilder();
toExcludeList.forEach(sfn -> {
- sb.append(boundary).append(crNl);
- sb.append(contentDisposition);
- sb.append(sfn).append(crNl);
+ sb.append(boundary).append(crNl)
+ .append(contentDisposition)
+ .append(sfn).append(crNl);
});
sb.append(endBoundary);
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
index ec1d522b721e..fa801804dd82 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
@@ -4505,9 +4505,9 @@ private void doMultipartUpload(OzoneBucket bucket, String keyName, byte val,
// Combine all parts data, and check is it matching with get key data.
String part1 = new String(data, UTF_8);
String part2 = new String(data, UTF_8);
- sb.append(part1);
- sb.append(part2);
- sb.append(part3);
+ sb.append(part1)
+ .append(part2)
+ .append(part3);
assertEquals(sb.toString(), new String(fileContent, UTF_8));
OmKeyArgs keyArgs = new OmKeyArgs.Builder()
diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java
index ac12189b1f62..8acc63de9aaf 100644
--- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java
+++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMDbCheckpointServlet.java
@@ -642,9 +642,9 @@ private void setupPostMethod(Collection toExcludeList)
String endBoundary = boundary + "--" + crNl;
StringBuilder sb = new StringBuilder();
toExcludeList.forEach(sfn -> {
- sb.append(boundary).append(crNl);
- sb.append(contentDisposition);
- sb.append(sfn).append(crNl);
+ sb.append(boundary).append(crNl)
+ .append(contentDisposition)
+ .append(sfn).append(crNl);
});
sb.append(endBoundary);
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
index ca6a8fd51797..520d9d6ef9e4 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java
@@ -1237,16 +1237,16 @@ private String getPartName(PartKeyInfo partKeyInfo, String volName,
String partFileName = OzoneFSUtils.getFileName(partKeyInfo.getPartName());
StringBuilder fullKeyPartName = new StringBuilder();
- fullKeyPartName.append(OZONE_URI_DELIMITER);
- fullKeyPartName.append(volName);
- fullKeyPartName.append(OZONE_URI_DELIMITER);
- fullKeyPartName.append(buckName);
+ fullKeyPartName.append(OZONE_URI_DELIMITER)
+ .append(volName)
+ .append(OZONE_URI_DELIMITER)
+ .append(buckName);
if (StringUtils.isNotEmpty(parentDir)) {
- fullKeyPartName.append(OZONE_URI_DELIMITER);
- fullKeyPartName.append(parentDir);
+ fullKeyPartName.append(OZONE_URI_DELIMITER)
+ .append(parentDir);
}
- fullKeyPartName.append(OZONE_URI_DELIMITER);
- fullKeyPartName.append(partFileName);
+ fullKeyPartName.append(OZONE_URI_DELIMITER)
+ .append(partFileName);
return fullKeyPartName.toString();
}
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java
index 9dc01e54f997..74349feee3dc 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java
@@ -604,9 +604,8 @@ public VolumeBucketId getVolumeBucketIdPairFSO(String fsoKey) throws IOException
@Override
public String getOzoneKey(String volume, String bucket, String key) {
StringBuilder builder = new StringBuilder()
- .append(OM_KEY_PREFIX).append(volume);
- // TODO : Throw if the Bucket is null?
- builder.append(OM_KEY_PREFIX).append(bucket);
+ .append(OM_KEY_PREFIX).append(volume)
+ .append(OM_KEY_PREFIX).append(bucket); // TODO : Throw if the Bucket is null?
if (StringUtils.isNotBlank(key)) {
builder.append(OM_KEY_PREFIX);
if (!key.equals(OM_KEY_PREFIX)) {
@@ -1769,11 +1768,11 @@ public String getOpenFileName(long volumeId, long bucketId,
long parentID, String fileName,
String clientId) {
StringBuilder openKey = new StringBuilder();
- openKey.append(OM_KEY_PREFIX).append(volumeId);
- openKey.append(OM_KEY_PREFIX).append(bucketId);
- openKey.append(OM_KEY_PREFIX).append(parentID);
- openKey.append(OM_KEY_PREFIX).append(fileName);
- openKey.append(OM_KEY_PREFIX).append(clientId);
+ openKey.append(OM_KEY_PREFIX).append(volumeId)
+ .append(OM_KEY_PREFIX).append(bucketId)
+ .append(OM_KEY_PREFIX).append(parentID)
+ .append(OM_KEY_PREFIX).append(fileName)
+ .append(OM_KEY_PREFIX).append(clientId);
return openKey.toString();
}
@@ -1781,9 +1780,9 @@ public String getOpenFileName(long volumeId, long bucketId,
public String getRenameKey(String volumeName, String bucketName,
long objectID) {
StringBuilder renameKey = new StringBuilder();
- renameKey.append(OM_KEY_PREFIX).append(volumeName);
- renameKey.append(OM_KEY_PREFIX).append(bucketName);
- renameKey.append(OM_KEY_PREFIX).append(objectID);
+ renameKey.append(OM_KEY_PREFIX).append(volumeName)
+ .append(OM_KEY_PREFIX).append(bucketName)
+ .append(OM_KEY_PREFIX).append(objectID);
return renameKey.toString();
}
@@ -1798,11 +1797,11 @@ public String getMultipartKey(long volumeId, long bucketId,
long parentID, String fileName,
String uploadId) {
StringBuilder openKey = new StringBuilder();
- openKey.append(OM_KEY_PREFIX).append(volumeId);
- openKey.append(OM_KEY_PREFIX).append(bucketId);
- openKey.append(OM_KEY_PREFIX).append(parentID);
- openKey.append(OM_KEY_PREFIX).append(fileName);
- openKey.append(OM_KEY_PREFIX).append(uploadId);
+ openKey.append(OM_KEY_PREFIX).append(volumeId)
+ .append(OM_KEY_PREFIX).append(bucketId)
+ .append(OM_KEY_PREFIX).append(parentID)
+ .append(OM_KEY_PREFIX).append(fileName)
+ .append(OM_KEY_PREFIX).append(uploadId);
return openKey.toString();
}
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java
index 84722ede6294..aa1402052f3a 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/file/OMFileRequest.java
@@ -197,8 +197,8 @@ public static OMPathInfoWithFSO verifyDirectoryKeysInPath(
StringBuilder fullKeyPath = new StringBuilder(bucketKey);
while (elements.hasNext()) {
String fileName = elements.next().toString();
- fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX);
- fullKeyPath.append(fileName);
+ fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX)
+ .append(fileName);
if (!missing.isEmpty()) {
// Add all the sub-dirs to the missing list except the leaf element.
// For example, /vol1/buck1/a/b/c/d/e/f/file1.txt.
diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java
index 2f8562642982..e8a17d2e74fe 100644
--- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java
+++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeysDeleteRequest.java
@@ -362,9 +362,9 @@ protected static void addDeletedKeys(Map auditMap,
StringBuilder keys = new StringBuilder();
for (int i = 0; i < deletedKeyInfos.size(); i++) {
OmKeyInfo key = deletedKeyInfos.get(i);
- keys.append('{').append(KEY).append('=').append(key.getKeyName()).append(", ");
- keys.append(DATA_SIZE).append('=').append(key.getDataSize()).append(", ");
- keys.append(REPLICATION_CONFIG).append('=').append(key.getReplicationConfig()).append('}');
+ keys.append('{').append(KEY).append('=').append(key.getKeyName()).append(", ")
+ .append(DATA_SIZE).append('=').append(key.getDataSize()).append(", ")
+ .append(REPLICATION_CONFIG).append('=').append(key.getReplicationConfig()).append('}');
if (i < deletedKeyInfos.size() - 1) {
keys.append(", ");
}
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java
index 32e567dd44b8..2fb0f56ae890 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis_snapshot/TestOmRatisSnapshotProvider.java
@@ -119,8 +119,8 @@ public void testWriteFormDataWithSstFile() throws IOException {
OmRatisSnapshotProvider.writeFormData(connection, sstFiles);
- sb.append(fileName).append(CR_NL);
- sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append("--").append(CR_NL);
+ sb.append(fileName).append(CR_NL)
+ .append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append("--").append(CR_NL);
assertEquals(sb.toString(),
new String(outputStream.toByteArray(), StandardCharsets.UTF_8));
}
@@ -141,8 +141,8 @@ public void testWriteFormDataWithoutSstFile() throws IOException {
private static StringBuilder getStringBuilder() {
StringBuilder sb = new StringBuilder();
- sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append(CR_NL);
- sb.append(CONTENT_DISPOSITION);
+ sb.append("--").append(MULTIPART_FORM_DATA_BOUNDARY).append(CR_NL)
+ .append(CONTENT_DISPOSITION);
return sb;
}
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java
index 4abeaf25d0c8..164ccf661cce 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/file/TestOMDirectoryCreateRequestWithFSO.java
@@ -790,8 +790,8 @@ private String createDirKey(List dirs, int depth) {
for (int i = 0; i < depth; i++) {
String dirName = RandomStringUtils.secure().nextAlphabetic(5);
dirs.add(dirName);
- buf.append(OzoneConsts.OM_KEY_PREFIX);
- buf.append(dirName);
+ buf.append(OzoneConsts.OM_KEY_PREFIX)
+ .append(dirName);
}
return buf.toString();
}
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java
index 93622ef70ba4..236da45a64bb 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequestWithFSO.java
@@ -162,8 +162,8 @@ protected long checkIntermediatePaths(Path keyPath) throws Exception {
StringBuilder fullKeyPath = new StringBuilder(bucketKey);
while (elements.hasNext()) {
String fileName = elements.next().toString();
- fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX);
- fullKeyPath.append(fileName);
+ fullKeyPath.append(OzoneConsts.OM_KEY_PREFIX)
+ .append(fileName);
String dbNodeName = omMetadataManager.getOzonePathKey(volumeId,
omBucketInfo.getObjectID(), lastKnownParentId, fileName);
OmDirectoryInfo omDirInfo = omMetadataManager.getDirectoryTable().
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java
index 677a722bfe17..adf3184d7edb 100644
--- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/validation/TestOMValidatorProcessor.java
@@ -504,9 +504,9 @@ private & Versioned> String annotationOf(
Class> annotationClass,
V applyBeforeVersion) {
StringBuilder annotation = new StringBuilder();
- annotation.append('@').append(annotationClass.getName()).append('(');
- annotation.append("processingPhase = ").append(phase);
- annotation.append(", requestType = ").append(reqType.name());
+ annotation.append('@').append(annotationClass.getName()).append('(')
+ .append("processingPhase = ").append(phase)
+ .append(", requestType = ").append(reqType.name());
if (applyBeforeVersion != null) {
annotation.append(", applyBefore = ").append(applyBeforeVersion.name());
}
@@ -548,11 +548,11 @@ private StringBuilder buildMethodSignature(
for (String modifier : modifiers) {
signature.append(modifier).append(' ');
}
- signature.append(returnType).append(' ');
- signature.append("validatorMethod(");
- signature.append(createParameterList(paramspecs));
- signature.append(") ");
- signature.append(createThrowsClause(exceptions));
+ signature.append(returnType).append(' ')
+ .append("validatorMethod(")
+ .append(createParameterList(paramspecs))
+ .append(") ")
+ .append(createThrowsClause(exceptions));
return signature.append(" {");
}
diff --git a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java
index 4acf198c687c..329e5e24c247 100644
--- a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java
+++ b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/TableNamingStrategy.java
@@ -30,17 +30,13 @@ public class TableNamingStrategy extends DefaultGeneratorStrategy {
@Override
public String getJavaClassName(Definition definition, Mode mode) {
if (definition instanceof TableDefinition && mode == Mode.DEFAULT) {
- StringBuilder result = new StringBuilder();
-
- result.append(StringUtils.toCamelCase(
+ return new StringBuilder().append(StringUtils.toCamelCase(
definition.getOutputName()
.replace(' ', '_')
.replace('-', '_')
- .replace('.', '_')
- ));
-
- result.append("Table");
- return result.toString();
+ .replace('.', '_')))
+ .append("Table")
+ .toString();
} else {
return super.getJavaClassName(definition, mode);
}
diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java
index 233a001400ed..9c989c9f4320 100644
--- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java
+++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/StringToSignProducer.java
@@ -116,8 +116,8 @@ public static String createSignatureBase(
LOG.error("DateTime Header not found.");
throw S3_AUTHINFO_CREATION_ERROR;
}
- strToSign.append(signatureInfo.getDateTime()).append(NEWLINE);
- strToSign.append(credentialScope).append(NEWLINE);
+ strToSign.append(signatureInfo.getDateTime()).append(NEWLINE)
+ .append(credentialScope).append(NEWLINE);
String canonicalRequest = buildCanonicalRequest(
scheme,
@@ -175,12 +175,12 @@ public static String buildCanonicalRequest(
StringBuilder canonicalHeaders = new StringBuilder();
for (String header : StringUtils.split(signedHeaders, ';')) {
- canonicalHeaders.append(header.toLowerCase());
- canonicalHeaders.append(':');
+ canonicalHeaders.append(header.toLowerCase())
+ .append(':');
if (headers.containsKey(header)) {
String headerValue = headers.get(header);
- canonicalHeaders.append(headerValue);
- canonicalHeaders.append(NEWLINE);
+ canonicalHeaders.append(headerValue)
+ .append(NEWLINE);
// Set for testing purpose only to skip date and host validation.
try {
@@ -306,10 +306,9 @@ private static String getQueryParamString(
if (result.length() > 0) {
result.append('&');
}
- result.append(urlEncode(p));
- result.append('=');
-
- result.append(urlEncode(queryMap.get(p)));
+ result.append(urlEncode(p))
+ .append('=')
+ .append(urlEncode(queryMap.get(p)));
}
}
return result.toString();