Skip to content

Commit 86a00a3

Browse files
author
Zhe Zhang
committed
Fix compilation.
1 parent 1d2f8ed commit 86a00a3

File tree

5 files changed

+8
-12
lines changed

5 files changed

+8
-12
lines changed

hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/ExternalBlockReader.java

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@
2424
import org.apache.hadoop.classification.InterfaceAudience;
2525
import org.apache.hadoop.fs.ReadOption;
2626
import org.apache.hadoop.hdfs.shortcircuit.ClientMmap;
27+
import org.apache.hadoop.util.DataChecksum;
2728

2829
/**
2930
* An ExternalBlockReader uses pluggable ReplicaAccessor objects to read from
@@ -117,4 +118,9 @@ public ClientMmap getClientMmap(EnumSet<ReadOption> opts) {
117118
// For now, pluggable ReplicaAccessors do not support zero-copy.
118119
return null;
119120
}
121+
122+
@Override
123+
public DataChecksum getDataChecksum() {
124+
return null;
125+
}
120126
}

hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ErasureCodingZoneManager.java

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ ErasureCodingZone getErasureCodingZone(INodesInPath iip) throws IOException {
8484
new ArrayList<XAttr>(0)
8585
: inode.getXAttrFeature().getXAttrs();
8686
for (XAttr xAttr : xAttrs) {
87-
if (XATTR_ERASURECODING_ZONE.equals(XAttrHelper.getPrefixName(xAttr))) {
87+
if (XATTR_ERASURECODING_ZONE.equals(XAttrHelper.getPrefixedName(xAttr))) {
8888
ByteArrayInputStream bIn=new ByteArrayInputStream(xAttr.getValue());
8989
DataInputStream dIn=new DataInputStream(bIn);
9090
String ecPolicyName = WritableUtils.readString(dIn);

hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeFsck.java

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -258,12 +258,8 @@ public void blockIdCK(String blockId) {
258258
NumberReplicas numberReplicas= bm.countNodes(blockInfo);
259259
out.println("Block Id: " + blockId);
260260
out.println("Block belongs to: "+iNode.getFullPathName());
261-
<<<<<<< HEAD
262261
out.println("No. of Expected Replica: " +
263-
bm.getExpectedReplicaNum(bc, blockInfo));
264-
=======
265-
out.println("No. of Expected Replica: " + blockInfo.getReplication());
266-
>>>>>>> cbb249534aa72ff6c290c4f99766415aeea9d6f5
262+
bm.getExpectedReplicaNum(blockInfo));
267263
out.println("No. of live Replica: " + numberReplicas.liveReplicas());
268264
out.println("No. of excess Replica: " + numberReplicas.excessReplicas());
269265
out.println("No. of stale Replica: " +

hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/FileDiffList.java

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,10 +22,7 @@
2222

2323
import org.apache.hadoop.hdfs.protocol.Block;
2424
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfo;
25-
<<<<<<< HEAD
2625
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous;
27-
=======
28-
>>>>>>> cbb249534aa72ff6c290c4f99766415aeea9d6f5
2926
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
3027
import org.apache.hadoop.hdfs.server.namenode.INode;
3128
import org.apache.hadoop.hdfs.server.namenode.INode.BlocksMapUpdateInfo;

hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileTruncate.java

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -54,10 +54,7 @@
5454
import org.apache.hadoop.hdfs.protocol.HdfsConstants.SafeModeAction;
5555
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
5656
import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
57-
<<<<<<< HEAD
5857
import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous;
59-
=======
60-
>>>>>>> cbb249534aa72ff6c290c4f99766415aeea9d6f5
6158
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants;
6259
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
6360
import org.apache.hadoop.security.UserGroupInformation;

0 commit comments

Comments
 (0)