forked from Qortal/qortal
More miscellaneous refactoring
This commit is contained in:
parent
09e783fbf6
commit
8dac3ebf96
@ -119,7 +119,7 @@ public class ArbitraryDataFile {
|
||||
return arbitraryDataFile;
|
||||
|
||||
} catch (IOException e) {
|
||||
LOGGER.error("Couldn't compute digest for DataFile");
|
||||
LOGGER.error("Couldn't compute digest for ArbitraryDataFile");
|
||||
}
|
||||
}
|
||||
return null;
|
||||
@ -187,7 +187,7 @@ public class ArbitraryDataFile {
|
||||
// Validate the file size
|
||||
long fileSize = Files.size(path);
|
||||
if (fileSize > MAX_FILE_SIZE) {
|
||||
LOGGER.error(String.format("DataFile is too large: %d bytes (max size: %d bytes)", fileSize, MAX_FILE_SIZE));
|
||||
LOGGER.error(String.format("ArbitraryDataFile is too large: %d bytes (max size: %d bytes)", fileSize, MAX_FILE_SIZE));
|
||||
return ArbitraryDataFile.ValidationResult.FILE_TOO_LARGE;
|
||||
}
|
||||
|
||||
@ -468,7 +468,7 @@ public class ArbitraryDataFile {
|
||||
return Crypto.digest(fileContent);
|
||||
|
||||
} catch (IOException e) {
|
||||
LOGGER.error("Couldn't compute digest for DataFile");
|
||||
LOGGER.error("Couldn't compute digest for ArbitraryDataFile");
|
||||
}
|
||||
}
|
||||
return null;
|
||||
|
@ -166,7 +166,7 @@ public class ArbitraryDataReader {
|
||||
private void fetchFromFileHash() {
|
||||
// Load data file directly from the hash
|
||||
ArbitraryDataFile arbitraryDataFile = ArbitraryDataFile.fromHash58(resourceId);
|
||||
// Set filePath to the location of the DataFile
|
||||
// Set filePath to the location of the ArbitraryDataFile
|
||||
this.filePath = Paths.get(arbitraryDataFile.getFilePath());
|
||||
}
|
||||
|
||||
@ -233,7 +233,7 @@ public class ArbitraryDataReader {
|
||||
if (!Arrays.equals(arbitraryDataFile.digest(), digest)) {
|
||||
throw new IllegalStateException("Unable to validate complete file hash");
|
||||
}
|
||||
// Set filePath to the location of the DataFile
|
||||
// Set filePath to the location of the ArbitraryDataFile
|
||||
this.filePath = Paths.get(arbitraryDataFile.getFilePath());
|
||||
}
|
||||
|
||||
@ -247,7 +247,7 @@ public class ArbitraryDataReader {
|
||||
AES.decryptFile("AES", aesKey, this.filePath.toString(), this.unencryptedPath.toString());
|
||||
|
||||
// Replace filePath pointer with the encrypted file path
|
||||
// Don't delete the original DataFile, as this is handled in the cleanup phase
|
||||
// Don't delete the original ArbitraryDataFile, as this is handled in the cleanup phase
|
||||
this.filePath = this.unencryptedPath;
|
||||
|
||||
} catch (NoSuchAlgorithmException | InvalidAlgorithmParameterException | NoSuchPaddingException
|
||||
|
@ -370,13 +370,13 @@ public class ArbitraryDataManager extends Thread {
|
||||
ArbitraryDataFile arbitraryDataFile = ArbitraryDataFile.fromHash(hash);
|
||||
if (chunkHashes != null && chunkHashes.length > 0) {
|
||||
arbitraryDataFile.addChunkHashes(chunkHashes);
|
||||
for (ArbitraryDataFileChunk dataFileChunk : arbitraryDataFile.getChunks()) {
|
||||
if (dataFileChunk.exists()) {
|
||||
hashes.add(dataFileChunk.getHash());
|
||||
//LOGGER.info("Added hash {}", dataFileChunk.getHash58());
|
||||
for (ArbitraryDataFileChunk chunk : arbitraryDataFile.getChunks()) {
|
||||
if (chunk.exists()) {
|
||||
hashes.add(chunk.getHash());
|
||||
//LOGGER.info("Added hash {}", chunk.getHash58());
|
||||
}
|
||||
else {
|
||||
LOGGER.info("Couldn't add hash {} because it doesn't exist", dataFileChunk.getHash58());
|
||||
LOGGER.info("Couldn't add hash {} because it doesn't exist", chunk.getHash58());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -133,7 +133,7 @@ public class HSQLDBArbitraryRepository implements ArbitraryRepository {
|
||||
arbitraryTransactionData.setData(dataHash);
|
||||
arbitraryTransactionData.setDataType(DataType.DATA_HASH);
|
||||
|
||||
// Create DataFile
|
||||
// Create ArbitraryDataFile
|
||||
ArbitraryDataFile arbitraryDataFile = new ArbitraryDataFile(rawData);
|
||||
|
||||
// Verify that the data file is valid, and that it matches the expected hash
|
||||
|
Loading…
x
Reference in New Issue
Block a user