Hi - I am currently using Calculator to create the SHA-1 checksum of my input files. Input files are between 1K and 1G. Everything was working fine, till I got to the larger files. With a larger file it fails instantly with the following exception:
With CRC-32 Adler-32 it is successful after ~25 seconds. It also fails immediately with MD5.
Does anyone have a work around for successful generation of SHA-1 for large files, or quicker generation from the other algorithms?
I know I could probably just add memory, but it would be a shame to do that for just this function when the actual file contents processes just fine.
Thanks!
Code:
2015/11/19 14:08:55 - Calculator.0 - ERROR (version 5.3.0.0-213, build 1 from 2015-02-02_12-17-08 by buildguy) : UnexpectedError:
2015/11/19 14:08:55 - Calculator.0 - ERROR (version 5.3.0.0-213, build 1 from 2015-02-02_12-17-08 by buildguy) : java.lang.OutOfMemoryError: Java heap space
2015/11/19 14:08:55 - Calculator.0 - at org.pentaho.di.core.row.ValueDataUtil.createChecksum(ValueDataUtil.java:310)
2015/11/19 14:08:55 - Calculator.0 - at org.pentaho.di.trans.steps.calculator.Calculator.calcFields(Calculator.java:394)
2015/11/19 14:08:55 - Calculator.0 - at org.pentaho.di.trans.steps.calculator.Calculator.processRow(Calculator.java:162)
2015/11/19 14:08:55 - Calculator.0 - at org.pentaho.di.trans.step.RunThread.run(RunThread.java:62)
2015/11/19 14:08:55 - Calculator.0 - at java.lang.Thread.run(Unknown Source)
Does anyone have a work around for successful generation of SHA-1 for large files, or quicker generation from the other algorithms?
I know I could probably just add memory, but it would be a shame to do that for just this function when the actual file contents processes just fine.
Thanks!