parent
07c052d822
commit
a32644bdb7
19
build.gradle
19
build.gradle
|
@ -55,6 +55,7 @@ configurations.all {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
allprojects { Project proj ->
|
allprojects { Project proj ->
|
||||||
apply plugin: 'com.google.osdetector'
|
apply plugin: 'com.google.osdetector'
|
||||||
|
|
||||||
|
@ -161,3 +162,21 @@ allprojects { Project proj ->
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
task aggregatedJavadocs(type: Javadoc, description: 'Generate javadocs from all child projects as if it was a single project', group: 'Documentation') {
|
||||||
|
subprojects.each { proj ->
|
||||||
|
proj.tasks.withType(Javadoc).each { javadocTask ->
|
||||||
|
logger.quiet("Adding javadoc for project " + proj.name)
|
||||||
|
source += javadocTask.source
|
||||||
|
classpath += javadocTask.classpath
|
||||||
|
excludes += javadocTask.excludes
|
||||||
|
includes += javadocTask.includes
|
||||||
|
}
|
||||||
|
}
|
||||||
|
destinationDir = file("$buildDir/docs/javadoc")
|
||||||
|
title = "$project.name $version API"
|
||||||
|
options.author true
|
||||||
|
options.links 'http://docs.oracle.com/javase/8/docs/api/'
|
||||||
|
options.addStringOption('Xdoclint:none', '-quiet')
|
||||||
|
}
|
|
@ -106,7 +106,7 @@ public class Text extends BinaryComparable implements WritableComparable<BinaryC
|
||||||
if (position < 0)
|
if (position < 0)
|
||||||
return -1; // duh.
|
return -1; // duh.
|
||||||
|
|
||||||
ByteBuffer bb = ByteBuffer.wrap(bytes).position(position);
|
ByteBuffer bb = (ByteBuffer) ByteBuffer.wrap(bytes).position(position);
|
||||||
return bytesToCodePoint(bb.slice());
|
return bytesToCodePoint(bb.slice());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1462,7 +1462,7 @@ public class SameDiff extends SDBaseOps {
|
||||||
* A special case of {@link #fit()}.
|
* A special case of {@link #fit()}.
|
||||||
*
|
*
|
||||||
* @param iter The iterator to train the SameDiff instance with
|
* @param iter The iterator to train the SameDiff instance with
|
||||||
* @param numEpochs The number of epochs for training. Must be > 0
|
* @param numEpochs The number of epochs for training. Must be > 0
|
||||||
* @param validationIter The DataSetIterator to use for validation (null to skip validation)
|
* @param validationIter The DataSetIterator to use for validation (null to skip validation)
|
||||||
* @param validationFrequency The frequency with which to run validation. 1 is every epoch, 2 is every other, etc.
|
* @param validationFrequency The frequency with which to run validation. 1 is every epoch, 2 is every other, etc.
|
||||||
* @param listeners Additional listeners to use during this operation
|
* @param listeners Additional listeners to use during this operation
|
||||||
|
@ -1479,7 +1479,7 @@ public class SameDiff extends SDBaseOps {
|
||||||
* A special case of {@link #fit()}.
|
* A special case of {@link #fit()}.
|
||||||
*
|
*
|
||||||
* @param iter The iterator to train the SameDiff instance with
|
* @param iter The iterator to train the SameDiff instance with
|
||||||
* @param numEpochs The number of epochs for training. Must be > 0
|
* @param numEpochs The number of epochs for training. Must be > 0
|
||||||
* @param listeners Additional listeners to use during this operation
|
* @param listeners Additional listeners to use during this operation
|
||||||
* @return a {@link History} object containing the history information for this training operation
|
* @return a {@link History} object containing the history information for this training operation
|
||||||
* (evaluations specified in the {@link TrainingConfig}, loss values, and timing information).
|
* (evaluations specified in the {@link TrainingConfig}, loss values, and timing information).
|
||||||
|
@ -1497,7 +1497,7 @@ public class SameDiff extends SDBaseOps {
|
||||||
* A special case of {@link #fit()}.
|
* A special case of {@link #fit()}.
|
||||||
*
|
*
|
||||||
* @param iter The iterator to train the SameDiff instance with
|
* @param iter The iterator to train the SameDiff instance with
|
||||||
* @param numEpochs The number of epochs for training. Must be > 0
|
* @param numEpochs The number of epochs for training. Must be > 0
|
||||||
* @param validationIter The MultiDataSetIterator to use for validation (null to skip validation)
|
* @param validationIter The MultiDataSetIterator to use for validation (null to skip validation)
|
||||||
* @param validationFrequency The frequency with which to run validation. 1 is every epoch, 2 is every other, etc.
|
* @param validationFrequency The frequency with which to run validation. 1 is every epoch, 2 is every other, etc.
|
||||||
* @param listeners Additional listeners to use during this operation
|
* @param listeners Additional listeners to use during this operation
|
||||||
|
@ -1514,7 +1514,7 @@ public class SameDiff extends SDBaseOps {
|
||||||
* A special case of {@link #fit()}.
|
* A special case of {@link #fit()}.
|
||||||
*
|
*
|
||||||
* @param iter The iterator to train the SameDiff instance with
|
* @param iter The iterator to train the SameDiff instance with
|
||||||
* @param numEpochs The number of epochs for training. Must be > 0
|
* @param numEpochs The number of epochs for training. Must be > 0
|
||||||
* @param listeners Additional listeners to use during this operation
|
* @param listeners Additional listeners to use during this operation
|
||||||
* @return a {@link History} object containing the history information for this training operation
|
* @return a {@link History} object containing the history information for this training operation
|
||||||
* (evaluations specified in the {@link TrainingConfig}, loss values, and timing information).
|
* (evaluations specified in the {@link TrainingConfig}, loss values, and timing information).
|
||||||
|
@ -3036,7 +3036,6 @@ public class SameDiff extends SDBaseOps {
|
||||||
* See also: {@link VariableType}
|
* See also: {@link VariableType}
|
||||||
*
|
*
|
||||||
* @param variables Variables to convert to constants
|
* @param variables Variables to convert to constants
|
||||||
* @return The (now constant) SDVariables
|
|
||||||
*/
|
*/
|
||||||
public void convertToConstants(List<SDVariable> variables) {
|
public void convertToConstants(List<SDVariable> variables) {
|
||||||
if (variables.size() == 0)
|
if (variables.size() == 0)
|
||||||
|
@ -3201,7 +3200,7 @@ public class SameDiff extends SDBaseOps {
|
||||||
* For example, {@code z(float) = x(float)+y(float)}, changing both x and y to double results in {@code z(double) = x(double)+y(double)}
|
* For example, {@code z(float) = x(float)+y(float)}, changing both x and y to double results in {@code z(double) = x(double)+y(double)}
|
||||||
* without doing anything to change z's datatype directly (z datatype is inferred from x + y + add op).<br>
|
* without doing anything to change z's datatype directly (z datatype is inferred from x + y + add op).<br>
|
||||||
* ARRAY type SDVariables cannot be converted directly, as their datatypes are determined by the function +
|
* ARRAY type SDVariables cannot be converted directly, as their datatypes are determined by the function +
|
||||||
* input datatypes.<b>
|
* input datatypes.<br>
|
||||||
* Note that this method should be used with caution: incorrect datatype modifications may leave your network
|
* Note that this method should be used with caution: incorrect datatype modifications may leave your network
|
||||||
* in an incorrect state. For example, {@code op(x(float),y(float)) -> op(x(double),y(float))} may not be
|
* in an incorrect state. For example, {@code op(x(float),y(float)) -> op(x(double),y(float))} may not be
|
||||||
* supported by all ops.
|
* supported by all ops.
|
||||||
|
|
|
@ -382,7 +382,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Cast the array to a new datatype - for example, Integer -> Float<br>
|
* Cast the array to a new datatype - for example, Integer -> Float<br>
|
||||||
*
|
*
|
||||||
* @param arg Input variable to cast (NDARRAY type)
|
* @param arg Input variable to cast (NDARRAY type)
|
||||||
* @param datatype Datatype to cast to
|
* @param datatype Datatype to cast to
|
||||||
|
@ -393,7 +393,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Cast the array to a new datatype - for example, Integer -> Float<br>
|
* Cast the array to a new datatype - for example, Integer -> Float<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param arg Input variable to cast (NDARRAY type)
|
* @param arg Input variable to cast (NDARRAY type)
|
||||||
|
@ -654,7 +654,7 @@ public class SDBaseOps {
|
||||||
*
|
*
|
||||||
* @param x Input variable (NUMERIC type)
|
* @param x Input variable (NUMERIC type)
|
||||||
* @param partitions 1D input with values 0 to numPartitions-1 (INT type)
|
* @param partitions 1D input with values 0 to numPartitions-1 (INT type)
|
||||||
* @param numPartitions Number of partitions, >= 1
|
* @param numPartitions Number of partitions, >= 1
|
||||||
*/
|
*/
|
||||||
public SDVariable[] dynamicPartition(SDVariable x, SDVariable partitions, int numPartitions) {
|
public SDVariable[] dynamicPartition(SDVariable x, SDVariable partitions, int numPartitions) {
|
||||||
SDValidation.validateNumerical("dynamicPartition", "x", x);
|
SDValidation.validateNumerical("dynamicPartition", "x", x);
|
||||||
|
@ -676,7 +676,7 @@ public class SDBaseOps {
|
||||||
* @param names names May be null. Arrays of names for the output variables.
|
* @param names names May be null. Arrays of names for the output variables.
|
||||||
* @param x Input variable (NUMERIC type)
|
* @param x Input variable (NUMERIC type)
|
||||||
* @param partitions 1D input with values 0 to numPartitions-1 (INT type)
|
* @param partitions 1D input with values 0 to numPartitions-1 (INT type)
|
||||||
* @param numPartitions Number of partitions, >= 1
|
* @param numPartitions Number of partitions, >= 1
|
||||||
*/
|
*/
|
||||||
public SDVariable[] dynamicPartition(String[] names, SDVariable x, SDVariable partitions,
|
public SDVariable[] dynamicPartition(String[] names, SDVariable x, SDVariable partitions,
|
||||||
int numPartitions) {
|
int numPartitions) {
|
||||||
|
@ -689,7 +689,7 @@ public class SDBaseOps {
|
||||||
/**
|
/**
|
||||||
* Dynamically merge the specified input arrays into a single array, using the specified indices<br>
|
* Dynamically merge the specified input arrays into a single array, using the specified indices<br>
|
||||||
*
|
*
|
||||||
* @param indices Indices to use when merging. Must be >= 1, same length as input variables (INT type)
|
* @param indices Indices to use when merging. Must be >= 1, same length as input variables (INT type)
|
||||||
* @param x Input variables. (NUMERIC type)
|
* @param x Input variables. (NUMERIC type)
|
||||||
* @return output Merged output variable (NUMERIC type)
|
* @return output Merged output variable (NUMERIC type)
|
||||||
*/
|
*/
|
||||||
|
@ -705,7 +705,7 @@ public class SDBaseOps {
|
||||||
* Dynamically merge the specified input arrays into a single array, using the specified indices<br>
|
* Dynamically merge the specified input arrays into a single array, using the specified indices<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param indices Indices to use when merging. Must be >= 1, same length as input variables (INT type)
|
* @param indices Indices to use when merging. Must be >= 1, same length as input variables (INT type)
|
||||||
* @param x Input variables. (NUMERIC type)
|
* @param x Input variables. (NUMERIC type)
|
||||||
* @return output Merged output variable (NUMERIC type)
|
* @return output Merged output variable (NUMERIC type)
|
||||||
*/
|
*/
|
||||||
|
@ -943,7 +943,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than operation: elementwise x > y<br>
|
* Greater than operation: elementwise x > y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -957,7 +957,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than operation: elementwise x > y<br>
|
* Greater than operation: elementwise x > y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -973,7 +973,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than operation: elementwise x > y<br>
|
* Greater than operation: elementwise x > y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -993,7 +993,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than operation: elementwise x > y<br>
|
* Greater than operation: elementwise x > y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1015,7 +1015,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than or equals operation: elementwise x >= y<br>
|
* Greater than or equals operation: elementwise x >= y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1029,7 +1029,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than or equals operation: elementwise x >= y<br>
|
* Greater than or equals operation: elementwise x >= y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1045,7 +1045,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than or equal to operation: elementwise x >= y<br>
|
* Greater than or equal to operation: elementwise x >= y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1065,7 +1065,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Greater than or equal to operation: elementwise x >= y<br>
|
* Greater than or equal to operation: elementwise x >= y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1232,7 +1232,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than operation: elementwise x < y<br>
|
* Less than operation: elementwise x < y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1246,7 +1246,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than operation: elementwise x < y<br>
|
* Less than operation: elementwise x < y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1262,7 +1262,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than operation: elementwise x < y<br>
|
* Less than operation: elementwise x < y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1282,7 +1282,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than operation: elementwise x < y<br>
|
* Less than operation: elementwise x < y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1304,7 +1304,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than or equals operation: elementwise x <= y<br>
|
* Less than or equals operation: elementwise x <= y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1318,7 +1318,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than or equals operation: elementwise x <= y<br>
|
* Less than or equals operation: elementwise x <= y<br>
|
||||||
*
|
*
|
||||||
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
* Return boolean array with values true where satisfied, or false otherwise.<br>
|
||||||
*
|
*
|
||||||
|
@ -1334,7 +1334,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than or equal to operation: elementwise x <= y<br>
|
* Less than or equal to operation: elementwise x <= y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -1354,7 +1354,7 @@ public class SDBaseOps {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Less than or equal to operation: elementwise x <= y<br>
|
* Less than or equal to operation: elementwise x <= y<br>
|
||||||
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
* If x and y arrays have equal shape, the output shape is the same as these inputs.<br>
|
||||||
*
|
*
|
||||||
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
* Note: supports broadcasting if x and y have different shapes and are broadcastable.<br>
|
||||||
|
@ -3590,7 +3590,7 @@ public class SDBaseOps {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
||||||
* Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)<br>
|
* {@code Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)}<br>
|
||||||
*
|
*
|
||||||
* @param lengths Lengths of the sequences (NUMERIC type)
|
* @param lengths Lengths of the sequences (NUMERIC type)
|
||||||
* @param maxLen Maximum sequence length
|
* @param maxLen Maximum sequence length
|
||||||
|
@ -3604,7 +3604,7 @@ public class SDBaseOps {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
||||||
* Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)<br>
|
* {@code Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)}<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param lengths Lengths of the sequences (NUMERIC type)
|
* @param lengths Lengths of the sequences (NUMERIC type)
|
||||||
|
@ -3620,7 +3620,7 @@ public class SDBaseOps {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
||||||
* Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)<br>
|
* {@code Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)}<br>
|
||||||
*
|
*
|
||||||
* @param lengths Lengths of the sequences (NUMERIC type)
|
* @param lengths Lengths of the sequences (NUMERIC type)
|
||||||
* @param maxLen Maximum sequence length (INT type)
|
* @param maxLen Maximum sequence length (INT type)
|
||||||
|
@ -3635,7 +3635,7 @@ public class SDBaseOps {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
* Generate a sequence mask (with values 0 or 1) based on the specified lengths <br>
|
||||||
* Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)<br>
|
* {@code Specifically, out[i, ..., k, j] = (j < lengths[i, ..., k] ? 1.0 : 0.0)}<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param lengths Lengths of the sequences (NUMERIC type)
|
* @param lengths Lengths of the sequences (NUMERIC type)
|
||||||
|
@ -3761,7 +3761,7 @@ public class SDBaseOps {
|
||||||
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
||||||
* [b]<br>
|
* [b]<br>
|
||||||
* [e]<br>
|
* [e]<br>
|
||||||
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
||||||
*
|
*
|
||||||
* @param input input Variable to get subset of (NUMERIC type)
|
* @param input input Variable to get subset of (NUMERIC type)
|
||||||
* @param begin Beginning index. Must be same length as rank of input array (Size: AtLeast(min=1))
|
* @param begin Beginning index. Must be same length as rank of input array (Size: AtLeast(min=1))
|
||||||
|
@ -3783,7 +3783,7 @@ public class SDBaseOps {
|
||||||
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
||||||
* [b]<br>
|
* [b]<br>
|
||||||
* [e]<br>
|
* [e]<br>
|
||||||
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param input input Variable to get subset of (NUMERIC type)
|
* @param input input Variable to get subset of (NUMERIC type)
|
||||||
|
@ -3807,7 +3807,7 @@ public class SDBaseOps {
|
||||||
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
||||||
* [b]<br>
|
* [b]<br>
|
||||||
* [e]<br>
|
* [e]<br>
|
||||||
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
||||||
*
|
*
|
||||||
* @param input input Variable to get subset of (NUMERIC type)
|
* @param input input Variable to get subset of (NUMERIC type)
|
||||||
* @param begin Beginning index. Must be same length as rank of input array (INT type)
|
* @param begin Beginning index. Must be same length as rank of input array (INT type)
|
||||||
|
@ -3829,7 +3829,7 @@ public class SDBaseOps {
|
||||||
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
* then slice(input, begin=[0,1], size=[2,1] will return:<br>
|
||||||
* [b]<br>
|
* [b]<br>
|
||||||
* [e]<br>
|
* [e]<br>
|
||||||
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
* Note that for each dimension i, begin[i] + size[i] <= input.size(i)<br>
|
||||||
*
|
*
|
||||||
* @param name name May be null. Name for the output variable
|
* @param name name May be null. Name for the output variable
|
||||||
* @param input input Variable to get subset of (NUMERIC type)
|
* @param input input Variable to get subset of (NUMERIC type)
|
||||||
|
|
|
@ -176,7 +176,7 @@ public class Evaluation extends BaseEvaluation<Evaluation> {
|
||||||
* Constructor to use for top N accuracy
|
* Constructor to use for top N accuracy
|
||||||
*
|
*
|
||||||
* @param labels Labels for the classes (may be null)
|
* @param labels Labels for the classes (may be null)
|
||||||
* @param topN Value to use for top N accuracy calculation (<=1: standard accuracy). Note that with top N
|
* @param topN Value to use for top N accuracy calculation (<=1: standard accuracy). Note that with top N
|
||||||
* accuracy, an example is considered 'correct' if the probability for the true class is one of the
|
* accuracy, an example is considered 'correct' if the probability for the true class is one of the
|
||||||
* highest N values
|
* highest N values
|
||||||
*/
|
*/
|
||||||
|
@ -1173,7 +1173,7 @@ public class Evaluation extends BaseEvaluation<Evaluation> {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* False Alarm Rate (FAR) reflects rate of misclassified to classified records
|
* False Alarm Rate (FAR) reflects rate of misclassified to classified records
|
||||||
* <a href="http://ro.ecu.edu.au/cgi/viewcontent.cgi?article=1058&context=isw">http://ro.ecu.edu.au/cgi/viewcontent.cgi?article=1058&context=isw</a><br>
|
* {@link }http://ro.ecu.edu.au/cgi/viewcontent.cgi?article=1058&context=isw}<br>
|
||||||
* Note: value returned will differ depending on number of classes and settings.<br>
|
* Note: value returned will differ depending on number of classes and settings.<br>
|
||||||
* 1. For binary classification, if the positive class is set (via default value of 1, via constructor,
|
* 1. For binary classification, if the positive class is set (via default value of 1, via constructor,
|
||||||
* or via {@link #setBinaryPositiveClass(Integer)}), the returned value will be for the specified positive class
|
* or via {@link #setBinaryPositiveClass(Integer)}), the returned value will be for the specified positive class
|
||||||
|
|
|
@ -34,7 +34,7 @@ public class BlasException extends Error {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Principal constructor - error message & error code
|
* Principal constructor - error message & error code
|
||||||
* @param message the error message to put into the Exception
|
* @param message the error message to put into the Exception
|
||||||
* @param errorCode the library error number
|
* @param errorCode the library error number
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -28,15 +28,15 @@ public interface Lapack {
|
||||||
* LU decomposiiton of a matrix
|
* LU decomposiiton of a matrix
|
||||||
* Factorize a matrix A
|
* Factorize a matrix A
|
||||||
*
|
*
|
||||||
* The matrix A is overridden by the L & U combined.
|
* The matrix A is overridden by the L & U combined.
|
||||||
* The permutation results are returned directly as a vector. To
|
* The permutation results are returned directly as a vector. To
|
||||||
* create the permutation matrix use getPFactor method
|
* create the permutation matrix use getPFactor method
|
||||||
* To split out the L & U matrix use getLFactor and getUFactor methods
|
* To split out the L & U matrix use getLFactor and getUFactor methods
|
||||||
*
|
*
|
||||||
* getrf = triangular factorization (TRF) of a general matrix (GE)
|
* getrf = triangular factorization (TRF) of a general matrix (GE)
|
||||||
*
|
*
|
||||||
* @param A the input matrix, it will be overwritten with the factors
|
* @param A the input matrix, it will be overwritten with the factors
|
||||||
* @returns Permutation array
|
* @return Permutation array
|
||||||
* @throws Error - with a message to indicate failure (usu. bad params)
|
* @throws Error - with a message to indicate failure (usu. bad params)
|
||||||
*/
|
*/
|
||||||
INDArray getrf(INDArray A);
|
INDArray getrf(INDArray A);
|
||||||
|
@ -53,7 +53,7 @@ public interface Lapack {
|
||||||
* matrix Q and an upper triangular R matrix
|
* matrix Q and an upper triangular R matrix
|
||||||
*
|
*
|
||||||
* @param A the input matrix, it will be overwritten with the factors
|
* @param A the input matrix, it will be overwritten with the factors
|
||||||
* @param The R array if null R is not returned
|
* @param R The R array if null R is not returned
|
||||||
* @throws Error - with a message to indicate failure (usu. bad params)
|
* @throws Error - with a message to indicate failure (usu. bad params)
|
||||||
*/
|
*/
|
||||||
void geqrf(INDArray A, INDArray R);
|
void geqrf(INDArray A, INDArray R);
|
||||||
|
@ -71,8 +71,7 @@ public interface Lapack {
|
||||||
* lower L ( or upper U ) triangular matrix
|
* lower L ( or upper U ) triangular matrix
|
||||||
*
|
*
|
||||||
* @param A the input matrix, it will be overwritten with the factors
|
* @param A the input matrix, it will be overwritten with the factors
|
||||||
* @param whether to return the upper (false) or lower factor
|
* @param lower whether to return the upper (false) or lower factor
|
||||||
* @returns Permutation array
|
|
||||||
* @throws Error - with a message to indicate failure (usu. bad params)
|
* @throws Error - with a message to indicate failure (usu. bad params)
|
||||||
*/
|
*/
|
||||||
void potrf(INDArray A, boolean lower);
|
void potrf(INDArray A, boolean lower);
|
||||||
|
@ -122,7 +121,7 @@ public interface Lapack {
|
||||||
*
|
*
|
||||||
* @param M - the size of the permutation matrix ( usu. the # rows in factored matrix )
|
* @param M - the size of the permutation matrix ( usu. the # rows in factored matrix )
|
||||||
* @param ipiv - the vector returned from a refactoring
|
* @param ipiv - the vector returned from a refactoring
|
||||||
* @returned the square permutation matrix - size is the M x M
|
* @return the square permutation matrix - size is the M x M
|
||||||
*/
|
*/
|
||||||
INDArray getPFactor(int M, INDArray ipiv);
|
INDArray getPFactor(int M, INDArray ipiv);
|
||||||
|
|
||||||
|
@ -131,8 +130,8 @@ public interface Lapack {
|
||||||
* extracts the L (lower triangular) matrix from the LU factor result
|
* extracts the L (lower triangular) matrix from the LU factor result
|
||||||
* L will be the same dimensions as A
|
* L will be the same dimensions as A
|
||||||
*
|
*
|
||||||
* @param A - the combined L & U matrices returned from factorization
|
* @param A - the combined L & U matrices returned from factorization
|
||||||
* @returned the lower triangular with unit diagonal
|
* @return the lower triangular with unit diagonal
|
||||||
*/
|
*/
|
||||||
INDArray getLFactor(INDArray A);
|
INDArray getLFactor(INDArray A);
|
||||||
|
|
||||||
|
@ -141,8 +140,8 @@ public interface Lapack {
|
||||||
* extracts the U (upper triangular) matrix from the LU factor result
|
* extracts the U (upper triangular) matrix from the LU factor result
|
||||||
* U will be n x n matrix where n = num cols in A
|
* U will be n x n matrix where n = num cols in A
|
||||||
*
|
*
|
||||||
* @param A - the combined L & U matrices returned from factorization
|
* @param A - the combined L & U matrices returned from factorization
|
||||||
* @returned the upper triangular matrix
|
* @return the upper triangular matrix
|
||||||
*/
|
*/
|
||||||
INDArray getUFactor(INDArray A);
|
INDArray getUFactor(INDArray A);
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.nd4j.linalg.api.ndarray.INDArray;
|
||||||
public interface Level1 {
|
public interface Level1 {
|
||||||
/**
|
/**
|
||||||
* computes a vector-vector dot product.
|
* computes a vector-vector dot product.
|
||||||
* @param n
|
* @param N
|
||||||
* @param alpha
|
* @param alpha
|
||||||
* @param X
|
* @param X
|
||||||
* @param Y
|
* @param Y
|
||||||
|
@ -65,7 +65,7 @@ public interface Level1 {
|
||||||
/**
|
/**
|
||||||
* finds the element of a
|
* finds the element of a
|
||||||
* vector that has the largest absolute value.
|
* vector that has the largest absolute value.
|
||||||
* @param n the length to iterate for
|
* @param N the length to iterate for
|
||||||
* @param arr the array to get the max
|
* @param arr the array to get the max
|
||||||
* index for
|
* index for
|
||||||
* @param stride the stride for the array
|
* @param stride the stride for the array
|
||||||
|
@ -105,7 +105,7 @@ public interface Level1 {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* computes a vector-scalar product and adds the result to a vector.
|
* computes a vector-scalar product and adds the result to a vector.
|
||||||
* @param n
|
* @param N
|
||||||
* @param alpha
|
* @param alpha
|
||||||
* @param x
|
* @param x
|
||||||
* @param y
|
* @param y
|
||||||
|
@ -115,7 +115,7 @@ public interface Level1 {
|
||||||
/**
|
/**
|
||||||
* computes a vector-scalar product and adds the result to a vector.
|
* computes a vector-scalar product and adds the result to a vector.
|
||||||
* y = a*x + y
|
* y = a*x + y
|
||||||
* @param n number of operations
|
* @param N number of operations
|
||||||
* @param alpha
|
* @param alpha
|
||||||
* @param x X
|
* @param x X
|
||||||
* @param offsetX offset of first element of X in buffer
|
* @param offsetX offset of first element of X in buffer
|
||||||
|
|
|
@ -1297,7 +1297,7 @@ public abstract class BaseDataBuffer implements DataBuffer {
|
||||||
if (offset() == 0) {
|
if (offset() == 0) {
|
||||||
return wrappedBuffer().asIntBuffer();
|
return wrappedBuffer().asIntBuffer();
|
||||||
} else
|
} else
|
||||||
return wrappedBuffer().asIntBuffer().position((int) offset());
|
return (IntBuffer) wrappedBuffer().asIntBuffer().position((int) offset());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -1308,7 +1308,7 @@ public abstract class BaseDataBuffer implements DataBuffer {
|
||||||
if (offset() == 0) {
|
if (offset() == 0) {
|
||||||
return wrappedBuffer().asLongBuffer();
|
return wrappedBuffer().asLongBuffer();
|
||||||
} else
|
} else
|
||||||
return wrappedBuffer().asLongBuffer().position((int) offset());
|
return (LongBuffer) wrappedBuffer().asLongBuffer().position((int) offset());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -1319,7 +1319,7 @@ public abstract class BaseDataBuffer implements DataBuffer {
|
||||||
if (offset() == 0) {
|
if (offset() == 0) {
|
||||||
return wrappedBuffer().asDoubleBuffer();
|
return wrappedBuffer().asDoubleBuffer();
|
||||||
} else {
|
} else {
|
||||||
return wrappedBuffer().asDoubleBuffer().position((int) (offset()));
|
return (DoubleBuffer) wrappedBuffer().asDoubleBuffer().position((int) (offset()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1331,7 +1331,8 @@ public abstract class BaseDataBuffer implements DataBuffer {
|
||||||
if (offset() == 0) {
|
if (offset() == 0) {
|
||||||
return wrappedBuffer().asFloatBuffer();
|
return wrappedBuffer().asFloatBuffer();
|
||||||
} else {
|
} else {
|
||||||
return wrappedBuffer().asFloatBuffer().position((int) (offset()));
|
return (FloatBuffer) wrappedBuffer().asFloatBuffer()
|
||||||
|
.position((int) (offset()));
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -34,7 +34,7 @@ import java.util.*;
|
||||||
public class NDArrayCreationUtil {
|
public class NDArrayCreationUtil {
|
||||||
private NDArrayCreationUtil() {}
|
private NDArrayCreationUtil() {}
|
||||||
|
|
||||||
/** Get an array of INDArrays (2d) all with the specified shape. Pair<INDArray,String> returned to aid
|
/** Get an array of INDArrays (2d) all with the specified shape. {@code Pair<INDArray,String>} returned to aid
|
||||||
* debugging: String contains information on how to reproduce the matrix (i.e., which function, and arguments)
|
* debugging: String contains information on how to reproduce the matrix (i.e., which function, and arguments)
|
||||||
* Each NDArray in the returned array has been obtained by applying an operation such as transpose, tensorAlongDimension,
|
* Each NDArray in the returned array has been obtained by applying an operation such as transpose, tensorAlongDimension,
|
||||||
* etc to an original array.
|
* etc to an original array.
|
||||||
|
|
|
@ -302,7 +302,7 @@ public interface DataSet extends Iterable<org.nd4j.linalg.dataset.DataSet>, Seri
|
||||||
* Get the example metadata, or null if no metadata has been set
|
* Get the example metadata, or null if no metadata has been set
|
||||||
*
|
*
|
||||||
* @return List of metadata instances
|
* @return List of metadata instances
|
||||||
* @see {@link #getExampleMetaData(Class)} for convenience method for types
|
* {@link #getExampleMetaData(Class)} for convenience method for types
|
||||||
*/
|
*/
|
||||||
List<Serializable> getExampleMetaData();
|
List<Serializable> getExampleMetaData();
|
||||||
|
|
||||||
|
|
|
@ -180,7 +180,7 @@ public interface MultiDataSet extends Serializable {
|
||||||
* Get the example metadata, or null if no metadata has been set
|
* Get the example metadata, or null if no metadata has been set
|
||||||
*
|
*
|
||||||
* @return List of metadata instances
|
* @return List of metadata instances
|
||||||
* @see {@link #getExampleMetaData(Class)} for convenience method for types
|
* {@link #getExampleMetaData(Class)} for convenience method for types
|
||||||
*/
|
*/
|
||||||
List<Serializable> getExampleMetaData();
|
List<Serializable> getExampleMetaData();
|
||||||
|
|
||||||
|
|
|
@ -35,23 +35,27 @@ import java.util.*;
|
||||||
|
|
||||||
public class ND4JTestUtils {
|
public class ND4JTestUtils {
|
||||||
|
|
||||||
private ND4JTestUtils(){ }
|
private ND4JTestUtils() {
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
@AllArgsConstructor
|
@AllArgsConstructor
|
||||||
@Data
|
@Data
|
||||||
public static class ComparisonResult {
|
public static class ComparisonResult {
|
||||||
List<Triple<File,File,Boolean>> allResults;
|
|
||||||
List<Triple<File,File,Boolean>> passed;
|
List<Triple<File, File, Boolean>> allResults;
|
||||||
List<Triple<File,File,Boolean>> failed;
|
List<Triple<File, File, Boolean>> passed;
|
||||||
|
List<Triple<File, File, Boolean>> failed;
|
||||||
List<File> skippedDir1;
|
List<File> skippedDir1;
|
||||||
List<File> skippedDir2;
|
List<File> skippedDir2;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A function for use with {@link #validateSerializedArrays(File, File, boolean, BiFunction)} using {@code INDArray#equals(Object)}
|
* A function for use with {@link #validateSerializedArrays(File, File, boolean, BiFunction)}
|
||||||
|
* using {@code INDArray#equals(Object)}
|
||||||
*/
|
*/
|
||||||
public static class EqualsFn implements BiFunction<INDArray,INDArray,Boolean> {
|
public static class EqualsFn implements BiFunction<INDArray, INDArray, Boolean> {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Boolean apply(INDArray i1, INDArray i2) {
|
public Boolean apply(INDArray i1, INDArray i2) {
|
||||||
return i1.equals(i2);
|
return i1.equals(i2);
|
||||||
|
@ -59,10 +63,12 @@ public class ND4JTestUtils {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A function for use with {@link #validateSerializedArrays(File, File, boolean, BiFunction)} using {@link INDArray#equalsWithEps(Object, double)}
|
* A function for use with {@link #validateSerializedArrays(File, File, boolean, BiFunction)}
|
||||||
|
* using {@link INDArray#equalsWithEps(Object, double)}
|
||||||
*/
|
*/
|
||||||
@AllArgsConstructor
|
@AllArgsConstructor
|
||||||
public static class EqualsWithEpsFn implements BiFunction<INDArray,INDArray,Boolean> {
|
public static class EqualsWithEpsFn implements BiFunction<INDArray, INDArray, Boolean> {
|
||||||
|
|
||||||
private final double eps;
|
private final double eps;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -72,90 +78,99 @@ public class ND4JTestUtils {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Scan the specified directories for matching files (i.e., same path relative to their respective root directories)
|
* Scan the specified directories for matching files (i.e., same path relative to their respective
|
||||||
* and compare the contents using INDArray.equals (via {@link EqualsFn}
|
* root directories) and compare the contents using INDArray.equals (via {@link EqualsFn} Assumes
|
||||||
* Assumes the saved files represent INDArrays saved with {@link Nd4j#saveBinary(INDArray, File)}
|
* the saved files represent INDArrays saved with {@link Nd4j#saveBinary(INDArray, File)}
|
||||||
|
*
|
||||||
* @param dir1 First directory
|
* @param dir1 First directory
|
||||||
* @param dir2 Second directory
|
* @param dir2 Second directory
|
||||||
* @param recursive Whether to search recursively (i.e., include files in subdirectories
|
* @param recursive Whether to search recursively (i.e., include files in subdirectories
|
||||||
* @return Comparison results
|
* @return Comparison results
|
||||||
*/
|
*/
|
||||||
public static ComparisonResult validateSerializedArrays(File dir1, File dir2, boolean recursive) throws Exception {
|
public static ComparisonResult validateSerializedArrays(File dir1, File dir2, boolean recursive)
|
||||||
|
throws Exception {
|
||||||
return validateSerializedArrays(dir1, dir2, recursive, new EqualsFn());
|
return validateSerializedArrays(dir1, dir2, recursive, new EqualsFn());
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Scan the specified directories for matching files (i.e., same path relative to their respective root directories)
|
* Scan the specified directories for matching files (i.e., same path relative to their respective
|
||||||
* and compare the contents using a provided function.<br>
|
* root directories) and compare the contents using a provided function.<br> Assumes the saved
|
||||||
* Assumes the saved files represent INDArrays saved with {@link Nd4j#saveBinary(INDArray, File)}
|
* files represent INDArrays saved with {@link Nd4j#saveBinary(INDArray, File)}
|
||||||
|
*
|
||||||
* @param dir1 First directory
|
* @param dir1 First directory
|
||||||
* @param dir2 Second directory
|
* @param dir2 Second directory
|
||||||
* @param recursive Whether to search recursively (i.e., include files in subdirectories
|
* @param recursive Whether to search recursively (i.e., include files in subdirectories
|
||||||
* @return Comparison results
|
* @return Comparison results
|
||||||
*/
|
*/
|
||||||
public static ComparisonResult validateSerializedArrays(File dir1, File dir2, boolean recursive, BiFunction<INDArray,INDArray,Boolean> evalFn) throws Exception {
|
public static ComparisonResult validateSerializedArrays(File dir1, File dir2, boolean recursive,
|
||||||
|
BiFunction<INDArray, INDArray, Boolean> evalFn) throws Exception {
|
||||||
File[] f1 = FileUtils.listFiles(dir1, null, recursive).toArray(new File[0]);
|
File[] f1 = FileUtils.listFiles(dir1, null, recursive).toArray(new File[0]);
|
||||||
File[] f2 = FileUtils.listFiles(dir2, null, recursive).toArray(new File[0]);
|
File[] f2 = FileUtils.listFiles(dir2, null, recursive).toArray(new File[0]);
|
||||||
|
|
||||||
Preconditions.checkState(f1.length > 0, "No files found for directory 1: %s", dir1.getAbsolutePath() );
|
Preconditions.checkState(f1.length > 0, "No files found for directory 1: %s",
|
||||||
Preconditions.checkState(f2.length > 0, "No files found for directory 2: %s", dir2.getAbsolutePath() );
|
dir1.getAbsolutePath());
|
||||||
|
Preconditions.checkState(f2.length > 0, "No files found for directory 2: %s",
|
||||||
|
dir2.getAbsolutePath());
|
||||||
|
|
||||||
Map<String,File> relativized1 = new HashMap<>();
|
Map<String, File> relativized1 = new HashMap<>();
|
||||||
Map<String,File> relativized2 = new HashMap<>();
|
Map<String, File> relativized2 = new HashMap<>();
|
||||||
|
|
||||||
URI u = dir1.toURI();
|
URI u = dir1.toURI();
|
||||||
for(File f : f1){
|
for (File f : f1) {
|
||||||
if(!f.isFile())
|
if (!f.isFile()) {
|
||||||
continue;
|
continue;
|
||||||
|
}
|
||||||
String relative = u.relativize(f.toURI()).getPath();
|
String relative = u.relativize(f.toURI()).getPath();
|
||||||
relativized1.put(relative, f);
|
relativized1.put(relative, f);
|
||||||
}
|
}
|
||||||
|
|
||||||
u = dir2.toURI();
|
u = dir2.toURI();
|
||||||
for(File f : f2){
|
for (File f : f2) {
|
||||||
if(!f.isFile())
|
if (!f.isFile()) {
|
||||||
continue;
|
continue;
|
||||||
|
}
|
||||||
String relative = u.relativize(f.toURI()).getPath();
|
String relative = u.relativize(f.toURI()).getPath();
|
||||||
relativized2.put(relative, f);
|
relativized2.put(relative, f);
|
||||||
}
|
}
|
||||||
|
|
||||||
List<File> skipped1 = new ArrayList<>();
|
List<File> skipped1 = new ArrayList<>();
|
||||||
for(String s : relativized1.keySet()){
|
for (String s : relativized1.keySet()) {
|
||||||
if(!relativized2.containsKey(s)){
|
if (!relativized2.containsKey(s)) {
|
||||||
skipped1.add(relativized1.get(s));
|
skipped1.add(relativized1.get(s));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
List<File> skipped2 = new ArrayList<>();
|
List<File> skipped2 = new ArrayList<>();
|
||||||
for(String s : relativized2.keySet()){
|
for (String s : relativized2.keySet()) {
|
||||||
if(!relativized1.containsKey(s)){
|
if (!relativized1.containsKey(s)) {
|
||||||
skipped2.add(relativized1.get(s));
|
skipped2.add(relativized1.get(s));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
List<Triple<File,File,Boolean>> allResults = new ArrayList<>();
|
List<Triple<File, File, Boolean>> allResults = new ArrayList<>();
|
||||||
List<Triple<File,File,Boolean>> passed = new ArrayList<>();
|
List<Triple<File, File, Boolean>> passed = new ArrayList<>();
|
||||||
List<Triple<File,File,Boolean>> failed = new ArrayList<>();
|
List<Triple<File, File, Boolean>> failed = new ArrayList<>();
|
||||||
for(Map.Entry<String,File> e : relativized1.entrySet()){
|
for (Map.Entry<String, File> e : relativized1.entrySet()) {
|
||||||
File file1 = e.getValue();
|
File file1 = e.getValue();
|
||||||
File file2 = relativized2.get(e.getKey());
|
File file2 = relativized2.get(e.getKey());
|
||||||
|
|
||||||
if(file2 == null)
|
if (file2 == null) {
|
||||||
continue;
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
INDArray i1 = Nd4j.readBinary(file1);
|
INDArray i1 = Nd4j.readBinary(file1);
|
||||||
INDArray i2 = Nd4j.readBinary(file2);
|
INDArray i2 = Nd4j.readBinary(file2);
|
||||||
boolean b = evalFn.apply(i1, i2);
|
boolean b = evalFn.apply(i1, i2);
|
||||||
Triple<File,File,Boolean> t = new Triple<>(file1, file2, b);
|
Triple<File, File, Boolean> t = new Triple<>(file1, file2, b);
|
||||||
allResults.add(t);
|
allResults.add(t);
|
||||||
if(b){
|
if (b) {
|
||||||
passed.add(t);
|
passed.add(t);
|
||||||
} else {
|
} else {
|
||||||
failed.add(t);
|
failed.add(t);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
Comparator<Triple<File,File,Boolean>> c = new Comparator<Triple<File, File, Boolean>>() {
|
Comparator<Triple<File, File, Boolean>> c = new Comparator<Triple<File, File, Boolean>>() {
|
||||||
@Override
|
@Override
|
||||||
public int compare(Triple<File, File, Boolean> o1, Triple<File, File, Boolean> o2) {
|
public int compare(Triple<File, File, Boolean> o1, Triple<File, File, Boolean> o2) {
|
||||||
return o1.getFirst().compareTo(o2.getFirst());
|
return o1.getFirst().compareTo(o2.getFirst());
|
||||||
|
@ -168,7 +183,6 @@ public class ND4JTestUtils {
|
||||||
Collections.sort(skipped1);
|
Collections.sort(skipped1);
|
||||||
Collections.sort(skipped2);
|
Collections.sort(skipped2);
|
||||||
|
|
||||||
|
|
||||||
return new ComparisonResult(allResults, passed, failed, skipped1, skipped2);
|
return new ComparisonResult(allResults, passed, failed, skipped1, skipped2);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -211,7 +211,7 @@ public class EmnistDataSetIterator extends BaseDatasetIterator {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the labels as a List<String>
|
* Get the labels as a {@code List<String>}
|
||||||
*
|
*
|
||||||
* @return Labels
|
* @return Labels
|
||||||
*/
|
*/
|
||||||
|
@ -244,7 +244,7 @@ public class EmnistDataSetIterator extends BaseDatasetIterator {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the label assignments for the given set as a List<String>
|
* Get the label assignments for the given set as a {@code List<String>}
|
||||||
*
|
*
|
||||||
* @param dataSet DataSet to get the label assignment for
|
* @param dataSet DataSet to get the label assignment for
|
||||||
* @return Label assignment and given dataset
|
* @return Label assignment and given dataset
|
||||||
|
|
|
@ -31,59 +31,91 @@ import java.util.Random;
|
||||||
|
|
||||||
public class LFWDataSetIterator extends RecordReaderDataSetIterator {
|
public class LFWDataSetIterator extends RecordReaderDataSetIterator {
|
||||||
|
|
||||||
/** Loads subset of images with given imgDim returned by the generator. */
|
/**
|
||||||
|
* Loads subset of images with given imgDim returned by the generator.
|
||||||
|
*/
|
||||||
public LFWDataSetIterator(int[] imgDim) {
|
public LFWDataSetIterator(int[] imgDim) {
|
||||||
this(LFWLoader.SUB_NUM_IMAGES, LFWLoader.SUB_NUM_IMAGES, imgDim, LFWLoader.SUB_NUM_LABELS, false,
|
this(LFWLoader.SUB_NUM_IMAGES, LFWLoader.SUB_NUM_IMAGES, imgDim, LFWLoader.SUB_NUM_LABELS,
|
||||||
|
false,
|
||||||
new ParentPathLabelGenerator(), true, 1, null, new Random(System.currentTimeMillis()));
|
new ParentPathLabelGenerator(), true, 1, null, new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples returned by the generator. */
|
/**
|
||||||
|
* Loads images with given batchSize, numExamples returned by the generator.
|
||||||
|
*/
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples) {
|
public LFWDataSetIterator(int batchSize, int numExamples) {
|
||||||
this(batchSize, numExamples, new int[] {LFWLoader.HEIGHT, LFWLoader.WIDTH, LFWLoader.CHANNELS},
|
this(batchSize, numExamples, new int[]{LFWLoader.HEIGHT, LFWLoader.WIDTH, LFWLoader.CHANNELS},
|
||||||
LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN, true, 1, null,
|
LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN, true, 1, null,
|
||||||
new Random(System.currentTimeMillis()));
|
new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples, imgDim returned by the generator. */
|
/**
|
||||||
|
* Loads images with given batchSize, numExamples, imgDim returned by the generator.
|
||||||
|
*/
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim) {
|
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim) {
|
||||||
this(batchSize, numExamples, imgDim, LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN, true, 1, null,
|
this(batchSize, numExamples, imgDim, LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN, true,
|
||||||
|
1, null,
|
||||||
new Random(System.currentTimeMillis()));
|
new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, imgDim, useSubset, returned by the generator. */
|
/**
|
||||||
|
* Loads images with given batchSize, imgDim, useSubset, returned by the generator.
|
||||||
|
*/
|
||||||
public LFWDataSetIterator(int batchSize, int[] imgDim, boolean useSubset) {
|
public LFWDataSetIterator(int batchSize, int[] imgDim, boolean useSubset) {
|
||||||
this(batchSize, useSubset ? LFWLoader.SUB_NUM_IMAGES : LFWLoader.NUM_IMAGES, imgDim,
|
this(batchSize, useSubset ? LFWLoader.SUB_NUM_IMAGES : LFWLoader.NUM_IMAGES, imgDim,
|
||||||
useSubset ? LFWLoader.SUB_NUM_LABELS : LFWLoader.NUM_LABELS, useSubset, LFWLoader.LABEL_PATTERN,
|
useSubset ? LFWLoader.SUB_NUM_LABELS : LFWLoader.NUM_LABELS, useSubset,
|
||||||
|
LFWLoader.LABEL_PATTERN,
|
||||||
true, 1, null, new Random(System.currentTimeMillis()));
|
true, 1, null, new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples, imgDim, train, & splitTrainTest returned by the generator. */
|
/**
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, boolean train, double splitTrainTest) {
|
* Loads images with given batchSize, numExamples, imgDim, train, & splitTrainTest returned
|
||||||
this(batchSize, numExamples, imgDim, LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN, train,
|
* by the generator.
|
||||||
|
*/
|
||||||
|
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, boolean train,
|
||||||
|
double splitTrainTest) {
|
||||||
|
this(batchSize, numExamples, imgDim, LFWLoader.NUM_LABELS, false, LFWLoader.LABEL_PATTERN,
|
||||||
|
train,
|
||||||
splitTrainTest, null, new Random(System.currentTimeMillis()));
|
splitTrainTest, null, new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples, numLabels, train, & splitTrainTest returned by the generator. */
|
/**
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int numLabels, boolean train, double splitTrainTest) {
|
* Loads images with given batchSize, numExamples, numLabels, train, & splitTrainTest
|
||||||
this(batchSize, numExamples, new int[] {LFWLoader.HEIGHT, LFWLoader.WIDTH, LFWLoader.CHANNELS}, numLabels,
|
* returned by the generator.
|
||||||
|
*/
|
||||||
|
public LFWDataSetIterator(int batchSize, int numExamples, int numLabels, boolean train,
|
||||||
|
double splitTrainTest) {
|
||||||
|
this(batchSize, numExamples, new int[]{LFWLoader.HEIGHT, LFWLoader.WIDTH, LFWLoader.CHANNELS},
|
||||||
|
numLabels,
|
||||||
false, null, train, splitTrainTest, null, new Random(System.currentTimeMillis()));
|
false, null, train, splitTrainTest, null, new Random(System.currentTimeMillis()));
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples, imgDim, numLabels, useSubset, train, splitTrainTest & Random returned by the generator. */
|
/**
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels, boolean useSubset,
|
* Loads images with given batchSize, numExamples, imgDim, numLabels, useSubset, train,
|
||||||
|
* splitTrainTest & Random returned by the generator.
|
||||||
|
*/
|
||||||
|
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels,
|
||||||
|
boolean useSubset,
|
||||||
boolean train, double splitTrainTest, Random rng) {
|
boolean train, double splitTrainTest, Random rng) {
|
||||||
this(batchSize, numExamples, imgDim, numLabels, useSubset, LFWLoader.LABEL_PATTERN, train, splitTrainTest, null,
|
this(batchSize, numExamples, imgDim, numLabels, useSubset, LFWLoader.LABEL_PATTERN, train,
|
||||||
|
splitTrainTest, null,
|
||||||
rng);
|
rng);
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Loads images with given batchSize, numExamples, imgDim, numLabels, useSubset, train, splitTrainTest & Random returned by the generator. */
|
/**
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels, boolean useSubset,
|
* Loads images with given batchSize, numExamples, imgDim, numLabels, useSubset, train,
|
||||||
|
* splitTrainTest & Random returned by the generator.
|
||||||
|
*/
|
||||||
|
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels,
|
||||||
|
boolean useSubset,
|
||||||
PathLabelGenerator labelGenerator, boolean train, double splitTrainTest, Random rng) {
|
PathLabelGenerator labelGenerator, boolean train, double splitTrainTest, Random rng) {
|
||||||
this(batchSize, numExamples, imgDim, numLabels, useSubset, labelGenerator, train, splitTrainTest, null, rng);
|
this(batchSize, numExamples, imgDim, numLabels, useSubset, labelGenerator, train,
|
||||||
|
splitTrainTest, null, rng);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create LFW data specific iterator
|
* Create LFW data specific iterator
|
||||||
|
*
|
||||||
* @param batchSize the batch size of the examples
|
* @param batchSize the batch size of the examples
|
||||||
* @param numExamples the overall number of examples
|
* @param numExamples the overall number of examples
|
||||||
* @param imgDim an array of height, width and channels
|
* @param imgDim an array of height, width and channels
|
||||||
|
@ -93,13 +125,14 @@ public class LFWDataSetIterator extends RecordReaderDataSetIterator {
|
||||||
* @param train true if use train value
|
* @param train true if use train value
|
||||||
* @param splitTrainTest the percentage to split data for train and remainder goes to test
|
* @param splitTrainTest the percentage to split data for train and remainder goes to test
|
||||||
* @param imageTransform how to transform the image
|
* @param imageTransform how to transform the image
|
||||||
|
|
||||||
* @param rng random number to lock in batch shuffling
|
* @param rng random number to lock in batch shuffling
|
||||||
* */
|
*/
|
||||||
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels, boolean useSubset,
|
public LFWDataSetIterator(int batchSize, int numExamples, int[] imgDim, int numLabels,
|
||||||
|
boolean useSubset,
|
||||||
PathLabelGenerator labelGenerator, boolean train, double splitTrainTest,
|
PathLabelGenerator labelGenerator, boolean train, double splitTrainTest,
|
||||||
ImageTransform imageTransform, Random rng) {
|
ImageTransform imageTransform, Random rng) {
|
||||||
super(new LFWLoader(imgDim, imageTransform, useSubset).getRecordReader(batchSize, numExamples, imgDim,
|
super(new LFWLoader(imgDim, imageTransform, useSubset).getRecordReader(batchSize, numExamples,
|
||||||
|
imgDim,
|
||||||
numLabels, labelGenerator, train, splitTrainTest, rng), batchSize, 1, numLabels);
|
numLabels, labelGenerator, train, splitTrainTest, rng), batchSize, 1, numLabels);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -12,12 +12,15 @@ configurations.archives.artifacts.with { archives ->
|
||||||
dependencies {
|
dependencies {
|
||||||
//Todo clean this
|
//Todo clean this
|
||||||
api platform(project(":cavis-common-platform"))
|
api platform(project(":cavis-common-platform"))
|
||||||
api "org.bytedeco:javacpp:1.5.7" //for some reason we needed to apply version numbers here, they do not end up in POM otherwise
|
//api "org.bytedeco:javacpp:1.5.7" //for some reason we needed to apply version numbers here, they do not end up in POM otherwise
|
||||||
api "com.fasterxml.jackson.datatype:jackson-datatype-joda:2.10.5"
|
api "com.fasterxml.jackson.datatype:jackson-datatype-joda:2.10.5"
|
||||||
api 'org.slf4j:slf4j-simple:2.0.3'
|
api 'org.slf4j:slf4j-simple:2.0.3'
|
||||||
api 'org.slf4j:slf4j-api:2.0.3'
|
api 'org.slf4j:slf4j-api:2.0.3'
|
||||||
//api group: "org.bytedeco", name: "javacpp", classifier: "linux-x64_86"
|
//TODO for the two below.. either platform specific uber jars or a single big one with all platforms
|
||||||
|
api group: "org.bytedeco", name: "javacpp", version: "1.5.7", classifier: "linux-x86_64"
|
||||||
|
//api group: "org.bytedeco", name: "javacpp", version: "1.5.7"
|
||||||
|
// api group: 'net.brutex.cavis-native', name: 'cavis-native-lib', version: '1.0.0-SNAPSHOT', classifier: "linux-x86_64-avx2-cpu"
|
||||||
|
//api group: 'net.brutex.cavis-native', name: 'cavis-native-lib', version: '1.0.0-SNAPSHOT'
|
||||||
rootProject.getAllprojects().each { Project sproj ->
|
rootProject.getAllprojects().each { Project sproj ->
|
||||||
if(!sproj.name.equals(name) && !sproj.name.equals("cavis-common-platform")
|
if(!sproj.name.equals(name) && !sproj.name.equals("cavis-common-platform")
|
||||||
&& !sproj.name.equals("Cavis")
|
&& !sproj.name.equals("Cavis")
|
||||||
|
|
|
@ -256,7 +256,7 @@ public class NDArrayMessage implements Serializable {
|
||||||
String messageId = UUID.randomUUID().toString();
|
String messageId = UUID.randomUUID().toString();
|
||||||
for (int i = 0; i < ret.length; i++) {
|
for (int i = 0; i < ret.length; i++) {
|
||||||
//data: only grab a chunk of the data
|
//data: only grab a chunk of the data
|
||||||
ByteBuffer view = wholeBuffer.byteBuffer().asReadOnlyBuffer().position(i * chunkSize);
|
ByteBuffer view = (ByteBuffer) wholeBuffer.byteBuffer().asReadOnlyBuffer().position(i * chunkSize);
|
||||||
view.limit(Math.min(i * chunkSize + chunkSize, wholeBuffer.capacity()));
|
view.limit(Math.min(i * chunkSize + chunkSize, wholeBuffer.capacity()));
|
||||||
view.order(ByteOrder.nativeOrder());
|
view.order(ByteOrder.nativeOrder());
|
||||||
view = view.slice();
|
view = view.slice();
|
||||||
|
|
|
@ -664,8 +664,8 @@ public class RoutedTransport extends BaseTransport {
|
||||||
|
|
||||||
|
|
||||||
public static class RemoteConnectionBuilder {
|
public static class RemoteConnectionBuilder {
|
||||||
private final Object locker = new Object();
|
private Object locker = new Object();
|
||||||
private final AtomicBoolean activated = new AtomicBoolean();
|
private AtomicBoolean activated = new AtomicBoolean();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue