Re-enable datavec-geo and datavec-hadoop as module in pom.xml; update imports (#425)

* Re-enable datavec-geo and datavec-hadoop as module in pom.xml; update imports

Signed-off-by: Alex Black <blacka101@gmail.com>

* Fix parent pom and cyclical dependency

Signed-off-by: Alex Black <blacka101@gmail.com>
master
Alex Black 2020-05-01 17:32:29 +10:00 committed by GitHub
parent 668d773993
commit e64fbbdc9b
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
16 changed files with 30 additions and 26 deletions

View File

@ -67,6 +67,13 @@
<version>${nd4j.version}</version> <version>${nd4j.version}</version>
<scope>test</scope> <scope>test</scope>
</dependency> </dependency>
<dependency>
<groupId>org.datavec</groupId>
<artifactId>datavec-local</artifactId>
<version>${project.version}</version>
<scope>test</scope>
</dependency>
</dependencies> </dependencies>
<profiles> <profiles>

View File

@ -18,7 +18,7 @@
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<parent> <parent>
<artifactId>datavec-parent</artifactId> <artifactId>datavec-data</artifactId>
<groupId>org.datavec</groupId> <groupId>org.datavec</groupId>
<version>1.0.0-SNAPSHOT</version> <version>1.0.0-SNAPSHOT</version>
</parent> </parent>

View File

@ -27,7 +27,7 @@ import org.datavec.api.transform.schema.Schema;
import org.datavec.api.writable.DoubleWritable; import org.datavec.api.writable.DoubleWritable;
import org.datavec.api.writable.Text; import org.datavec.api.writable.Text;
import org.datavec.api.writable.Writable; import org.datavec.api.writable.Writable;
import org.nd4j.linalg.function.Supplier; import org.nd4j.common.function.Supplier;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collections; import java.util.Collections;

View File

@ -17,8 +17,8 @@
package org.datavec.api.transform.transform.geo; package org.datavec.api.transform.transform.geo;
import org.apache.commons.io.FileUtils; import org.apache.commons.io.FileUtils;
import org.nd4j.base.Preconditions; import org.nd4j.common.base.Preconditions;
import org.nd4j.util.ArchiveUtils; import org.nd4j.common.util.ArchiveUtils;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;

View File

@ -16,9 +16,10 @@
package org.datavec.api.transform; package org.datavec.api.transform;
import lombok.extern.slf4j.Slf4j; import lombok.extern.slf4j.Slf4j;
import org.nd4j.common.tests.AbstractAssertTestsClass;
import org.nd4j.common.tests.BaseND4JTest;
import java.util.*; import java.util.*;
import org.nd4j.AbstractAssertTestsClass;
import org.nd4j.BaseND4JTest;
/** /**
* This class checks that all test classes (i.e., anything with one or more methods annotated with @Test) * This class checks that all test classes (i.e., anything with one or more methods annotated with @Test)

View File

@ -18,7 +18,7 @@
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<parent> <parent>
<artifactId>datavec-parent</artifactId> <artifactId>datavec-data</artifactId>
<groupId>org.datavec</groupId> <groupId>org.datavec</groupId>
<version>1.0.0-SNAPSHOT</version> <version>1.0.0-SNAPSHOT</version>
</parent> </parent>

View File

@ -19,7 +19,7 @@ package org.datavec.hadoop.records.reader.mapfile;
import org.apache.hadoop.io.MapFile; import org.apache.hadoop.io.MapFile;
import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.Writable;
import org.apache.hadoop.io.WritableComparable; import org.apache.hadoop.io.WritableComparable;
import org.nd4j.linalg.primitives.Pair; import org.nd4j.common.primitives.Pair;
import java.io.IOException; import java.io.IOException;
import java.util.List; import java.util.List;

View File

@ -23,9 +23,9 @@ import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.Writable;
import org.apache.hadoop.io.WritableComparable; import org.apache.hadoop.io.WritableComparable;
import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.ReflectionUtils;
import org.nd4j.linalg.primitives.Pair;
import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey; import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey;
import org.datavec.hadoop.records.reader.mapfile.record.RecordWritable; import org.datavec.hadoop.records.reader.mapfile.record.RecordWritable;
import org.nd4j.common.primitives.Pair;
import java.io.Closeable; import java.io.Closeable;
import java.io.IOException; import java.io.IOException;

View File

@ -26,7 +26,7 @@ import org.datavec.api.split.InputSplit;
import org.datavec.api.writable.Writable; import org.datavec.api.writable.Writable;
import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey; import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey;
import org.datavec.hadoop.records.reader.mapfile.record.RecordWritable; import org.datavec.hadoop.records.reader.mapfile.record.RecordWritable;
import org.nd4j.linalg.util.MathUtils; import org.nd4j.common.util.MathUtils;
import java.io.DataInputStream; import java.io.DataInputStream;
import java.io.File; import java.io.File;

View File

@ -28,7 +28,7 @@ import org.datavec.api.split.InputSplit;
import org.datavec.api.writable.Writable; import org.datavec.api.writable.Writable;
import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey; import org.datavec.hadoop.records.reader.mapfile.index.LongIndexToKey;
import org.datavec.hadoop.records.reader.mapfile.record.SequenceRecordWritable; import org.datavec.hadoop.records.reader.mapfile.record.SequenceRecordWritable;
import org.nd4j.linalg.util.MathUtils; import org.nd4j.common.util.MathUtils;
import java.io.DataInputStream; import java.io.DataInputStream;
import java.io.File; import java.io.File;

View File

@ -18,11 +18,10 @@ package org.datavec.hadoop.records.reader.mapfile.index;
import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.MapFile; import org.apache.hadoop.io.MapFile;
import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.Writable;
import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.ReflectionUtils;
import org.nd4j.linalg.primitives.Pair;
import org.datavec.hadoop.records.reader.mapfile.IndexToKey; import org.datavec.hadoop.records.reader.mapfile.IndexToKey;
import org.nd4j.common.primitives.Pair;
import java.io.IOException; import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;

View File

@ -16,9 +16,10 @@
package org.datavec.hadoop; package org.datavec.hadoop;
import lombok.extern.slf4j.Slf4j; import lombok.extern.slf4j.Slf4j;
import org.nd4j.common.tests.AbstractAssertTestsClass;
import org.nd4j.common.tests.BaseND4JTest;
import java.util.*; import java.util.*;
import org.nd4j.AbstractAssertTestsClass;
import org.nd4j.BaseND4JTest;
/** /**
* This class checks that all test classes (i.e., anything with one or more methods annotated with @Test) * This class checks that all test classes (i.e., anything with one or more methods annotated with @Test)
* extends BaseND4jTest - either directly or indirectly. * extends BaseND4jTest - either directly or indirectly.

View File

@ -16,6 +16,7 @@
package org.datavec.hadoop.records.reader; package org.datavec.hadoop.records.reader;
import org.nd4j.common.util.MathUtils;
import org.nd4j.shade.guava.io.Files; import org.nd4j.shade.guava.io.Files;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -36,7 +37,6 @@ import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Test; import org.junit.Test;
import org.nd4j.linalg.factory.Nd4j; import org.nd4j.linalg.factory.Nd4j;
import org.nd4j.linalg.util.MathUtils;
import java.io.File; import java.io.File;
import java.io.IOException; import java.io.IOException;

View File

@ -16,6 +16,8 @@
package org.datavec.hadoop.records.reader; package org.datavec.hadoop.records.reader;
import org.nd4j.common.primitives.Pair;
import org.nd4j.common.util.MathUtils;
import org.nd4j.shade.guava.io.Files; import org.nd4j.shade.guava.io.Files;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -36,8 +38,6 @@ import org.datavec.hadoop.records.reader.mapfile.record.SequenceRecordWritable;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Test; import org.junit.Test;
import org.nd4j.linalg.primitives.Pair;
import org.nd4j.linalg.util.MathUtils;
import java.io.File; import java.io.File;
import java.io.IOException; import java.io.IOException;

View File

@ -16,6 +16,8 @@
package org.datavec.hadoop.records.reader; package org.datavec.hadoop.records.reader;
import org.nd4j.common.primitives.Pair;
import org.nd4j.common.util.MathUtils;
import org.nd4j.shade.guava.io.Files; import org.nd4j.shade.guava.io.Files;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
@ -36,8 +38,6 @@ import org.datavec.hadoop.records.reader.mapfile.record.SequenceRecordWritable;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Test; import org.junit.Test;
import org.nd4j.linalg.primitives.Pair;
import org.nd4j.linalg.util.MathUtils;
import java.io.File; import java.io.File;
import java.io.IOException; import java.io.IOException;

View File

@ -33,6 +33,8 @@
<module>datavec-data-codec</module> <module>datavec-data-codec</module>
<module>datavec-data-image</module> <module>datavec-data-image</module>
<module>datavec-data-nlp</module> <module>datavec-data-nlp</module>
<module>datavec-geo</module>
<module>datavec-hadoop</module>
</modules> </modules>
<properties> <properties>
@ -45,12 +47,6 @@
<artifactId>nd4j-api</artifactId> <artifactId>nd4j-api</artifactId>
<version>${nd4j.version}</version> <version>${nd4j.version}</version>
</dependency> </dependency>
<dependency>
<groupId>org.datavec</groupId>
<artifactId>datavec-local</artifactId>
<version>${project.version}</version>
<scope>test</scope>
</dependency>
</dependencies> </dependencies>
<profiles> <profiles>