Skip to content

Commit de5f2bb

Browse files
authored
chore: remove unused import (#246)
1 parent d66ffe4 commit de5f2bb

30 files changed

+49
-130
lines changed

spark/.scalafix.conf

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
rules = [
2+
RemoveUnused
3+
]

spark/pom.xml

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -197,11 +197,29 @@
197197
<groupId>net.alchim31.maven</groupId>
198198
<artifactId>scala-maven-plugin</artifactId>
199199
<version>4.8.0</version>
200+
<executions>
201+
<execution>
202+
<goals>
203+
<goal>compile</goal>
204+
<goal>testCompile</goal>
205+
</goals>
206+
</execution>
207+
</executions>
200208
<configuration>
201209
<jvmArgs>
202210
<jvmArg>-Xms64m</jvmArg>
203211
<jvmArg>-Xmx1024m</jvmArg>
204212
</jvmArgs>
213+
<args>
214+
<arg>-Ywarn-unused</arg>
215+
</args>
216+
<compilerPlugins>
217+
<compilerPlugin>
218+
<groupId>org.scalameta</groupId>
219+
<artifactId>semanticdb-scalac_2.12.10</artifactId>
220+
<version>4.3.24</version>
221+
</compilerPlugin>
222+
</compilerPlugins>
205223
</configuration>
206224
</plugin>
207225
<plugin>
@@ -225,6 +243,11 @@
225243
</scala>
226244
</configuration>
227245
</plugin>
246+
<plugin>
247+
<groupId>io.github.evis</groupId>
248+
<artifactId>scalafix-maven-plugin_2.13</artifactId>
249+
<version>0.1.8_0.11.0</version>
250+
</plugin>
228251
</plugins>
229252
</build>
230253
<packaging>jar</packaging>

spark/src/main/scala/com/alibaba/graphar/EdgeInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/GraphInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/VertexInfo.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import java.io.{File, FileInputStream}
20-
import org.apache.hadoop.fs.{Path, FileSystem}
19+
import org.apache.hadoop.fs.Path
2120
import org.apache.spark.sql.{SparkSession}
2221
import org.yaml.snakeyaml.{Yaml, DumperOptions}
2322
import org.yaml.snakeyaml.constructor.Constructor

spark/src/main/scala/com/alibaba/graphar/datasources/GarCommitProtocol.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,9 +21,7 @@ import com.alibaba.graphar.GeneralParams
2121
import org.json4s._
2222
import org.json4s.jackson.JsonMethods._
2323

24-
import org.apache.spark.internal.io.FileCommitProtocol
2524
import org.apache.spark.sql.execution.datasources.SQLHadoopMapReduceCommitProtocol
26-
import org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl
2725
import org.apache.hadoop.mapreduce._
2826
import org.apache.spark.internal.Logging
2927

spark/src/main/scala/com/alibaba/graphar/datasources/GarDataSource.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,6 @@ import org.apache.spark.sql.SparkSession
2929
import org.apache.spark.sql.execution.datasources.csv.CSVFileFormat
3030
import org.apache.spark.sql.execution.datasources.orc.OrcFileFormat
3131
import org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat
32-
import org.apache.spark.sql.execution.datasources.v2._
3332
import org.apache.spark.sql.types.StructType
3433
import org.apache.spark.sql.util.CaseInsensitiveStringMap
3534
import org.apache.spark.sql.sources.DataSourceRegister

spark/src/main/scala/com/alibaba/graphar/datasources/GarWriterBuilder.scala

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -26,12 +26,9 @@ import org.apache.hadoop.conf.Configuration
2626
import org.apache.hadoop.fs.Path
2727
import org.apache.hadoop.mapreduce.Job
2828
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
29-
import org.apache.hadoop.mapreduce.{Job, TaskAttemptContext}
29+
import org.apache.hadoop.mapreduce.Job
3030

31-
import org.apache.spark.sql.execution.datasources.{
32-
OutputWriter,
33-
OutputWriterFactory
34-
}
31+
import org.apache.spark.sql.execution.datasources.OutputWriterFactory
3532
import org.apache.spark.sql.SparkSession
3633
import org.apache.spark.sql.catalyst.InternalRow
3734
import org.apache.spark.sql.catalyst.util.{CaseInsensitiveMap, DateTimeUtils}
@@ -49,7 +46,6 @@ import org.apache.spark.sql.execution.datasources.{
4946
import org.apache.spark.sql.execution.metric.SQLMetric
5047
import org.apache.spark.sql.internal.SQLConf
5148
import org.apache.spark.sql.types.{DataType, StructType}
52-
import org.apache.spark.sql.util.SchemaUtils
5349
import org.apache.spark.util.SerializableConfiguration
5450
import org.apache.spark.sql.execution.datasources.v2.FileBatchWrite
5551
import org.apache.spark.sql.catalyst.expressions.AttributeReference

spark/src/main/scala/com/alibaba/graphar/datasources/csv/CSVWriterBuilder.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,6 @@ import org.apache.spark.sql.execution.datasources.{
3030
import org.apache.spark.sql.execution.datasources.csv.CsvOutputWriter
3131
import org.apache.spark.sql.internal.SQLConf
3232
import org.apache.spark.sql.types.{DataType, StructType}
33-
import org.apache.spark.sql.sources.Filter
34-
import org.apache.spark.sql.connector.write.SupportsOverwrite
3533

3634
import com.alibaba.graphar.datasources.GarWriteBuilder
3735

spark/src/main/scala/com/alibaba/graphar/example/GraphAr2Neo4j.scala

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -16,18 +16,11 @@
1616

1717
package com.alibaba.graphar.example
1818

19-
import com.alibaba.graphar.datasources._
20-
import com.alibaba.graphar.reader.{VertexReader, EdgeReader}
2119
import com.alibaba.graphar.graph.GraphReader
2220
import com.alibaba.graphar.{GeneralParams, GraphInfo}
2321
import com.alibaba.graphar.util.Utils
2422

25-
import java.io.{File, FileInputStream}
26-
import org.yaml.snakeyaml.Yaml
27-
import org.yaml.snakeyaml.constructor.Constructor
28-
import scala.beans.BeanProperty
2923
import org.apache.spark.sql.{DataFrame, SaveMode, SparkSession}
30-
import org.apache.hadoop.fs.{Path, FileSystem}
3124

3225
object GraphAr2Neo4j {
3326

spark/src/main/scala/com/alibaba/graphar/example/Neo4j2GraphAr.scala

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,9 @@
1616

1717
package com.alibaba.graphar.example
1818

19-
import com.alibaba.graphar.datasources._
2019
import com.alibaba.graphar.graph.GraphWriter
2120

22-
import org.apache.spark.sql.{DataFrame, SparkSession}
21+
import org.apache.spark.sql.SparkSession
2322

2423
object Neo4j2GraphAr {
2524

spark/src/main/scala/com/alibaba/graphar/graph/GraphReader.scala

Lines changed: 1 addition & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -16,19 +16,10 @@
1616

1717
package com.alibaba.graphar.graph
1818

19-
import com.alibaba.graphar.{
20-
GeneralParams,
21-
AdjListType,
22-
GraphInfo,
23-
VertexInfo,
24-
EdgeInfo
25-
}
19+
import com.alibaba.graphar.{GraphInfo, VertexInfo, EdgeInfo}
2620
import com.alibaba.graphar.reader.{VertexReader, EdgeReader}
27-
import com.alibaba.graphar.util.IndexGenerator
2821

2922
import org.apache.spark.sql.{DataFrame, SparkSession}
30-
import org.apache.spark.sql.types._
31-
import org.apache.spark.sql.functions._
3223

3324
/**
3425
* The helper object for reading graph through the definitions of graph info.

spark/src/main/scala/com/alibaba/graphar/graph/GraphTransformer.scala

Lines changed: 2 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -16,19 +16,11 @@
1616

1717
package com.alibaba.graphar.graph
1818

19-
import com.alibaba.graphar.{
20-
GeneralParams,
21-
AdjListType,
22-
GraphInfo,
23-
VertexInfo,
24-
EdgeInfo
25-
}
19+
import com.alibaba.graphar.{AdjListType, GraphInfo, VertexInfo, EdgeInfo}
2620
import com.alibaba.graphar.reader.{VertexReader, EdgeReader}
2721
import com.alibaba.graphar.writer.{VertexWriter, EdgeWriter}
2822

29-
import org.apache.spark.sql.{DataFrame, SparkSession}
30-
import org.apache.spark.sql.types._
31-
import org.apache.spark.sql.functions._
23+
import org.apache.spark.sql.SparkSession
3224

3325
/**
3426
* The helper object for transforming graphs through the definitions of their

spark/src/main/scala/com/alibaba/graphar/graph/GraphWriter.scala

Lines changed: 1 addition & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -16,22 +16,14 @@
1616

1717
package com.alibaba.graphar.graph
1818

19-
import com.alibaba.graphar.{
20-
AdjListType,
21-
GraphInfo,
22-
VertexInfo,
23-
EdgeInfo,
24-
GeneralParams
25-
}
19+
import com.alibaba.graphar.{AdjListType, GraphInfo, GeneralParams}
2620
import com.alibaba.graphar.writer.{VertexWriter, EdgeWriter}
2721
import com.alibaba.graphar.util.IndexGenerator
2822
import com.alibaba.graphar.util.Utils
2923

3024
import org.apache.spark.sql.{DataFrame, SparkSession}
3125
import org.apache.spark.sql.types._
32-
import org.apache.spark.sql.functions._
3326

34-
import org.apache.hadoop.conf.Configuration
3527
import org.apache.hadoop.fs.{FileSystem, Path}
3628
import java.io.{BufferedWriter, OutputStreamWriter}
3729

spark/src/main/scala/com/alibaba/graphar/reader/EdgeReader.scala

Lines changed: 1 addition & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -17,19 +17,11 @@
1717
package com.alibaba.graphar.reader
1818

1919
import com.alibaba.graphar.util.{IndexGenerator, DataFrameConcat}
20-
import com.alibaba.graphar.{
21-
GeneralParams,
22-
EdgeInfo,
23-
FileType,
24-
AdjListType,
25-
PropertyGroup
26-
}
27-
import com.alibaba.graphar.datasources._
20+
import com.alibaba.graphar.{EdgeInfo, FileType, AdjListType, PropertyGroup}
2821
import com.alibaba.graphar.util.FileSystem
2922

3023
import org.apache.spark.sql.{DataFrame, SparkSession}
3124
import org.apache.spark.sql.types._
32-
import org.apache.spark.sql.functions._
3325

3426
/**
3527
* Reader for edge chunks.

spark/src/main/scala/com/alibaba/graphar/reader/VertexReader.scala

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -17,14 +17,11 @@
1717
package com.alibaba.graphar.reader
1818

1919
import com.alibaba.graphar.util.{IndexGenerator, DataFrameConcat}
20-
import com.alibaba.graphar.{GeneralParams, VertexInfo, FileType, PropertyGroup}
21-
import com.alibaba.graphar.datasources._
20+
import com.alibaba.graphar.{VertexInfo, PropertyGroup}
2221
import com.alibaba.graphar.util.FileSystem
2322

2423
import org.apache.spark.sql.{DataFrame, SparkSession}
2524
import org.apache.spark.sql.types._
26-
import org.apache.spark.sql.functions._
27-
import org.apache.spark.sql.Row
2825

2926
/**
3027
* Reader for vertex chunks.

spark/src/main/scala/com/alibaba/graphar/util/DataFrameConcat.scala

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -16,11 +16,8 @@
1616

1717
package com.alibaba.graphar.util
1818

19-
import com.alibaba.graphar.GeneralParams
20-
21-
import org.apache.spark.sql.{DataFrame, SparkSession}
19+
import org.apache.spark.sql.DataFrame
2220
import org.apache.spark.sql.types._
23-
import org.apache.spark.sql.functions._
2421
import org.apache.spark.sql.Row
2522
import org.apache.spark.rdd.RDD
2623

spark/src/main/scala/com/alibaba/graphar/util/FileSystem.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@ package com.alibaba.graphar.util
1919
import org.json4s._
2020
import org.json4s.jackson.Serialization.write
2121

22-
import org.apache.spark.sql.SparkSession
2322
import org.apache.spark.sql.DataFrame
2423
import org.apache.hadoop.fs.Path
2524
import org.apache.hadoop.conf.Configuration

spark/src/main/scala/com/alibaba/graphar/util/IndexGenerator.scala

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -18,15 +18,11 @@ package com.alibaba.graphar.util
1818

1919
import com.alibaba.graphar.GeneralParams
2020

21-
import org.apache.spark.sql.SparkSession
2221
import org.apache.spark.sql.types._
23-
import org.apache.spark.sql.functions._
2422
import org.apache.spark.sql.DataFrame
2523
import org.apache.spark.sql.Row
26-
import org.apache.spark.rdd.RDD
2724

2825
import scala.collection.SortedMap
29-
import scala.collection.mutable.ArrayBuffer
3026

3127
/**
3228
* IndexGenerator is an object to help generating the indices for vertex/edge

spark/src/main/scala/com/alibaba/graphar/util/Patitioner.scala

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,6 @@
1616

1717
package com.alibaba.graphar.util
1818

19-
import org.apache.spark.sql.types._
2019
import org.apache.spark.Partitioner
2120

2221
/**

spark/src/main/scala/com/alibaba/graphar/writer/EdgeWriter.scala

Lines changed: 1 addition & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -16,11 +16,7 @@
1616

1717
package com.alibaba.graphar.writer
1818

19-
import com.alibaba.graphar.util.{
20-
FileSystem,
21-
ChunkPartitioner,
22-
EdgeChunkPartitioner
23-
}
19+
import com.alibaba.graphar.util.{FileSystem, EdgeChunkPartitioner}
2420
import com.alibaba.graphar.{
2521
GeneralParams,
2622
EdgeInfo,
@@ -38,8 +34,6 @@ import org.apache.spark.sql.types.{
3834
StructType,
3935
StructField
4036
}
41-
import org.apache.spark.util.Utils
42-
import org.apache.spark.rdd.RDD
4337
import org.apache.spark.sql.functions._
4438

4539
import scala.collection.SortedMap
@@ -54,7 +48,6 @@ object EdgeWriter {
5448
adjListType: AdjListType.Value,
5549
vertexNumOfPrimaryVertexLabel: Long
5650
): (DataFrame, Seq[DataFrame], Array[Long], Map[Long, Int]) = {
57-
import spark.implicits._
5851
val edgeSchema = edgeDf.schema
5952
val colName = if (
6053
adjListType == AdjListType.ordered_by_source || adjListType == AdjListType.unordered_by_source

spark/src/main/scala/com/alibaba/graphar/writer/VertexWriter.scala

Lines changed: 2 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -16,25 +16,14 @@
1616

1717
package com.alibaba.graphar.writer
1818

19-
import com.alibaba.graphar.util.{FileSystem, ChunkPartitioner, IndexGenerator}
20-
import com.alibaba.graphar.{
21-
GeneralParams,
22-
VertexInfo,
23-
FileType,
24-
AdjListType,
25-
PropertyGroup
26-
}
19+
import com.alibaba.graphar.util.{FileSystem, ChunkPartitioner}
20+
import com.alibaba.graphar.{GeneralParams, VertexInfo, PropertyGroup}
2721

28-
import org.apache.spark.sql.SparkSession
2922
import org.apache.spark.sql.types._
3023
import org.apache.spark.sql.functions._
3124
import org.apache.spark.sql.DataFrame
32-
import org.apache.spark.sql.Row
33-
import org.apache.spark.rdd.RDD
34-
import org.apache.spark.HashPartitioner
3525
import org.apache.spark.sql.types.{LongType, StructField}
3626

37-
import scala.collection.SortedMap
3827
import scala.collection.mutable.ArrayBuffer
3928

4029
/** Helper object for VertexWriter class. */

spark/src/test/scala/com/alibaba/graphar/ComputeExample.scala

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -18,13 +18,8 @@ package com.alibaba.graphar
1818

1919
import com.alibaba.graphar.reader.{VertexReader, EdgeReader}
2020

21-
import java.io.{File, FileInputStream}
22-
import scala.beans.BeanProperty
23-
import org.yaml.snakeyaml.Yaml
24-
import org.yaml.snakeyaml.constructor.Constructor
25-
import org.apache.spark.sql.{DataFrame, SparkSession}
21+
import org.apache.spark.sql.SparkSession
2622
import org.apache.spark.graphx._
27-
import org.apache.spark.rdd.RDD
2823
import org.scalatest.funsuite.AnyFunSuite
2924

3025
class ComputeExampleSuite extends AnyFunSuite {

spark/src/test/scala/com/alibaba/graphar/TestGraphInfo.scala

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -16,11 +16,7 @@
1616

1717
package com.alibaba.graphar
1818

19-
import org.yaml.snakeyaml.Yaml
20-
import org.yaml.snakeyaml.constructor.Constructor
21-
import scala.beans.BeanProperty
2219
import org.scalatest.funsuite.AnyFunSuite
23-
import org.apache.hadoop.fs.{Path, FileSystem}
2420
import org.apache.spark.sql.SparkSession
2521

2622
class GraphInfoSuite extends AnyFunSuite {

0 commit comments

Comments
 (0)