@@ -19,8 +19,6 @@ package org.apache.spark.sql.delta
19
19
import java .io .File
20
20
import java .net .URI
21
21
22
- import org .apache .spark .SparkThrowable
23
-
24
22
// scalastyle:off import.ordering.noEmptyLine
25
23
import org .apache .spark .sql .delta .DeltaOperations .Delete
26
24
import org .apache .spark .sql .delta .commands .DeltaGenerateCommand
@@ -33,10 +31,12 @@ import org.apache.hadoop.fs._
33
31
import org .apache .hadoop .fs .permission .FsPermission
34
32
import org .apache .hadoop .util .Progressable
35
33
34
+ import org .apache .spark .SparkThrowable
36
35
import org .apache .spark .sql ._
37
36
import org .apache .spark .sql .catalyst .TableIdentifier
38
37
import org .apache .spark .sql .functions ._
39
38
import org .apache .spark .sql .test .SharedSparkSession
39
+ // scalastyle:on import.ordering.noEmptyLine
40
40
41
41
class DeltaGenerateSymlinkManifestSuite
42
42
extends DeltaGenerateSymlinkManifestSuiteBase
@@ -601,7 +601,6 @@ trait DeltaGenerateSymlinkManifestSuiteBase extends QueryTest
601
601
subClass = ExistingDeletionVectorsWithIncrementalManifestGeneration ) {
602
602
setEnabledIncrementalManifest(tablePath, enabled = true )
603
603
}
604
-
605
604
// Run optimize to delete the DVs and rewrite the data files
606
605
withSQLConf(DeltaSQLConf .DELTA_OPTIMIZE_MAX_DELETED_ROWS_RATIO .key -> " 0.00001" ) {
607
606
spark.sql(s " OPTIMIZE delta.` $tablePath` " )
0 commit comments