@@ -4,8 +4,6 @@ import is.hail.backend.Backend
44import is .hail .backend .spark .SparkBackend
55import is .hail .expr .ir .functions .IRFunctionRegistry
66import is .hail .io .fs .FS
7- import is .hail .io .vcf ._
8- import is .hail .types .virtual ._
97import is .hail .utils ._
108
119import scala .reflect .ClassTag
@@ -17,9 +15,6 @@ import org.apache.log4j.{LogManager, PropertyConfigurator}
1715import org .apache .spark ._
1816import org .apache .spark .executor .InputMetrics
1917import org .apache .spark .rdd .RDD
20- import org .json4s .Extraction
21- import org .json4s .jackson .JsonMethods
22- import sourcecode .Enclosing
2318
2419case class FilePartition (index : Int , file : String ) extends Partition
2520
@@ -42,8 +37,6 @@ object HailContext {
4237
4338 def backend : Backend = get.backend
4439
45- def sparkBackend (implicit E : Enclosing ): SparkBackend = get.backend.asSpark
46-
4740 def configureLogging (logFile : String , quiet : Boolean , append : Boolean ): Unit = {
4841 org.apache.log4j.helpers.LogLog .setInternalDebugging(true )
4942 org.apache.log4j.helpers.LogLog .setQuietMode(false )
@@ -194,12 +187,4 @@ class HailContext private (
194187 : Array [(String , Array [String ])] =
195188 fileAndLineCounts(fs : FS , regex, files, maxLines).mapValues(_.map(_.value)).toArray
196189
197- def parseVCFMetadata (fs : FS , file : String ): Map [String , Map [String , Map [String , String ]]] =
198- LoadVCF .parseHeaderMetadata(fs, Set .empty, TFloat64 , file)
199-
200- def pyParseVCFMetadataJSON (fs : FS , file : String ): String = {
201- val metadata = LoadVCF .parseHeaderMetadata(fs, Set .empty, TFloat64 , file)
202- implicit val formats = defaultJSONFormats
203- JsonMethods .compact(Extraction .decompose(metadata))
204- }
205190}
0 commit comments