spark functions 源码
spark functions 代码
文件路径:/connector/protobuf/src/main/scala/org/apache/spark/sql/protobuf/functions.scala
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.sql.protobuf
import scala.collection.JavaConverters._
import org.apache.spark.annotation.Experimental
import org.apache.spark.sql.Column
// scalastyle:off: object.name
object functions {
// scalastyle:on: object.name
/**
* Converts a binary column of Protobuf format into its corresponding catalyst value. The
* specified schema must match actual schema of the read data, otherwise the behavior is
* undefined: it may fail or return arbitrary result. To deserialize the data with a compatible
* and evolved schema, the expected Protobuf schema can be set via the option protoSchema.
*
* @param data
* the binary column.
* @param descFilePath
* the protobuf descriptor in Message GeneratedMessageV3 format.
* @param messageName
* the protobuf message name to look for in descriptorFile.
* @since 3.4.0
*/
@Experimental
def from_protobuf(
data: Column,
descFilePath: String,
messageName: String,
options: java.util.Map[String, String]): Column = {
new Column(
ProtobufDataToCatalyst(data.expr, descFilePath, messageName, options.asScala.toMap))
}
/**
* Converts a binary column of Protobuf format into its corresponding catalyst value. The
* specified schema must match actual schema of the read data, otherwise the behavior is
* undefined: it may fail or return arbitrary result. To deserialize the data with a compatible
* and evolved schema, the expected Protobuf schema can be set via the option protoSchema.
*
* @param data
* the binary column.
* @param descFilePath
* the protobuf descriptor in Message GeneratedMessageV3 format.
* @param messageName
* the protobuf MessageName to look for in descriptorFile.
* @since 3.4.0
*/
@Experimental
def from_protobuf(data: Column, descFilePath: String, messageName: String): Column = {
new Column(ProtobufDataToCatalyst(data.expr, descFilePath, messageName, Map.empty))
}
/**
* Converts a column into binary of protobuf format.
*
* @param data
* the data column.
* @param descFilePath
* the protobuf descriptor in Message GeneratedMessageV3 format.
* @param messageName
* the protobuf MessageName to look for in descriptorFile.
* @since 3.4.0
*/
@Experimental
def to_protobuf(data: Column, descFilePath: String, messageName: String): Column = {
new Column(CatalystDataToProtobuf(data.expr, descFilePath, messageName))
}
}
相关信息
相关文章
spark CatalystDataToProtobuf 源码
0
赞
- 所属分类: 前端技术
- 本文标签:
热门推荐
-
2、 - 优质文章
-
3、 gate.io
-
8、 golang
-
9、 openharmony
-
10、 Vue中input框自动聚焦