geoanalytics.sql.functions

aggr_convex_hull

geoanalytics.sql.functions.aggr_convex_hull(geometry)

Operates on a grouped DataFrame and returns the convex hull of geometries in each group. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_ConvexHull

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Polygon column representing the convex hull of all of the geometries.

Return type

pyspark.sql.Column

aggr_intersection

geoanalytics.sql.functions.aggr_intersection(geometry)

Operates on a grouped DataFrame and returns the intersection of geometries in each group. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement. An empty geometry is returned when no geometries intersect.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_Intersection

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the intersection of all of the geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

aggr_linestring

geoanalytics.sql.functions.aggr_linestring(point, order_by)

Operates on a grouped DataFrame and returns a linestring containing the points ordered by the order_by column. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_Linestring

Parameters
  • point (pyspark.sql.Column) – Point geometry column.

  • order_by (pyspark.sql.Column) – Column to sort by

Returns

Linestring column representing the sorted input points.

Return type

pyspark.sql.Column

aggr_mean_center

geoanalytics.sql.functions.aggr_mean_center(geometry, weight=None)

Operates on a grouped DataFrame and returns the weighted aggregated centroid (mean center) of geometries in each group. You can optionally specify a numeric weight column which is used to weight locations according to their relative importance. The default is unweighted. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_MeanCenter

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • weight (pyspark.sql.Column, optional) – Used to weight locations according to their relative importance, defaults to None. Can be a LongType, DoubleType or StringType column with numeric values.

Returns

Point column representing the weighted aggregate centroid.

Return type

pyspark.sql.Column

aggr_stdev_ellipse

geoanalytics.sql.functions.aggr_stdev_ellipse(geometry, num_stdev=1.0, weight=None, min_records=2)

Operates on a grouped DataFrame and returns the weighted aggregate standard deviational ellipse of geometries in each group. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_StdevEllipse

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • num_stdev (double, optional) – Size of the returned ellipse in standard deviations, defaults to 1.

  • weight (pyspark.sql.Column, optional) – Weights locations according to their relative importance, defaults to None. Can be a LongType, DoubleType or StringType column with numeric values.

  • min_records (int, optional) – Number of geometries that must be considered for a standard deviation to be calculated, defaults to 2.

Returns

Polygon column representing the weighted aggregate Standard Deviational Ellipse.

Return type

pyspark.sql.Column

aggr_union

geoanalytics.sql.functions.aggr_union(geometry)

Operates on a grouped DataFrame and returns the union of geometries in each group. All geometries are required to have the same type. For example having point, linestring, and polygon geometry types in the same column is not supported. You can group your DataFrame using DataFrame.groupBy() or with a GROUP BY clause in a SQL statement.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Aggr_Union

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the union of all of the geometries. The returned geometry column type will be the same type as the input geometries except in the case of point input which will return multipoint.

Return type

pyspark.sql.Column

area

geoanalytics.sql.functions.area(geometry)

Returns a double column with the planar area of each geometry in the input column. The unit of the area calculation is the same as the units of the input geometries. For example, if you have polygons in a spatial reference that uses meters, the result will be in square meters. If your input geometries are in a geographic coordinate system, it is recommended that you use ST_GeodesicArea. Geometry types other than polygon will return 0.0 for the area.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Area

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the area.

Return type

pyspark.sql.Column

as_binary

geoanalytics.sql.functions.as_binary(geometry)

Returns the well-known binary (WKB) representation of the geometry as a binary column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_AsBinary

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BinaryType column with the Well-Known Binary (WKB) representation of the geometry.

Return type

pyspark.sql.Column

as_esri_json

geoanalytics.sql.functions.as_esri_json(geometry)

Returns the Esri JSON representation of the geometry as a string column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_AsEsriJSON

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

StringType column with the Esri JSON representation of the geometry.

Return type

pyspark.sql.Column

as_geojson

geoanalytics.sql.functions.as_geojson(geometry)

Returns the GeoJSON representation of the geometry as a string column. Geometries that have an m-value and no z-coordinate will only return x,y coordinates.

Refer to the GeoAnalytics guide for examples and usage notes: ST_AsGeoJSON

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

StringType column with the GeoJSON representation of the geometry.

Return type

pyspark.sql.Column

as_shape

geoanalytics.sql.functions.as_shape(geometry)

Returns the shapefile representation of the geometry as a binary column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_AsShape

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BinaryType column with the shapefile representation of the geometry.

Return type

pyspark.sql.Column

as_text

geoanalytics.sql.functions.as_text(geometry)

Returns the well-known text (WKT) representation of the geometry as a string column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_AsText

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

StringType column with the well-known text (WKT) representation of the geometry.

Return type

pyspark.sql.Column

azimuth

geoanalytics.sql.functions.azimuth(geometry1, geometry2)

Returns a double column representing the normalized azimuth in degrees. The output azimuth angle is heading from the first geometry to the second geometry. The angle is referenced from the north and is positive clockwise. This function requires that the first geometry column have a spatial reference set.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Azimuth

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the normalized azimuth in degrees.

Return type

pyspark.sql.Column

bbox_intersects

geoanalytics.sql.functions.bbox_intersects(geometry, xmin, ymin, xmax, ymax)

Returns a boolean column where the result is True if the geometry intersects the defined envelope; otherwise, it returns False. The four numeric values represent the minimum and maximum x,y coordinates of an axis-aligned rectangle, also known as an envelope. The x,y coordinates should be specified in the same units as the input geometry column. For example, if the input geometry is in a spatial reference that uses degrees, xmin, ymin, xmax, and ymax should all be in degrees.

Refer to the GeoAnalytics guide for examples and usage notes: ST_BboxIntersects

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • xmin (int/float) – The minimum x-coordinate point for the Envelope.

  • ymin (int/float) – The minimum y-coordinate point for the Envelope.

  • xmax (int/float) – The maximum x-coordinate point for the Envelope.

  • ymax (int/float) – The maximum y-coordinate point for the Envelope.

Returns

BooleanType column. True if geometry intersects the Envelope, False otherwise.

Return type

pyspark.sql.Column

bin_center

geoanalytics.sql.functions.bin_center(bin)

Returns a point column representing the center point of each bin.

Refer to the GeoAnalytics guide for examples and usage notes: ST_BinCenter

Parameters

bin (pyspark.sql.Column) – Spatial Bin (bin2d) column with the key for the bin.

Returns

Point column representing the center point for the bin associated with the given key.

Return type

pyspark.sql.Column

bin_geometry

geoanalytics.sql.functions.bin_geometry(bin)

Returns a polygon column representing the geometry of each bin.

Refer to the GeoAnalytics guide for examples and usage notes: ST_BinGeometry

Parameters

bin (pyspark.sql.Column) – Spatial Bin (bin2d) column with the key for the bin.

Returns

Polygon column representing the polygon for the bin associated with the given key.

Return type

pyspark.sql.Column

bin_id

geoanalytics.sql.functions.bin_id(bin)

Returns a long column representing the unique identifier of each bin.

Refer to the GeoAnalytics guide for examples and usage notes: ST_BinId

Parameters

bin (pyspark.sql.Column) – Spatial Bin (bin2d) column with the key for the bin.

Returns

LongType column representing the ID of the given key.

Return type

pyspark.sql.Column

boundary

geoanalytics.sql.functions.boundary(geometry)

Returns a geometry column representing the topological boundary of the given geometry. The function will return a linestring if the input is a polygon and a multipoint if the input is a linestring. Point and multipoint inputs are not supported.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Boundary

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the boundary. The returned geometry column type will be multipoint for linestring and multilinestring input and linestring for polygon and multipolygon input.

Return type

pyspark.sql.Column

buffer

geoanalytics.sql.functions.buffer(geometry, distance)

Returns a polygon column with buffer polygons that represent the area that is less than or equal to the specified planar distance from each input geometry. The distance can be specified as a single value or a numeric column. The distance value should be in the same units as the input geometry. For example, if your input geometry is in a spatial reference that uses meters, you should specify the distance in meters. To create a buffer polygon using geodesic distance calculations use ST_GeodesicBuffer.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Buffer

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • distance (pyspark.sql.Column/int/float) – Distance used to create the buffer. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value.

Returns

Polygon column representing the buffer around the input geometry.

Return type

pyspark.sql.Column

cast

geoanalytics.sql.functions.cast(geometry, geometry_type)

Returns a geometry column that contains the input geometries cast to the geometry type specified by the string value. The string value can be ‘point’, ‘multipoint’, ‘linestring’, ‘polygon’, or ‘geometry’. The function will return null when the input geometries cannot be cast to the specified type.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Cast

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • geometry_type (str) – Geometry type to cast the input geometry to.

Returns

Geometry column representing the cast geometry type.

Return type

pyspark.sql.Column

centerline

geoanalytics.sql.functions.centerline(geometry)

Creates a centerline of a polygon geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Centerline

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Geometry column with the centerline of the polygon feature.

Return type

pyspark.sql.Column

centroid

geoanalytics.sql.functions.centroid(geometry)

Returns a point column that represents the centroid of each input geometry. The result point is not guaranteed to be on the surface of the geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Centroid

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Point column representing the centroid.

Return type

pyspark.sql.Column

contains

geoanalytics.sql.functions.contains(geometry1, geometry2)

Returns a boolean column where the result is True if the geometry in the first column completely contains the second; otherwise, it is False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Contains

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 contains geometry2, False otherwise.

Return type

pyspark.sql.Column

closest_point

geoanalytics.sql.functions.closest_point(geometry1, geometry2)

Returns a point column representing the point on the first geometry that is closest to the second geometry. This function calculates the planar distance between the two geometries to identify the closest point on the first geometry in relation to the second geometry. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations. If the two input geometry columns are in different spatial references, the spatial reference of the output geometry would be the same as the first geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_ClosestPoint

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Point column representing the closest point.

Return type

pyspark.sql.Column

convex_hull

geoanalytics.sql.functions.convex_hull(geometry)

Returns a geometry column that represents the convex hull of the input geometries in each record. A convex hull is the smallest geometry having only interior angles measuring less than 180° that encloses each input geometry. For multipoint, linestring, and polygon geometries the result will be a polygon. For point geometries, the result is a point. The result column will always have the generic geometry type.

Refer to the GeoAnalytics guide for examples and usage notes: ST_ConvexHull

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Generic geometry column representing the convex hull of the given geometry.

Return type

pyspark.sql.Column

coord_dim

geoanalytics.sql.functions.coord_dim(geometry)

Returns an integer column representing the dimensionality of the coordinates in the input geometry. For example, an input geometry with x,y coordinates only will return 2, while a geometry with x,y,z coordinates will return 3.

Refer to the GeoAnalytics guide for examples and usage notes: ST_CoordDim

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

IntegerType column representing the dimensionality.

Return type

pyspark.sql.Column

create_distance

geoanalytics.sql.functions.create_distance(value, unit)

Returns a struct column representing a distance.

Refer to the GeoAnalytics guide for examples and usage notes: ST_CreateDistance

Parameters
  • value (pyspark.sql.Column/float/int) – A numeric value or column of numeric values.

  • unit (pyspark.sql.Column/str) – The distance unit. Choose from Meters, Kilometers, Feet, Yards, Miles, or NauticalMiles.

Returns

StructType column representing the distance.

Return type

pyspark.sql.Column

create_duration

geoanalytics.sql.functions.create_duration(value, unit)

Returns a struct column representing a duration.

Refer to the GeoAnalytics guide for examples and usage notes: ST_CreateDuration

Parameters
  • value (pyspark.sql.Column/int) – An integer value or column of integer values.

  • unit (pyspark.sql.Column/str) – The duration unit. Choose from Milliseconds, Seconds, Minutes, Hours, or Days.

Returns

StructType column representing the duration.

Return type

pyspark.sql.Column

crosses

geoanalytics.sql.functions.crosses(geometry1, geometry2)

Returns a boolean column where the result is True if the two geometries cross; otherwise, it returns False. Two geometries cross when their intersection is not empty and is not equal to either of the geometries. The intersection must also have a dimensionality less than the maximum dimension of the two input geometries.

This function is only relevant for the following combinations of geometries:

  • multipoint/linestring

  • multipoint/polygon

  • linestring/polygon

  • linestring/multipoint

  • linestring/linestring

  • polygon/multipoint

  • polygon/linestring

For all other combinations the function will always return False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Crosses

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 crosses geometry2, False otherwise.

Return type

pyspark.sql.Column

densify

geoanalytics.sql.functions.densify(geometry, max_segment_length)

Returns a geometry column of densified geometries. This function adds vertices along linestrings and polygons such that every segment within the geometry is no longer than max_segment_length with planar distance calculation. The max_segment_length is in the same units as the input geometry and should be greater than zero. To densify geometry using geodesic distance calculation, use ST_GeodesicDensify.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Densify

Parameters
  • geometry (pyspark.sql.Column) – Polygon or linestring geometry column

  • max_segment_length (pyspark.sql.Column/int/float) – Maximum length of all planar segments in the resulting polygon or linestring.

Returns

Geometry column representing the densified linestrings or polygons with planar distance calculation.

Return type

pyspark.sql.Column

difference

geoanalytics.sql.functions.difference(geometry1, geometry2)

Returns a geometry column representing the parts of the first geometry that do not intersect the second geometry. The result column geometry type will be that of the first geometry. If the first geometry is completely contained in the second geometry, then null geometry is returned.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Difference

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the part of geometry1 that does not intersect geometry2.

Return type

pyspark.sql.Column

dimension

geoanalytics.sql.functions.dimension(geometry)

Returns an integer column representing the dimensionality of the input geometry. Points and multipoints have a dimension of 0, lines 1, and polygons 2.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Dimension

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

IntegerType column representing the dimensionality.

Return type

pyspark.sql.Column

disjoint

geoanalytics.sql.functions.disjoint(geometry1, geometry2)

Returns a boolean column where the result is True if the first and second geometry are disjoint; otherwise, it returns False. Two geometries are disjoint if they are not overlapping, touching or intersecting each other.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Disjoint

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 and geometry2 are disjoint, False otherwise.

Return type

pyspark.sql.Column

distance

geoanalytics.sql.functions.distance(geometry1, geometry2)

Returns a double column representing the planar distance between the two input geometries. For multipoints, lines, and polygons, the distance is calculated from the nearest point between the geometries. The result will be in the same units as the input geometry data. For example, if your input geometries are in a spatial reference that uses meters, the result values will be in meters. If your input geometries are in a geographic coordinate system, use ST_GeodesicDistance to calculate distance.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Distance

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the planar distance.

Return type

pyspark.sql.Column

dwithin

geoanalytics.sql.functions.dwithin(geometry1, geometry2, distance, geodesic=False)

Returns a boolean column where the result is True if the two geometries are spatially within the given distance; otherwise, it returns False. For multipoints, lines, and polygons, the distance is calculated from the nearest point between the geometries. You can optionally provide a boolean value that determines if geodesic distances will be used by the function. Planar distances will be used by default.

When using planar calculations, the distance value is specified in the same units as the input geometry. When geodesic is set to True, the distance value is specified in meters.

Refer to the GeoAnalytics guide for examples and usage notes: ST_DWithin

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

  • distance (pyspark.sql.Column/int/float) – Distance value to use. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value.

  • geodesic (bool, optional) – Geodesic distance will be used between geometries instead of planar distance, defaults to False.

Returns

BooleanType column. True if geometry1 and geometry2 are spatially within a given distance, False otherwise.

Return type

pyspark.sql.Column

end_point

geoanalytics.sql.functions.end_point(linestring)

Returns a point column representing the last point of the input linestring.

Refer to the GeoAnalytics guide for examples and usage notes: ST_EndPoint

Parameters

linestring (pyspark.sql.Column) – Linestring geometry column.

Returns

Point column representing the ending point.

Return type

pyspark.sql.Column

env_intersects

geoanalytics.sql.functions.env_intersects(geometry, *args, **kwargs)

Returns a boolean column where the result is True if the envelopes of two geometries spatially intersect; otherwise, it returns False.

Note

The behavior of ST_EnvIntersects was changed at version 1.2.0. ST_EnvIntersects will not support defining envelopes with minimum and maximum x,y coordinates in the future. Use ST_BboxIntersects to do this instead.

Refer to the GeoAnalytics guide for examples and usage notes: ST_EnvIntersects

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the envelopes of two geometries spatially intersect in 2D, False otherwise.

Return type

pyspark.sql.Column

envelope

geoanalytics.sql.functions.envelope(geometry)

Returns a polygon column representing an envelope for each geometry in the input column, where an envelope is the smallest rectangle that encompasses the input geometry and aligns to the x-axis and y-axis. To find the smallest rectangle that encompasses a geometry but is not axis-aligned, use ST_MinBoundingBox. If the input geometry is a single point, the function will create a degenerate polygon at the location of the point.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Envelope

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Polygon column representing the envelope.

Return type

pyspark.sql.Column

equals

geoanalytics.sql.functions.equals(geometry1, geometry2)

Returns a boolean column where the result is True if the first geometry and the second geometry are spatially equal; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Equals

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 and geometry2 are spatially equal, False otherwise.

Return type

pyspark.sql.Column

euclidean_distance

geoanalytics.sql.functions.euclidean_distance(linestring1, linestring2)

Returns a double column representing the Euclidean distance between the two input linestring geometries. The result will be in the same units as the input geometry data. For example, if the input geometries are in a spatial reference that uses meters, the result values will be in meters.

Refer to the GeoAnalytics guide for examples and usage notes: ST_EuclideanDistance

Parameters
  • linestring1 (pyspark.sql.Column) – Linestring geometry column.

  • linestring2 (pyspark.sql.Column) – Linestring geometry column.

Returns

DoubleType column representing the Euclidean distance between the two input linestrings.

Return type

pyspark.sql.Column

exterior_ring

geoanalytics.sql.functions.exterior_ring(polygon)

Returns a linestring column representing the exterior ring of the polygon. Multipart polygons will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_ExteriorRing

Parameters

polygon (pyspark.sql.Column) – Polygon geometry column.

Returns

Linestring column representing the exterior ring.

Return type

pyspark.sql.Column

flip

geoanalytics.sql.functions.flip(geometry, mode)

Flips the input geometry around an axis. There are three options:

  • X_AXIS - flips a geometry vertically around the horizontal axis.

  • Y_AXIS - flips a geometry horizontally around the vertical axis.

  • BOTH_AXES - flips a geometry horizontally around the vertical axis and vertically around the horizontal axis.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Flip

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • mode (str) – Choose from X_AXIS, Y_AXIS, or BOTH_AXES.

Returns

Geometry column with the flipped geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

fréchet_distance

geoanalytics.sql.functions.frechet_distance(linestring1, linestring2)

Returns a double column representing the discrete Fréchet distance between the two input linestring geometries.

The function can be used with or without the diacritic.

Refer to the GeoAnalytics guide for examples and usage notes: ST_FréchetDistance

Parameters
  • linestring1 (pyspark.sql.Column) – Linestring geometry column.

  • linestring2 (pyspark.sql.Column) – Linestring geometry column.

Returns

DoubleType column representing the discrete Fréchet distance between the two input linestrings.

Return type

pyspark.sql.Column

generalize

geoanalytics.sql.functions.generalize(geometry, tolerance)

Returns a geometry column that generalizes the input linestring or polygon geometry using the Douglas-Peucker algorithm with the specified tolerance. The result is the input geometry generalized to include only a subset of the original geometry’s vertices. Point and multipoint geometry types are not supported as input.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Generalize

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • tolerance (pyspark.sql.Column/int/float) – Numeric value that limits the distance the output geometry can differ from the input geometry. Can be a LongType, DoubleType or StringType column or a numeric value.

Returns

Geometry column with the generalized geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

geodesic_area

geoanalytics.sql.functions.geodesic_area(geometry)

Returns a double column containing the geodesic area of the input geometry in square meters. For point, multipoint, and linestring geometries this function will always return 0. This function is more accurate but less performant than ST_Area and requires that a spatial reference is set on the input geometry column. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicArea

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the geodesic area.

Return type

pyspark.sql.Column

geodesic_buffer

geoanalytics.sql.functions.geodesic_buffer(geometry, distance)

Returns a polygon column with buffer polygons representing the geodesic area that is less than or equal to the specified distance from each input geometry. The distance can be specified as a single value or a numeric column and should be specified in meters. The result will also be in meters. This function is more accurate but less performant than ST_Buffer and requires that a spatial reference is set on the input geometry column. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicBuffer

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • distance (pyspark.sql.Column/int/float) – Distance used to create the buffer. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value.

Returns

Polygon column representing the geodesic buffer around the input geometry.

Return type

pyspark.sql.Column

geodesic_closest_point

geoanalytics.sql.functions.geodesic_closest_point(geometry1, geometry2)

Returns a point column representing the point on the first geometry that is closest to the second geometry. This function calculates the geodesic distance between the two geometries to identify the closest point on the first geometry in relation to the second geometry. This function requires that a spatial reference is set on the input geometry columns. If the two geometry columns are in different spatial references, the function automatically transforms the second geometry into the spatial reference of the first. To learn more about the difference between planar and geodesic calculations, see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicClosestPoint

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Point column representing the geodesic closest point.

Return type

pyspark.sql.Column

geodesic_densify

geoanalytics.sql.functions.geodesic_densify(geometry, max_segment_length)

Returns a geometry column of densified geometries. This function adds vertices along polygons or linestrings to create densified approximations of geodesic segments with each segment being no longer than max_segment_length. The max_segment_length should be specified in meters and greater than zero. This function is more accurate but less performant than ST_Densify and requires that a spatial reference is set on the input geometry column. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicDensify

Parameters
  • geometry (pyspark.sql.Column) – Polygon or linestring geometry column

  • max_segment_length (pyspark.sql.Column/int/float) – Maximum length in meters of all geodesic segments in the resulting polygon or linestring.

Returns

Geometry column representing the densified linestrings or polygons with geodesic distance calculation.

Return type

pyspark.sql.Column

geodesic_distance

geoanalytics.sql.functions.geodesic_distance(geometry1, geometry2)

Returns a double column representing the geodesic distance between the two input geometries in meters. For multipoints, lines, and polygons, the distance is calculated from the nearest point between the geometries. This function is more accurate but less performant than ST_Distance and requires that a spatial reference is set on at least the first input geometry column. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations. If the two geometry columns are in different spatial references, the function will automatically transform the second geometry into the spatial reference of the first.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicDistance

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the geodesic distance.

Return type

pyspark.sql.Column

geodesic_length

geoanalytics.sql.functions.geodesic_length(geometry)

Returns a double column that represents the geodesic length of the input geometry. The length is calculated in meters. For point and multipoint geometries this function will always return 0. For polygon geometries this function will return the geodesic length of the perimeter of the polygon. This function is more accurate but less performant than ST_Length and requires that a spatial reference is set on the input geometry column. To learn more about the difference between planar and geodesic calculations see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicLength

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the geodesic length.

Return type

pyspark.sql.Column

geodesic_shortest_line

geoanalytics.sql.functions.geodesic_shortest_line(geometry1, geometry2)

Returns a linestring column representing the shortest line that touches two geometries, using geodesic distance calculation. This function returns only one shortest line if there are more than one. If the two input geometries intersect, an empty line geometry is returned. This function requires that a spatial reference ID is set on the input geometry columns. If the two geometry columns are in different spatial references, the function automatically transforms the second geometry into the spatial reference of the first. To learn more about the difference between planar and geodesic calculations, see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeodesicShortestLine

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the geodesic shortest line.

Return type

pyspark.sql.Column

geohash_bin

geoanalytics.sql.functions.geohash_bin(geometry, precision)

Returns a bin column containing a single Geohash bin at the specified precision for each record in the input column. Use ST_BinGeometry to obtain the geometry of each result bin.

This function can also be called with a string column representing the ID of the bin (see ST_BinId). The bin ID will be cast to a bin column.

ST_GeohashBin requires the spatial reference of the geometry column to be GCS_WGS_1984 (EPSG:4326). If the input geometry is in a different spatial reference, this function automatically transforms the geometry into GCS_WGS_1984. To learn more about spatial references, see Coordinate systems and transformations.

Refer to the GeoAnalytics Engine guide for examples and usage notes: ST_GeohashBin

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • precision (int) – Numerical value representing the size of the Geohash bin.

Returns

Spatial bin (bin2d) column representing a single Geohash bin for each geometry.

Return type

pyspark.sql.Column

geohash_bins

geoanalytics.sql.functions.geohash_bins(geometry, precision, padding=0.0)

Returns an array column containing Geohash bins at the specified precision that cover the spatial extent of each record in the input column. You can optionally specify a numeric value for padding, which conceptually applies a buffer of the specified distance to the input geometry before creating the Geohash bins. The padding value is in meters. Use ST_BinGeometry to obtain the geometry of each result bin.

ST_GeohashBins requires the spatial reference to be set to GCS_WGS_1984 (EPSG:4326). If the input geometry is in a different spatial reference, this function automatically transforms the geometry into GCS_WGS_1984. To learn more about spatial references, see Coordinate systems and transformations.

Refer to the GeoAnalytics Engine guide for examples and usage notes: ST_GeohashBins

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • precision (int) – Numerical value representing the size of the Geohash bin.

  • padding (float, optional) – Numerical buffer value applied to the geometry before finding the intersecting bins, defaults to 0.0.

Returns

Array column representing an array of spatial bin (bin2d) Geohash bins.

Return type

pyspark.sql.Column

geom_from_binary

geoanalytics.sql.functions.geom_from_binary(wkb, sr=None)

Returns a geometry column. The input binary column must contain the well-known binary (WKB) representation of geometries. You can optionally specify a spatial reference for the result geometry column. The sr parameter value must be a valid SRID or WKT string. This function should only be used when you don’t know the geometry type represented by the input column or when the input column contains more than one geometry type. In other cases, use the function specific to the geometry type of your input data (i.e. ST_PointFromBinary, ST_LineFromBinary, ST_MPointFromBinary, or ST_PolyFromBinary).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromBinary

Parameters
  • wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned generic geometry, defaults to None.

Returns

Generic geometry column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

geom_from_esri_json

geoanalytics.sql.functions.geom_from_esri_json(json_str, sr=None)

Returns a geometry column. The input string column must contain the Esri JSON representation of geometries. You can optionally specify a spatial reference for the result geometry column. The sr parameter value must be a valid SRID or WKT string. Any spatial reference defined in the input strings will not be used. This function should only be used when you don’t know the geometry type represented by the input column or when the input column contains more than one geometry type. In other cases, use the function specific to the geometry type of your input data (i.e. ST_PointFromEsriJSON, ST_LineFromEsriJSON, ST_MPointFromEsriJSON, or ST_PolyFromEsriJSON).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromEsriJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the Esri JSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned generic geometry, defaults to None.

Returns

Generic geometry column from the Esri JSON representation.

Return type

pyspark.sql.Column

geom_from_geojson

geoanalytics.sql.functions.geom_from_geojson(json_str, sr=None)

Returns a geometry column. The input string column must contain the GeoJSON representation of geometries. You can optionally specify a spatial reference for the result geometry column. The sr parameter value must be a valid SRID or WKT string. Any spatial reference defined in the input strings will not be used. This function should only be used when you don’t know the geometry type represented by the input column or when the input column contains more than one geometry type. In other cases, use the function specific to the geometry type of your input data (i.e. ST_PointFromGeoJSON, ST_LineFromGeoJSON, ST_MPointFromGeoJSON, or ST_PolyFromGeoJSON).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromGeoJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the GeoJSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned generic geometry, defaults to None.

Returns

Generic geometry column from the GeoJSON representation.

Return type

pyspark.sql.Column

geom_from_shape

geoanalytics.sql.functions.geom_from_shape(shp, sr=None)

Returns a geometry column. The input binary column must contain the shapefile representation of geometries. You can optionally specify a spatial reference for the result geometry column. The sr parameter value must be a valid SRID or WKT string. This function should only be used when you don’t know the geometry type represented by the input column or when the input column contains more than one geometry type. In other cases, use the function specific to the geometry type of your input data (i.e. ST_PointFromShape, ST_LineFromShape, ST_MPointFromShape, or ST_PolyFromShape).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromShape

Parameters
  • shp (pyspark.sql.Column) – BinaryType column with the shapefile representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned generic geometry, defaults to None.

Returns

Generic geometry column from the shapefile representation.

Return type

pyspark.sql.Column

geom_from_text

geoanalytics.sql.functions.geom_from_text(wkt, sr=None)

Returns a geometry column. The string column must contain the well-known text (WKT) representation of geometries. You can optionally specify a spatial reference for the result geometry column. The sr parameter value must be a valid SRID or WKT string. This function should only be used when you don’t know the geometry type represented by the input column or when the input column contains more than one geometry type. In other cases, use the function specific to the geometry type of your input data (i.e. ST_PointFromText, ST_LineFromText, ST_MPointFromText, or ST_PolyFromText).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromText

Parameters
  • wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of geometries.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned generic geometry, defaults to None.

Returns

Generic geometry column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

geometries

geoanalytics.sql.functions.geometries(geometry)

Returns an array column. Multipoint geometries return an array of points, multipart linestrings return an array of single-path linestrings, and multipart polygons return an array of single-ring polygons.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Geometries

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Array column representing an array of single part geometries.

Return type

pyspark.sql.Column

geometry_n

geoanalytics.sql.functions.geometry_n(geometry, n)

Returns a geometry column. The output column contains the nth single-part geometry from a multipart geometry. When n=0, the first single-part geometry is returned. If the nth geometry doesn’t exist, null is returned.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeometryN

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • n (pyspark.sql.Column/int) – Index of the geometry to return. Can be an IntegerType column or an integer value.

Returns

Geometry column representing the nth geometry. The returned geometry column type will be the same type as the input geometries except in the case of multipoint input which will return point.

Return type

pyspark.sql.Column

geometry_type

geoanalytics.sql.functions.geometry_type(geometry)

Returns a string column. The string indicates the type of each input geometry (i.e. ‘Point’, ‘MultiPoint’, ‘Linestring’, or ‘Polygon’).

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeometryType

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

StringType column representing the geometry type.

Return type

pyspark.sql.Column

h3_bin

geoanalytics.sql.functions.h3_bin(geometry, h3_resolution)

Returns a bin column containing a single H3 bin at the specified resolution for each record in the input column. The centroid of the input geometry is guaranteed to intersect with the bin returned but is not necessarily coincident with the bin center. Use ST_BinGeometry to obtain the geometry of each result bin.

This function can also be called with a long column representing the ID of the bin (see ST_BinId). The bin ID will be cast to a bin column.

ST_H3Bin requires the spatial reference of the geometry column to be GCS_WGS_1984 (EPSG:4326). If the input geometry is in a different spatial reference, this function automatically transforms the geometry into GCS_WGS_1984. To learn more about spatial references, see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_H3Bin

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • h3_resolution (int) – H3 cell resolution, see H3 documentation for more information.

Returns

Spatial bin (bin2d) column representing a single H3 bin for each geometry.

Return type

pyspark.sql.Column

h3_bins

geoanalytics.sql.functions.h3_bins(geometry, h3_resolution, padding=0.0)

Returns an array column containing H3 bins at the specified resolution that cover the spatial extent of each record in the input column. You can optionally specify a numeric value for padding, which conceptually applies a buffer of the specified distance to the input geometry before creating the H3 bins. The padding value is in meters. Use ST_BinGeometry to obtain the geometry of each result bin.

ST_H3Bins requires the spatial reference to be set to GCS_WGS_1984 (EPSG:4326). If the input geometry is in a different spatial reference, this function automatically transforms the geometry into GCS_WGS_1984. To learn more about spatial references, see Coordinate systems and transformations.

Refer to the GeoAnalytics guide for examples and usage notes: ST_H3Bins

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • h3_resolution (int) – H3 cell resolution, see H3 documentation for more information.

  • padding (float, optional) – Numerical buffer value applied to the geometry before finding the intersecting bins, defaults to 0.0.

Returns

Array column representing an array of spatial bin (bin2d) H3 bins.

Return type

pyspark.sql.Column

hausdorff_distance

geoanalytics.sql.functions.hausdorff_distance(geometry1, geometry2)

Returns a double column representing the Hausdorff distance between the two input geometries.

Refer to the GeoAnalytics guide for examples and usage notes: ST_HausdorffDistance

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the Hausdorff distance between the two input geometries.

Return type

pyspark.sql.Column

hex_bin

geoanalytics.sql.functions.hex_bin(geometry, bin_size)

Returns a bin column containing a single hexagonal bin for each record in the input column. The specified bin size determines the height of each bin and is in the same units as the input geometry. The centroid of the input geometry is guaranteed to intersect with the bin returned but is not necessarily coincident with the bin center. Use ST_BinGeometry to obtain the geometry of each result bin.

This function can also be called with a long column representing the ID of the bin (see ST_BinId). The bin ID will be cast to a bin column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_HexBin

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • bin_size (int/float) – Numerical value representing the height of the hexagonal bin.

Returns

Spatial bin (bin2d) column representing a single hexagonal bin for each geometry.

Return type

pyspark.sql.Column

hex_bins

geoanalytics.sql.functions.hex_bins(geometry, bin_size, padding=0.0)

Returns an array column containing hexagonal bins that cover the spatial extent of each record in the input column. The specified bin size determines the height of each bin and is in the same units as the input geometry. You can optionally specify a numeric value for padding, which conceptually applies a buffer of the specified distance to the input geometry before creating the hexagonal bins.

Refer to the GeoAnalytics guide for examples and usage notes: ST_HexBins

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • bin_size (int/float) – Numerical value representing the height of the hexagonal bin.

  • padding (int/float/str, optional) – Numerical buffer value applied to the geometry before finding the intersecting bins, defaults to 0.0.

Returns

Array column representing an array of spatial bin (bin2d) hexagonal bins.

Return type

pyspark.sql.Column

interior_ring_n

geoanalytics.sql.functions.interior_ring_n(polygon, n)

Returns a linestring column. The output is the nth interior ring of the input polygon as a linestring. If there is more than one interior ring, the order of the interior rings is defined by the order in the input polygon. When n=0, the first interior ring is returned. If the index exceeds the number of interior rings in the polygon, null is returned. If the input is a multipolygon null is returned.

Refer to the GeoAnalytics guide for examples and usage notes: ST_InteriorRingN

Parameters
  • polygon (pyspark.sql.Column) – Polygon geometry column.

  • n (pyspark.sql.Column/int) – Index of the interior ring to return. Can be an IntegerType column or an integer value.

Returns

Linestring column representing the nth interior ring.

Return type

pyspark.sql.Column

intersection

geoanalytics.sql.functions.intersection(geometry1, geometry2, intersect_type=None)

Returns a geometry column containing the intersection of two input geometry records. You can optionally specify a string value that determines the geometry type of the result. The string can be one of: ‘multipoint’, ‘linestring’ or ‘polygon’. If no intersection type is specified, the function will return the same geometry type as the input geometry with the lowest dimension. For example, if you calculate the intersection of a polygon and a linestring the function will return a linestring.

The function will return an empty geometry if the two input geometries do not intersect or there is no intersection that matches the specified intersect type. If the intersection is a single point, the geometry type of the result column will be a multipoint.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Intersection

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

  • intersect_type (str, optional) – Sets the output geometry type, defaults to None.

Returns

Geometry column.

Return type

pyspark.sql.Column

intersects

geoanalytics.sql.functions.intersects(geometry1, geometry2)

Returns a boolean column where the result is True if the first geometry and the second geometry spatially intersect; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Intersects

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 and geometry2 spatially intersect in 2D, False otherwise.

Return type

pyspark.sql.Column

is_3d

geoanalytics.sql.functions.is_3d(geometry)

Returns a boolean column where the result is True if the geometry is three-dimensional; otherwise, it returns False. The geometry is considered three-dimensional if it has x,y,z coordinates.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Is3D

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the geometry is three-dimensional, False otherwise.

Return type

pyspark.sql.Column

is_closed

geoanalytics.sql.functions.is_closed(linestring)

Returns a boolean column where the result is True if the start and end point of a given linestring are coincident; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_IsClosed

Parameters

linestring (pyspark.sql.Column) – Linestring geometry column.

Returns

BooleanType column. True if the linestring is closed, False otherwise.

Return type

pyspark.sql.Column

is_empty

geoanalytics.sql.functions.is_empty(geometry)

Returns a boolean column where the result is True if the geometry is empty; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_IsEmpty

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the geometry is empty, False otherwise.

Return type

pyspark.sql.Column

is_measured

geoanalytics.sql.functions.is_measured(geometry)

Returns a boolean column where the result is True if the geometry has an m-value; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_IsMeasured

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the geometry has m-values, False otherwise.

Return type

pyspark.sql.Column

is_ring

geoanalytics.sql.functions.is_ring(geometry)

Returns a boolean column where the result is True if the input linestring is closed and simple; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_IsRing

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the linestring is both closed and simple, False otherwise.

Return type

pyspark.sql.Column

is_simple

geoanalytics.sql.functions.is_simple(geometry)

Returns a boolean column where the result is True if the given geometry is simple; otherwise, it returns False.

The criteria for simplicity vary for each geometry type and are as follows:

  • A point is always simple.

  • A multipoint is considered simple if no two points are coincident.

  • A linestring is considered simple if it does not cross the same point twice, except for start and end points. A multipart linestring is only considered simple if the parts do not intersect except at the start or end points of the parts.

  • A polygon or multipart polygon is considered simple if each ring does not cross the same point twice and no two rings cross or intersect except at a single point (i.e. they are tangent at a point but not a line).

Refer to the GeoAnalytics guide for examples and usage notes: ST_IsSimple

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if the geometry has self-intersection or self-tangency, False otherwise.

Return type

pyspark.sql.Column

length

geoanalytics.sql.functions.length(geometry)

Returns a double column representing the planar length of the input geometry. The length is calculated in the same units as the input geometry. For point and multipoint geometries the function will always return 0. For polygon geometries this function will return the length of the perimeter of the polygon.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Length

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the planar length.

Return type

pyspark.sql.Column

line_from_binary

geoanalytics.sql.functions.line_from_binary(wkb, sr=None)

Returns a linestring column. The input binary column must contain the well-known binary (WKB) representation of linestring geometries. You can optionally specify a spatial reference for the result linestring column. The sr parameter value must be a valid SRID or WKT string. If a linestring cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_LineFromBinary

Parameters
  • wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned linestring geometry, defaults to None.

Returns

Linestring column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

line_from_esri_json

geoanalytics.sql.functions.line_from_esri_json(json_str, sr=None)

Returns a linestring column. The input string column must contain the Esri JSON representation of linestring geometries. You can optionally specify a spatial reference for the result linestring column. The sr parameter value must be a valid SRID or WKT string. Any SRID defined in the input strings will not be used. If a linestring cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_LineFromEsriJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the Esri JSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned linestring geometry, defaults to None.

Returns

Linestring column from the Esri JSON representation.

Return type

pyspark.sql.Column

line_from_geojson

geoanalytics.sql.functions.line_from_geojson(json_str, sr=None)

Returns a linestring column. The input string column must contain the GeoJSON representation of linestring geometries. You can optionally specify a spatial reference for the result linestring column. The sr parameter value must be a valid SRID or WKT string. If a linestring cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_LineFromGeoJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the GeoJSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned linestring geometry, defaults to None.

Returns

Linestring column from the GeoJSON representation.

Return type

pyspark.sql.Column

line_from_shape

geoanalytics.sql.functions.line_from_shape(shp, sr=None)

Returns a linestring column. The input binary column must contain the shapefile representation of linestring geometries. You can optionally specify a spatial reference for the result linestring column. The sr parameter value must be a valid SRID or WKT string. If a linestring cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_LineFromShape

Parameters
  • shp (pyspark.sql.Column) – BinaryType column with the shapefile representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned linestring geometry, defaults to None.

Returns

Linestring column from the shapefile representation.

Return type

pyspark.sql.Column

line_from_text

geoanalytics.sql.functions.line_from_text(wkt, sr=None)

Returns a linestring column. The string column must contain the well-known text (WKT) representation of linestring geometries. You can optionally specify a spatial reference for the result linestring column. The sr parameter value must be a valid SRID or WKT string.

Refer to the GeoAnalytics guide for examples and usage notes: ST_LineFromText

Parameters
  • wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of linestring geometries.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned linestring geometry, defaults to None.

Returns

Linestring column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

linestring

geoanalytics.sql.functions.linestring(points)

Returns a linestring column. The input arrays must be arrays of point geometries. The function creates a linestring geometry by connecting the point geometries in the same order that they are stored in the input array.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Linestring

Parameters

points (pyspark.sql.Column) – Array of point geometries.

Returns

Linestring column representing the array of points.

Return type

pyspark.sql.Column

m

geoanalytics.sql.functions.m(point, new_value=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a point column and returns a double column containing the m-values of the input points. If a point does not have an m-value the function returns NaN.

Setter: Takes a point column and a numeric value and returns a point column containing the input points with the m-values set to the numeric value.

Refer to the GeoAnalytics guide for examples and usage notes: ST_M

Parameters
  • point (pyspark.sql.Column) – Point geometry column.

  • new_value (pyspark.sql.Column/int/float, optional) – M-value to set. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

Returns

  • Getter: DoubleType column representing the m-value.

  • Setter: Point column representing the updated m-value.

Return type

pyspark.sql.Column

make_point

geoanalytics.sql.functions.make_point(x, y, z=None, m=None)

Returns a point column. The two input columns must contain the x,y coordinates of the points respectively. You can optionally specify two additional input columns with z-coordinates and m-values. The spatial reference of the result column will always be 0 and should be set to a valid ID using ST_SRID.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MakePoint

Parameters
  • x (pyspark.sql.Column/int/float) – X-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (x-coordinates) or a numeric value.

  • y (pyspark.sql.Column/int/float) – Y-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (y-coordinates) or a numeric value.

  • z (pyspark.sql.Column/int/float, optional) – Z-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

  • m (pyspark.sql.Column/int/float, optional) – M-value for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

Returns

Point column representing the point geometries.

Return type

pyspark.sql.Column

max_m

geoanalytics.sql.functions.max_m(geometry)

Returns a double column containing the maximum m-value of each input geometry. If the input geometry does not have m-values the function will return NaN.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MaxM

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the maximum m-value for the envelope.

Return type

pyspark.sql.Column

max_x

geoanalytics.sql.functions.max_x(geometry)

Returns a double column containing the maximum x-coordinate of each input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MaxX

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the maximum x-coordinate for the envelope.

Return type

pyspark.sql.Column

max_y

geoanalytics.sql.functions.max_y(geometry)

Returns a double column containing the maximum y-coordinate of each input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MaxY

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the maximum y-coordinate for the envelope.

Return type

pyspark.sql.Column

max_z

geoanalytics.sql.functions.max_z(geometry)

Returns a double column containing the maximum z-coordinate of each input geometry. If the input geometry does not have z-coordinates the function will return NaN.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MaxZ

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the maximum z-coordinate for the envelope.

Return type

pyspark.sql.Column

min_bounding_box

geoanalytics.sql.functions.min_bounding_box(geometry, by_area=True)

Returns a polygon column containing a polygon for each geometry in the input column. The polygon is the smallest rectangle of arbitrary alignment that encompasses the input geometry. You can optionally provide a boolean value that determines how the rectangle will be created. There are two options:

  • True: Creates a rectangle with the minimum possible area. This is the default.

  • False: Creates a rectangle with the minimum possible width.

For point geometries, this function will return a degenerate polygon at the location of the point.

To find the minimum bounding box that aligns to the x and y axis, use ST_Envelope.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MinBoundingBox

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • by_area (pyspark.sql.Column/bool, optional) – BooleanType column or a boolean value. True minimizes the bounding box area, False minimizes the width. Defaults to True.

Returns

Polygon column representing the bounding box.

Return type

pyspark.sql.Column

min_m

geoanalytics.sql.functions.min_m(geometry)

Returns a double column containing the minimum m-value of each input geometry. If the input geometry does not have m-values the function will return NaN.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MinM

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the minimum m-value for the envelope.

Return type

pyspark.sql.Column

min_x

geoanalytics.sql.functions.min_x(geometry)

Returns a double column containing the minimum x-coordinate of each input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MinX

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the minimum x-coordinate for the envelope.

Return type

pyspark.sql.Column

min_y

geoanalytics.sql.functions.min_y(geometry)

Returns a double column containing the minimum y-coordinate of each input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MinY

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the minimum y-coordinate for the envelope.

Return type

pyspark.sql.Column

min_z

geoanalytics.sql.functions.min_z(geometry)

Returns a double column containing the minimum z-coordinate of each input geometry. If the input geometry does not have z-coordinates the function will return NaN.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MinZ

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

DoubleType column representing the minimum z-coordinate for the envelope.

Return type

pyspark.sql.Column

mpoint_from_binary

geoanalytics.sql.functions.mpoint_from_binary(wkb, sr=None)

Returns a multipoint column. The input binary column must contain the well-known binary (WKB) representation of multipoint geometries. You can optionally specify a spatial reference for the result multipoint column. The sr parameter value must be a valid SRID or WKT string. If amultipoint cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MPointFromBinary

Parameters
  • wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Multipoint column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

mpoint_from_esri_json

geoanalytics.sql.functions.mpoint_from_esri_json(json_str, sr=None)

Returns a multipoint column. The input string column must contain the Esri JSON representation of multipoint geometries. You can optionally specify a spatial reference for the result multipoint column. The sr parameter value must be a valid SRID or WKT string. Any SRID defined in the input strings will not be used. If a multipoint cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MPointFromEsriJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the Esri JSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Multipoint column from the Esri JSON representation.

Return type

pyspark.sql.Column

mpoint_from_geojson

geoanalytics.sql.functions.mpoint_from_geojson(json_str, sr=None)

Returns a multipoint column. The input string column must contain the GeoJSON representation of multipoint geometries. You can optionally specify a spatial reference for the result multipoint column. The sr parameter value must be a valid SRID or WKT string. If a multipoint cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MPointFromGeoJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the GeoJSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Multipoint column from the GeoJSON representation.

Return type

pyspark.sql.Column

mpoint_from_shape

geoanalytics.sql.functions.mpoint_from_shape(shp, sr=None)

Returns a multipoint column. The input binary column must contain the shapefile representation of multipoint geometries. You can optionally specify a spatial reference for the result multipoint column. The sr parameter value must be a valid SRID or WKT string. If a multipoint cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MPointFromShape

Parameters
  • shp (pyspark.sql.Column) – BinaryType column with the shapefile representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Multipoint column from the shapefile representation.

Return type

pyspark.sql.Column

mpoint_from_text

geoanalytics.sql.functions.mpoint_from_text(wkt, sr=None)

Returns a multipoint column. The string column must contain the well-known text (WKT) representation of multipoint geometries. You can optionally specify a spatial reference for the result multipoint column. The sr parameter value must be a valid SRID or WKT string.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MPointFromText

Parameters
  • wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of multipoint geometries.

  • sr (int/sr) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Multipoint column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

multilinestring

geoanalytics.sql.functions.multilinestring(*point_arrays)

Returns a linestring column. The input array column must contain an array of arrays of point geometries.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MultiLinestring

Parameters

point_arrays (pyspark.sql.Column) – Array of point geometry arrays.

Returns

Linestring column representing the array of points.

Return type

pyspark.sql.Column

multipoint

geoanalytics.sql.functions.multipoint(points)

Returns a multipoint column. The input array column must contain an array of point geometries.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MultiPoint

Parameters

points (pyspark.sql.Column) – Array of point geometries.

Returns

Multipoint column representing the array of points.

Return type

pyspark.sql.Column

multipolygon

geoanalytics.sql.functions.multipolygon(*point_arrays)

Returns a polygon column. The input column must contain an array of arrays of point geometries. The output polygon column represents the one or more rings created from the point arrays.

Refer to the GeoAnalytics guide for examples and usage notes: ST_MultiPolygon

Parameters

point_arrays (pyspark.sql.Column) – Array of point geometry arrays.

Returns

Polygon column representing the array of points.

Return type

pyspark.sql.Column

num_geometries

geoanalytics.sql.functions.num_geometries(geometry)

Returns an integer column representing the number of geometries in each record.

Refer to the GeoAnalytics guide for examples and usage notes: ST_NumGeometries

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

IntegerType column representing the number of geometries.

Return type

pyspark.sql.Column

num_interior_ring

geoanalytics.sql.functions.num_interior_ring(polygon)

OGC alias for ST_NumInteriorRings.

Refer to the GeoAnalytics guide for examples and usage notes: ST_NumInteriorRings

Parameters

polygon (pyspark.sql.Column) – Polygon Geometry column.

Returns

IntegerType column representing the number of interior rings.

Return type

pyspark.sql.Column

num_interior_rings

geoanalytics.sql.functions.num_interior_rings(polygon)

Returns an integer column representing the number of interior rings in the input polygon. The function will return null when the input is a multipart polygon.

Refer to the GeoAnalytics guide for examples and usage notes: ST_NumInteriorRings

Parameters

polygon (pyspark.sql.Column) – Polygon geometry column.

Returns

IntegerType column representing the number of interior rings.

Return type

pyspark.sql.Column

num_points

geoanalytics.sql.functions.num_points(geometry)

Returns an integer column representing the number of points in the input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_NumPoints

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

IntegerType column representing the number of points.

Return type

pyspark.sql.Column

overlaps

geoanalytics.sql.functions.overlaps(geometry1, geometry2)

Returns a boolean column where the result is True if the first geometry and the second geometry spatially overlap; otherwise, it returns False. Two geometries overlap when their intersection is the same geometry type as either of the inputs but not equal to either of the inputs.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Overlaps

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 and geometry2 spatially overlap, False otherwise.

Return type

pyspark.sql.Column

point

geoanalytics.sql.functions.point(x, y, sr=None)

Returns a point column. The two numeric columns or values must contain the x,y coordinates of the point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string.

To create point geometries with a z-coordinate and/or m-value, use ST_PointZ, ST_PointZM, ST_PointM, or ST_MakePoint.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Point

Parameters
  • x (pyspark.sql.Column/int/float) – X-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (x-coordinates) or a numeric value.

  • y (pyspark.sql.Column/int/float) – Y-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (y-coordinates) or a numeric value.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Point column representing the point geometries.

Return type

pyspark.sql.Column

point_from_binary

geoanalytics.sql.functions.point_from_binary(wkb, sr=None)

Returns a point column. The input binary column must contain the well-known binary (WKB) representation of point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string. If a point cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointFromBinary

Parameters
  • wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned point geometry, defaults to None.

Returns

Point column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

point_from_esri_json

geoanalytics.sql.functions.point_from_esri_json(json_str, sr=None)

Returns a point column. The input string column must contain the Esri JSON representation of point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string. Any SRID defined in the input strings will not be used. If a point cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointFromEsriJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the Esri JSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned point geometry, defaults to None.

Returns

Point column from the Esri JSON representation.

Return type

pyspark.sql.Column

point_from_geojson

geoanalytics.sql.functions.point_from_geojson(json_str, sr=None)

Returns a point column. The input string column must contain the GeoJSON representation of point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string. If a point cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointFromGeoJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the GeoJSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned point geometry, defaults to None.

Returns

Point column from the GeoJSON representation.

Return type

pyspark.sql.Column

point_from_shape

geoanalytics.sql.functions.point_from_shape(shp, sr=None)

Returns a point column. The input binary column must contain the shapefile representation of point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string. If a point cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointFromShape

Parameters
  • shp (pyspark.sql.Column) – BinaryType column with the shapefile representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned point geometry, defaults to None.

Returns

Point column from the shapefile representation.

Return type

pyspark.sql.Column

point_from_text

geoanalytics.sql.functions.point_from_text(wkt, sr=None)

Returns a point column. The string column must contain the well-known text (WKT) representation of point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointFromText

Parameters
  • wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of point geometries.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned point geometry, defaults to None.

Returns

Point column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

point_m

geoanalytics.sql.functions.point_m(x, y, m, sr=None)

Returns a point column. The three numeric columns or values must contain the x,y coordinates and m-values of the point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string.

To create point geometries without m-values or z-coordinates use ST_Point. To create point geometries with z-coordinates use ST_PointZ, ST_PointZM, or ST_MakePoint.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointM

Parameters
  • x (pyspark.sql.Column/int/float) – X-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (x-coordinates) or a numeric value.

  • y (pyspark.sql.Column/int/float) – Y-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (y-coordinates) or a numeric value.

  • m (pyspark.sql.Column/int/float) – M-value for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Point column representing the point geometries.

Return type

pyspark.sql.Column

point_n

geoanalytics.sql.functions.point_n(geometry, n)

Returns a point column. The output column represents the nth point in the input geometry, where 0 is the first point. If the nth point does not exist the function returns null. This function always returns null for multipart linestrings, and multipart polygons.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointN

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • n (pyspark.sql.Column/int) – Index of the point to return. Can be an IntegerType column or an integer value.

Returns

Point column representing the nth point.

Return type

pyspark.sql.Column

point_on_surface

geoanalytics.sql.functions.point_on_surface(geometry)

Returns a point column. The function returns a point that lies on the surface of linestring or polygon geometries.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointOnSurface

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Point column representing a point that lies on the surface.

Return type

pyspark.sql.Column

point_z

geoanalytics.sql.functions.point_z(x, y, z, sr=None)

Returns a point column. The three numeric columns or values must contain the x,y,z coordinates of the point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string.

To create point geometries without m-values or z-coordinates use ST_Point. To create point geometries with m-values use ST_PointM, ST_PointZM, or ST_MakePoint.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointZ

Parameters
  • x (pyspark.sql.Column/int/float) – X-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (x-coordinates) or a numeric value.

  • y (pyspark.sql.Column/int/float) – Y-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (y-coordinates) or a numeric value.

  • z (pyspark.sql.Column/int/float) – Z-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Point column representing the point geometries.

Return type

pyspark.sql.Column

point_zm

geoanalytics.sql.functions.point_zm(x, y, z, m, sr=None)

Returns a point column. The four numeric columns or values must contain the x,y,z coordinates and m-values of the point geometries. You can optionally specify a spatial reference for the result point column. The sr parameter value must be a valid SRID or WKT string.

To create point geometries without m-values or z-coordinates use ST_Point. To create point geometries with only m-values or only z-coordinates use ST_PointM, ST_PointZ, or ST_MakePoint.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PointZM

Parameters
  • x (pyspark.sql.Column/int/float) – X-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (x coordinates) or a numeric value.

  • y (pyspark.sql.Column/int/float) – Y-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values (y-coordinates) or a numeric value.

  • z (pyspark.sql.Column/int/float) – Z-coordinate for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

  • m (pyspark.sql.Column/int/float) – M-value for the point. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned multipoint geometry, defaults to None.

Returns

Point column representing the point geometries.

Return type

pyspark.sql.Column

points

geoanalytics.sql.functions.points(geometry)

Returns an array column. For linestring and polygon geometries the function returns the vertices of the input geometry as an array of points. For multipoint and point geometries the function returns an array of all points in the input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Points

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Array column representing an array of point geometries.

Return type

pyspark.sql.Column

poly_from_binary

geoanalytics.sql.functions.poly_from_binary(wkb, sr=None)

Returns a polygon column. The input binary column must contain the well-known binary (WKB) representation of polygon geometries. You can optionally specify a spatial reference for the result polygon column. The sr parameter value must be a valid SRID or WKT string. If a polygon cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PolyFromBinary

Parameters
  • wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned polygon geometry, defaults to None.

Returns

Polygon column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

poly_from_esri_json

geoanalytics.sql.functions.poly_from_esri_json(json_str, sr=None)

Returns a polygon column. The input string column must contain the Esri JSON representation of polygon geometries. You can optionally specify a spatial reference for the result polygon column. The sr parameter value must be a valid SRID or WKT string. Any SRID defined in the input strings will not be used. If a polygon cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PolyFromEsriJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the Esri JSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned polygon geometry, defaults to None.

Returns

Polygon column from the Esri JSON representation.

Return type

pyspark.sql.Column

poly_from_geojson

geoanalytics.sql.functions.poly_from_geojson(json_str, sr=None)

Returns a polygon column. The input string column must contain the GeoJSON representation of polygon geometries. You can optionally specify a spatial reference for the result polygon column. The sr parameter value must be a valid SRID or WKT string. If a polygon cannot be created from the input string the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PolyFromGeoJSON

Parameters
  • json_str (pyspark.sql.Column) – StringType column with the GeoJSON representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned polygon geometry, defaults to None.

Returns

Polygon column from the GeoJSON representation.

Return type

pyspark.sql.Column

poly_from_shape

geoanalytics.sql.functions.poly_from_shape(shp, sr=None)

Returns a polygon column. The input binary column must contain the shapefile representation of polygon geometries. You can optionally specify a spatial reference for the result polygon column. The sr parameter value must be a valid SRID or WKT string. If a polygon cannot be created from the input binary the function will return null.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PolyFromShape

Parameters
  • shp (pyspark.sql.Column) – BinaryType column with the shapefile representation.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned polygon geometry, defaults to None.

Returns

Polygon column from the shapefile representation.

Return type

pyspark.sql.Column

poly_from_text

geoanalytics.sql.functions.poly_from_text(wkt, sr=None)

Returns a polygon column. The string column must contain the well-known text (WKT) representation of polygon geometries. You can optionally specify a spatial reference for the result polygon column. The sr parameter value must be a valid SRID or WKT string.

Refer to the GeoAnalytics guide for examples and usage notes: ST_PolyFromText

Parameters
  • wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of polygon geometries.

  • sr (int/str, optional) – Spatial reference (SRID or WKT) to set on the returned polygon geometry, defaults to None.

Returns

Polygon column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

polygon

geoanalytics.sql.functions.polygon(points)

Returns a polygon column. The input column must contain an array of point geometries.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Polygon

Parameters

points (pyspark.sql.Column) – Array of point geometries.

Returns

Polygon column representing the array of points.

Return type

pyspark.sql.Column

relate

geoanalytics.sql.functions.relate(geometry1, geometry2, relation)

Returns a boolean column where the result is True if the first geometry and the second geometry satisfy the spatial relationship defined by the specified DE-9IM string code; otherwise, it returns False. The string code contains nine characters that represent the nine spatial relations of the dimensionally extended 9-intersection model (DE-9IM). The character values indicate the dimensionality of the relationship: 0 for points, 1 for linestrings, 2 for polygons, and ‘F’ to indicate an empty set.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Relate

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

  • relation (pyspark.sql.Column/str) – The DE-9IM matrix value that will be used to compare the spatial relationship. Can be a StringType column or string value.

Returns

BooleanType column. True if the spatial relationship of geometry1 and geometry2 match the DE-9IM matrix value, False otherwise.

Return type

pyspark.sql.Column

rotate

geoanalytics.sql.functions.rotate(geometry, angle_in_radians, rotation_center=None)

Rotates a geometry counterclockwise by an angle specified in radians.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Rotate

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • angle_in_radians (pyspark.sql.Column/int/float) – Rotation angle in radians. Can be a LongType, DoubleType or StringType column or a numeric value.

  • rotation_center (pyspark.sql.Column, optional) – if specified, the geometry is rotated about the rotation_center, else around the origin (0,0).

Returns

Geometry column with the rotated geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

scale

geoanalytics.sql.functions.scale(geometry, x_scale_factor, y_scale_factor)

Scales the geometry to a new size by multiplying the coordinates with the corresponding factor parameters.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Scale

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • x_scale_factor (pyspark.sql.Column/int/float) – Numeric value that specifies the scale factor in the X direction. Can be a LongType, DoubleType or StringType column or a numeric value.

  • y_scale_factor (pyspark.sql.Column/int/float) – Numeric value that specifies the scale factor in the Y direction. Can be a LongType, DoubleType or StringType column or a numeric value.

Returns

Geometry column with the scaled geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

segmentize

geoanalytics.sql.functions.segmentize(linestring, max_segment_length=2)

Returns an array column. This function creates an array of linestrings from the input linestring by breaking the input linestring into segments that are shorter than or equal to the maximum length specified. The maximum segment length is in the same units as the input geometry.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Segmentize

Parameters
  • linestring (pyspark.sql.Column) – Linestring geometry column.

  • max_segment_length (pyspark.sql.Column/int/float) – Maximum length for any segment created. Can be a LongType, DoubleType or StringType column or an integer or float value.

Returns

Array column representing an array of linestring segments.

Return type

pyspark.sql.Column

segments

geoanalytics.sql.functions.segments(linestring, num_points=2, step_size=1)

Returns an array column. The function creates an array of linestrings by splitting the input linestring at a certain number of vertices using a moving window. By default the function will create segments with two points and the moving window will move one point at a time (step size of 1). You can optionally include a larger number of points in each linestring by specifying a numeric value greater than 2. You can also increase the step size by setting it to a value greater than 1. Setting the step size to one less than the number of points will always result in segments that touch but do not overlap.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Segments

Parameters
  • linestring (pyspark.sql.Column) – Linestring geometry column.

  • num_points (pyspark.sql.Column/int, optional) – Numeric value representing the number of points in each segment, defaults to 2. Can be a LongType or StringType column or an integer value.

  • step_size (pyspark.sql.Column, optional) – Numeric value representing the number of points between the start of each new segment, defaults to 1. Can be a LongType or StringType column or an integer value.

Returns

Array column representing an array of linestring segments.

Return type

pyspark.sql.Column

shear

geoanalytics.sql.functions.shear(geometry, proportion_x, proportion_y)

Returns a geometry column that generalizes the input linestring or polygon geometry using the Douglas-Peucker algorithm with the specified tolerance. The result is the input geometry generalized to include only a subset of the original geometry’s vertices. Point and multipoint geometry types are not supported as input.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Shear

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • proportion_x (pyspark.sql.Column/int/float) – Numeric value that specifies the proportion of shearing in the X direction. Can be a LongType, DoubleType or StringType column or a numeric value.

  • proportion_y (pyspark.sql.Column/int/float) – Numeric value that specifies the proportion of shearing in the Y direction. Can be a LongType, DoubleType or StringType column or a numeric value.

Returns

Geometry column with the sheared geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

shortest_line

geoanalytics.sql.functions.shortest_line(geometry1, geometry2)

Returns a linestring column representing the shortest line that touches two geometries, using planar distance calculation. This function returns only one shortest line if there are more than one. If the two input geometries intersect, an empty line geometry is returned. If the two geometry columns are in different spatial references, the function automatically transforms the second geometry into the spatial reference of the first. To create a shortest line using geodesic distance calculation, use ST_GeodesicShortestLine.

Refer to the GeoAnalytics guide for examples and usage notes: ST_ShortestLine

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the shortest line.

Return type

pyspark.sql.Column

simplify

geoanalytics.sql.functions.simplify(geometry)

Returns a geometry column containing the simplified geometries. This function simplifies the input geometry according to the OpenGIS Simple Features Implementation Specification for SQL 1.2.1 (06-103r4).

Refer to the GeoAnalytics guide for examples and usage notes: ST_Simplify

Parameters

geometry (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the simplified geometry. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

split

geoanalytics.sql.functions.split(geometry, splitter)

Returns an array column from an input linestring or polygon column. This function splits the geometry with the splitter linestring and returns the resulting parts as an array of geometries. If the input geometry is a linestring, the output will be an array of linestrings. If the input geometry type is a polygon, the output will be an array of polygons. If a linestring is split by an equal linestring, an empty linestring along with the input linestring and the splitter linestring are returned.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Split

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • splitter (pyspark.sql.Column) – Linestring geometry column.

Returns

Array column representing an array of geometries resulting from the split. The returned geometry type will be the same as the geometry column type.

Return type

pyspark.sql.Column

square_bin

geoanalytics.sql.functions.square_bin(geometry, bin_size)

Returns a bin column containing a single square bin for each record in the input column. The specified bin size determines the height of each bin and is in the same units as the input geometry. The centroid of the input geometry is guaranteed to intersect with the bin returned but is not necessarily coincident with the bin center. Use ST_BinGeometry to obtain the geometry of each result bin.

This function can also be called with a long column representing the ID of the bin (see ST_BinId). The bin ID will be cast to a bin column.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SquareBin

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • bin_size (int/float) – Numeric value representing the size of the side of the square bin.

Returns

Spatial bin (bin2d) column representing a single square bin for each geometry.

Return type

pyspark.sql.Column

square_bins

geoanalytics.sql.functions.square_bins(geometry, bin_size, padding=0.0)

Returns an array column containing square bins that cover the spatial extent of each record in the input column. The specified bin size determines the height of each bin and is in the same units as the input geometry. You can optionally specify a numeric value for padding, which conceptually applies a buffer of the specified distance to the input geometry before creating the square bins.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SquareBins

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • bin_size (int/float) – Numerical value representing the size of the side of the square bin.

  • padding (int/float/str, optional) – Numerical buffer value applied to the geometry before finding the intersecting bins, defaults to 0.0.

Returns

Array column representing an array of spatial bin (bin2d) square bins.

Return type

pyspark.sql.Column

sr_text

geoanalytics.sql.functions.sr_text(geometry, wkt=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a geometry column and returns the spatial reference (WKT) of the column as a string column. If the spatial reference of the input geometry column has not been set, the function returns an empty string.

Setter: Takes a geometry column and a string value and returns the input geometry column with its spatial reference WKT set to the string value. This does not affect the geometry data in the column. To transform your geometry data from one spatial reference to another, use ST_Transform.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SRText

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • wkt (str, optional) – Spatial reference (WKT) to set on the geometry, defaults to None.

Returns

  • Getter: StringType column representing the spatial reference (WKT) for the geometry.

  • Setter: Geometry column representing the geometry with the updated spatial reference. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

srid

geoanalytics.sql.functions.srid(geometry, srid=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a geometry column and returns the spatial reference (SRID) of the column as an integer column. If the SRID of the input geometry column has not been set, the function returns 0.

Setter: Takes a geometry column and a numeric value and returns the input geometry column with its SRID set to the numeric value. This does not affect the geometry data in the column. To transform your geometry data from one spatial reference to another, use ST_Transform.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SRID

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • srid (int, optional) – Spatial reference (SRID) to set on the geometry, defaults to None.

Returns

  • Getter: IntegerType column representing the spatial reference (SRID) for the geometry.

  • Setter: Geometry column representing the geometry with the updated spatial reference. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

start_point

geoanalytics.sql.functions.start_point(linestring)

Returns a point column representing the first point of the input linestring.

Refer to the GeoAnalytics guide for examples and usage notes: ST_StartPoint

Parameters

linestring (pyspark.sql.Column) – Linestring geometry column.

Returns

Point column representing the starting point.

Return type

pyspark.sql.Column

sym_difference

geoanalytics.sql.functions.sym_difference(geometry1, geometry2)

Returns a geometry column containing the geometries that represent the portions of the input geometries that do not intersect. If one of the input geometry types is geometry, the output type will be the same. For all other cases the result geometry type will be the same as the input geometry type with the highest dimension.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SymDifference

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the portions of geometry1 and geometry2 that do not intersect. The returned geometry column type will be the highest dimension of the two input geometries. If one or both of the input geometries is a generic geometry type, then a generic geometry column type will be returned. For example linestring and polygon input will return a polygon geometry type.

Return type

pyspark.sql.Column

symmetric_diff

geoanalytics.sql.functions.symmetric_diff(geometry1, geometry2)

Esri alias for OGC ST_SymDifference.

Refer to the GeoAnalytics guide for examples and usage notes: ST_SymDifference

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

Geometry column representing the portions of geometry1 and geometry2 that do not intersect. The returned geometry column type will be the highest dimension of the two input geometries. If one or both of the input geometries is a generic geometry type, then a generic geometry column type will be returned. For example linestring and polygon input will return a polygon geometry type.

Return type

pyspark.sql.Column

touches

geoanalytics.sql.functions.touches(geometry1, geometry2)

Returns a boolean column where the result is True if the first geometry and the second geometry spatially touch on their boundaries (i.e., their intersection is a single point); otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Touches

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 and geometry2 spatially touch on their boundaries, False otherwise.

Return type

pyspark.sql.Column

transform

geoanalytics.sql.functions.transform(geometry, sr, *, extent=None, datum_transform=None)

Returns a geometry column. The input geometry column must have a spatial reference set. The sr parameter value must be a valid SRID or WKT string. The function returns the input geometries transformed into the specified spatial reference. It will also set the spatial reference of the result column. To learn more about what it means to transform your geometry data, see Coordinate systems and transformations. To set the spatial reference of a geometry column without transforming the geometries, use ST_SRID or ST_SRText.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Transform

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • sr (int/str) – The spatial reference (SRID or WKT) that the geometry will be projected into.

  • extent (BoundingBox, optional) – Extent of the area of analysis to use when determining the best transformation to use.

  • datum_transform (str, optional) – Transformation path to use when transforming between geographic spatial references. This parameter overrides the session level transform settings as well as the extent param.

Returns

Geometry column representing the projected geometry. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

translate

geoanalytics.sql.functions.translate(geometry, x_offset, y_offset)

Translates a geometry by given offsets.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Translate

Parameters
  • geometry (pyspark.sql.Column) – Geometry column.

  • x_offset (pyspark.sql.Column/int/float) – Numeric value that specifies the offset in the X direction. Can be a LongType, DoubleType or StringType column or a numeric value.

  • y_offset (pyspark.sql.Column/int/float) – Numeric value that specifies the offset in the Y direction. Can be a LongType, DoubleType or StringType column or a numeric value.

Returns

Geometry column with the translated geometries. The returned geometry column type will be the same as the input geometry column type.

Return type

pyspark.sql.Column

union

geoanalytics.sql.functions.union(*geometries)

Returns a geometry column containing the geometries that represent the spatial union of the geometries in each row of the input columns. The geometry types of the input columns must be the same. To find the union of all geometries in a group or column use ST_Aggr_Union.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Union

Parameters

geometries (pyspark.sql.Column) – Multiple geometry columns.

Returns

Geometry column representing the spatial union of the geometries. The returned geometry column type will be the same type as the input geometries except in the case of point input which will return multipoint. If one or both of the input geometries is a generic geometry type, then a generic geometry column type will be returned.

Return type

pyspark.sql.Column

within

geoanalytics.sql.functions.within(geometry1, geometry2)

Returns a boolean column where the result is True if the first geometry is completely inside the second geometry; otherwise, it returns False.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Within

Parameters
  • geometry1 (pyspark.sql.Column) – Geometry column.

  • geometry2 (pyspark.sql.Column) – Geometry column.

Returns

BooleanType column. True if geometry1 is within geometry2, False otherwise.

Return type

pyspark.sql.Column

wkb_to_sql

geoanalytics.sql.functions.wkb_to_sql(wkb)

OGC alias for ST_GeomFromBinary without SRID.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromBinary

Parameters

wkb (pyspark.sql.Column) – BinaryType column with the Well-Known Binary (WKB) representation.

Returns

Generic geometry column from the Well-Known Binary (WKB) representation.

Return type

pyspark.sql.Column

wkt_to_sql

geoanalytics.sql.functions.wkt_to_sql(wkt)

OGC alias for ST_GeomFromText without SRID.

Refer to the GeoAnalytics guide for examples and usage notes: ST_GeomFromText

Parameters

wkt (pyspark.sql.Column) – StringType column with the well-known text (WKT) representation of geometries.

Returns

Generic geometry column from the well-known text (WKT) representation.

Return type

pyspark.sql.Column

x

geoanalytics.sql.functions.x(point, new_value=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a point column and returns a double column containing the x-coordinate of the input points.

Setter: Takes a point column and a numeric value or column and returns a point column containing the input points with the x-coordinates set to the numeric value.

Refer to the GeoAnalytics guide for examples and usage notes: ST_X

Parameters
  • point (pyspark.sql.Column) – Point geometry column.

  • new_value (pyspark.sql.Column/int/float, optional) – X-coordinate to set. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

Returns

  • Getter: DoubleType column representing the x-coordinate.

  • Setter: Point column representing the updated x-coordinate.

Return type

pyspark.sql.Column

y

geoanalytics.sql.functions.y(point, new_value=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a point column and returns a double column containing the y-coordinate of the input points.

Setter: Takes a point column and a numeric value or column and returns a point column containing the input points with the y-coordinates set to the numeric value.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Y

Parameters
  • point (pyspark.sql.Column) – Point geometry column.

  • new_value (pyspark.sql.Column/int/float, optional) – Y-coordinate to set. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

Returns

  • Getter: DoubleType column representing the y-coordinate.

  • Setter: Point column representing the updated y-coordinate.

Return type

pyspark.sql.Column

z

geoanalytics.sql.functions.z(point, new_value=None)

Can work as a getter or a setter, depending on the inputs.

Getter: Takes a point column and returns a double column containing the z-coordinates of the input points. If a point does not have a z-coordinate the function returns NaN.

Setter: Takes a point column and a numeric value or column and returns a point column containing the input points with the z-coordinates set to the numeric value.

Refer to the GeoAnalytics guide for examples and usage notes: ST_Z

Parameters
  • point (pyspark.sql.Column) – Point geometry column.

  • new_value (pyspark.sql.Column/int/float, optional) – Z-coordinate to set. Can be a LongType, DoubleType or StringType column with numeric values or a numeric value, defaults to None.

Returns

  • Getter: DoubleType column representing the z-coordinate.

  • Setter: Point column representing the updated z-coordinate.

Return type

pyspark.sql.Column