pyspark.pandas.extensions.register_series_accessor¶
-
pyspark.pandas.extensions.
register_series_accessor
(name: str) → Callable[[Type[T]], Type[T]]¶ Register a custom accessor with a Series object
- Parameters
- namestr
name used when calling the accessor after its registered
- Returns
- callable
A class decorator.
See also
register_dataframe_accessor
Register a custom accessor on DataFrame objects
register_index_accessor
Register a custom accessor on Index objects
Notes
When accessed, your accessor will be initialiazed with the pandas-on-Spark object the user is interacting with. The code signature must be:
def __init__(self, pandas_on_spark_obj): # constructor logic ...
In the pandas API, if data passed to your accessor has an incorrect dtype, it’s recommended to raise an
AttributeError
for consistency purposes. In pandas-on-Spark,ValueError
is more frequently used to annotate when a value’s datatype is unexpected for a given method/function.Ultimately, you can structure this however you like, but pandas-on-Spark would likely do something like this:
>>> ps.Series(['a', 'b']).dt ... Traceback (most recent call last): ... ValueError: Cannot call DatetimeMethods on type StringType()
Examples
In your library code:
from pyspark.pandas.extensions import register_series_accessor @register_series_accessor("geo") class GeoAccessor: def __init__(self, pandas_on_spark_obj): self._obj = pandas_on_spark_obj @property def is_valid(self): # boolean check to see if series contains valid geometry return True
Then, in an ipython session:
>>> ## Import if the accessor is in the other file. >>> # from my_ext_lib import GeoAccessor >>> psdf = ps.DataFrame({"longitude": np.linspace(0,10), ... "latitude": np.linspace(0, 20)}) >>> psdf.longitude.geo.is_valid True