pyspark.sql.functions.log¶
-
pyspark.sql.functions.
log
(arg1: Union[ColumnOrName, float], arg2: Optional[ColumnOrName] = None) → pyspark.sql.column.Column¶ Returns the first argument-based logarithm of the second argument.
If there is only one argument, then this takes the natural logarithm of the argument.
Examples
>>> df.select(log(10.0, df.age).alias('ten')).rdd.map(lambda l: str(l.ten)[:7]).collect() ['0.30102', '0.69897']
>>> df.select(log(df.age).alias('e')).rdd.map(lambda l: str(l.e)[:7]).collect() ['0.69314', '1.60943']