pyspark.sql.Column.alias¶
-
Column.
alias
(*alias, **kwargs)[source]¶ Returns this column aliased with a new name or names (in the case of expressions that return more than one column, such as explode).
New in version 1.3.0.
- Parameters
- aliasstr
desired column names (collects all positional arguments passed)
- Other Parameters
- metadata: dict
a dict of information to be stored in
metadata
attribute of the correspondingStructField
(optional, keyword only argument)Changed in version 2.2.0: Added optional
metadata
argument.
Examples
>>> df.select(df.age.alias("age2")).collect() [Row(age2=2), Row(age2=5)] >>> df.select(df.age.alias("age3", metadata={'max': 99})).schema['age3'].metadata['max'] 99