pyspark.pandas.read_sql_query#

pyspark.pandas.read_sql_query(sql, con, index_col=None, **options)[source]#

Read SQL query into a DataFrame.

Returns a DataFrame corresponding to the result set of the query string. Optionally provide an index_col parameter to use one of the columns as the index, otherwise default index will be used.

Note

Some database might hit the issue of Spark: SPARK-27596

Parameters
sqlstring SQL query

SQL query to be executed.

constr

A JDBC URI could be provided as str.

Note

The URI must be JDBC URI instead of Python’s database URI.

index_colstring or list of strings, optional, default: None

Column(s) to set as index(MultiIndex).

optionsdict

All other options passed directly into Spark’s JDBC data source.

Returns
DataFrame

See also

read_sql_table

Read SQL database table into a DataFrame.

read_sql

Examples

>>> ps.read_sql_query('SELECT * FROM table_name', 'jdbc:postgresql:db_name')