You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: python/docs/source/user_guide/sql/python_data_source.rst
+86-25Lines changed: 86 additions & 25 deletions
Original file line number
Diff line number
Diff line change
@@ -535,7 +535,7 @@ Filter Pushdown in Python Data Sources
535
535
536
536
Filter pushdown is an optimization technique that allows data sources to handle filters natively, reducing the amount of data that needs to be transferred and processed by Spark.
537
537
538
-
The filter pushdown API is introduced in Spark 4.1, enabling DataSourceReader to selectively push down filters from the query to the source.
538
+
The filter pushdown API enables DataSourceReader to selectively push down filters from the query to the source.
539
539
540
540
You must turn on the configuration ``spark.sql.python.filterPushdown.enabled`` to enable filter pushdown.
541
541
@@ -554,32 +554,93 @@ To enable filter pushdown in your Python Data Source, implement the ``pushFilter
554
554
555
555
.. code-block:: python
556
556
557
-
from pyspark.sql.datasource import EqualTo, Filter, GreaterThan, LessThan
0 commit comments