Skip to content

Commit

Permalink
chore(deps): bump ruff to 0.3.0 (#8503)
Browse files Browse the repository at this point in the history
  • Loading branch information
cpcloud authored Mar 1, 2024
1 parent ee821b1 commit ef39aab
Show file tree
Hide file tree
Showing 56 changed files with 183 additions and 319 deletions.
3 changes: 1 addition & 2 deletions docs/posts/pydata-performance/step1.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,7 @@


@udf.scalar.builtin
def flatten(x: list[list[str]]) -> list[str]:
...
def flatten(x: list[list[str]]) -> list[str]: ...


expr = (
Expand Down
3 changes: 1 addition & 2 deletions docs/posts/pydata-performance/step2.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,7 @@


@udf.scalar.builtin
def flatten(x: list[list[str]]) -> list[str]:
...
def flatten(x: list[list[str]]) -> list[str]: ...


expr = (
Expand Down
1 change: 1 addition & 0 deletions docs/posts/torch/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@
Adapted from https://gist.github.com/pdet/e8d38734232c08e6c15aba79b4eb8368#file-taxi_prediction_example-py.
"""

from __future__ import annotations

import pyarrow as pa
Expand Down
1 change: 1 addition & 0 deletions ibis/__init__.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
"""Initialize Ibis module."""

from __future__ import annotations

__version__ = "8.0.0"
Expand Down
10 changes: 3 additions & 7 deletions ibis/backends/bigquery/converter.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,10 +11,6 @@ def convert_GeoSpatial(cls, s, dtype, pandas_type):

return gpd.GeoSeries(shp.from_wkt(s))

convert_Point = (
convert_LineString
) = (
convert_Polygon
) = (
convert_MultiLineString
) = convert_MultiPoint = convert_MultiPolygon = convert_GeoSpatial
convert_Point = convert_LineString = convert_Polygon = convert_MultiLineString = (
convert_MultiPoint
) = convert_MultiPolygon = convert_GeoSpatial
6 changes: 2 additions & 4 deletions ibis/backends/bigquery/tests/system/udf/test_udf_execute.py
Original file line number Diff line number Diff line change
Expand Up @@ -148,8 +148,7 @@ def test_builtin_scalar(con, value, expected):
from ibis import udf

@udf.scalar.builtin
def bit_count(x: bytes) -> int:
...
def bit_count(x: bytes) -> int: ...

expr = bit_count(value)
result = con.execute(expr)
Expand All @@ -168,8 +167,7 @@ def test_builtin_agg(con, where, expected):
from ibis import udf

@udf.agg.builtin(name="array_concat_agg")
def concat_agg(x, where: bool = True) -> dt.Array[str]:
...
def concat_agg(x, where: bool = True) -> dt.Array[str]: ...

t = ibis.memtable({"a": [list("abc"), list("def")]})
expr = concat_agg(t.a, **where)
Expand Down
3 changes: 1 addition & 2 deletions ibis/backends/bigquery/tests/unit/udf/test_builtin.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,7 @@


@ibis.udf.scalar.builtin
def farm_fingerprint(value: bytes) -> int:
...
def farm_fingerprint(value: bytes) -> int: ...


@ibis.udf.scalar.builtin(schema="fn", database="bqutil")
Expand Down
15 changes: 5 additions & 10 deletions ibis/backends/clickhouse/tests/test_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -262,18 +262,15 @@ def test_truncate_table(con, engine, temp_table):


@udf.scalar.builtin(name="arrayJaccardIndex")
def array_jaccard_index(a: dt.Array[dt.int64], b: dt.Array[dt.int64]) -> float:
...
def array_jaccard_index(a: dt.Array[dt.int64], b: dt.Array[dt.int64]) -> float: ...


@udf.scalar.builtin(name="arrayJaccardIndex")
def array_jaccard_index_no_input_types(a, b) -> float:
...
def array_jaccard_index_no_input_types(a, b) -> float: ...


@udf.scalar.builtin
def arrayJaccardIndex(a: dt.Array[dt.int64], b: dt.Array[dt.int64]) -> float:
...
def arrayJaccardIndex(a: dt.Array[dt.int64], b: dt.Array[dt.int64]) -> float: ...


@pytest.mark.parametrize(
Expand All @@ -292,13 +289,11 @@ def test_builtin_scalar_udf(con, func):


@udf.agg.builtin
def entropy(a) -> float:
...
def entropy(a) -> float: ...


@udf.agg.builtin(name="sumKahan")
def sum_kahan(a: float) -> float:
...
def sum_kahan(a: float) -> float: ...


@pytest.mark.parametrize("func", [entropy, sum_kahan])
Expand Down
6 changes: 2 additions & 4 deletions ibis/backends/clickhouse/tests/test_functions.py
Original file line number Diff line number Diff line change
Expand Up @@ -473,8 +473,7 @@ def test_hash(alltypes, snapshot):

def test_udf_in_array_map(alltypes):
@udf.scalar.builtin(name="plus")
def my_add(a: int, b: int) -> int:
...
def my_add(a: int, b: int) -> int: ...

n = 5
expr = (
Expand All @@ -489,8 +488,7 @@ def my_add(a: int, b: int) -> int:

def test_udf_in_array_filter(alltypes):
@udf.scalar.builtin(name="equals")
def my_eq(a: int, b: int) -> bool:
...
def my_eq(a: int, b: int) -> bool: ...

expr = alltypes.int_col.collect().filter(lambda x: my_eq(x, 1))
result = expr.execute()
Expand Down
6 changes: 3 additions & 3 deletions ibis/backends/dask/convert.py
Original file line number Diff line number Diff line change
Expand Up @@ -41,9 +41,9 @@ def convert_Integer(cls, s, dtype, pandas_type):

convert_SignedInteger = convert_UnsignedInteger = convert_Integer
convert_Int64 = convert_Int32 = convert_Int16 = convert_Int8 = convert_SignedInteger
convert_UInt64 = (
convert_UInt32
) = convert_UInt16 = convert_UInt8 = convert_UnsignedInteger
convert_UInt64 = convert_UInt32 = convert_UInt16 = convert_UInt8 = (
convert_UnsignedInteger
)

@classmethod
def convert_Floating(cls, s, dtype, pandas_type):
Expand Down
6 changes: 3 additions & 3 deletions ibis/backends/datafusion/compiler.py
Original file line number Diff line number Diff line change
Expand Up @@ -248,9 +248,9 @@ def visit_ExtractYearMonthQuarterDay(self, op, *, arg):
part = type(op).__name__[skip:].lower()
return self.f.date_part(part, arg)

visit_ExtractYear = (
visit_ExtractMonth
) = visit_ExtractQuarter = visit_ExtractDay = visit_ExtractYearMonthQuarterDay
visit_ExtractYear = visit_ExtractMonth = visit_ExtractQuarter = visit_ExtractDay = (
visit_ExtractYearMonthQuarterDay
)

def visit_ExtractDayOfYear(self, op, *, arg):
return self.f.date_part("doy", arg)
Expand Down
30 changes: 10 additions & 20 deletions ibis/backends/duckdb/tests/test_udf.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,33 +7,27 @@


@udf.scalar.builtin
def hamming(a: str, b: str) -> int:
...
def hamming(a: str, b: str) -> int: ...


@udf.scalar.builtin
def jaccard(a: str, b: str) -> float:
...
def jaccard(a: str, b: str) -> float: ...


@udf.scalar.builtin
def jaro_similarity(a: str, b: str) -> float:
...
def jaro_similarity(a: str, b: str) -> float: ...


@udf.scalar.builtin
def jaro_winkler_similarity(a: str, b: str) -> float:
...
def jaro_winkler_similarity(a: str, b: str) -> float: ...


@udf.scalar.builtin
def damerau_levenshtein(a: str, b: str) -> int:
...
def damerau_levenshtein(a: str, b: str) -> int: ...


@udf.scalar.builtin
def mismatches(a: str, b: str) -> int:
...
def mismatches(a: str, b: str) -> int: ...


@pytest.mark.parametrize(
Expand All @@ -57,26 +51,22 @@ def test_builtin_scalar(con, func):

def test_builtin_scalar_noargs(con):
@udf.scalar.builtin
def version() -> str:
...
def version() -> str: ...

expr = version()
assert con.execute(expr) == f"v{con.version}"


@udf.agg.builtin
def product(x, where: bool = True) -> float:
...
def product(x, where: bool = True) -> float: ...


@udf.agg.builtin
def fsum(x, where: bool = True) -> float:
...
def fsum(x, where: bool = True) -> float: ...


@udf.agg.builtin
def favg(x: float, where: bool = True) -> float:
...
def favg(x: float, where: bool = True) -> float: ...


@pytest.mark.parametrize("func", [product, fsum, favg])
Expand Down
12 changes: 4 additions & 8 deletions ibis/backends/flink/compiler.py
Original file line number Diff line number Diff line change
Expand Up @@ -496,9 +496,9 @@ def visit_TemporalTruncate(self, op, *, arg, unit):
unit_sql = unit_var.sql(self.dialect)
return self.f.floor(self.v[f"{arg_sql} TO {unit_sql}"])

visit_TimestampTruncate = (
visit_DateTruncate
) = visit_TimeTruncate = visit_TemporalTruncate
visit_TimestampTruncate = visit_DateTruncate = visit_TimeTruncate = (
visit_TemporalTruncate
)

def visit_StringContains(self, op, *, haystack, needle):
return self.f.instr(haystack, needle) > 0
Expand All @@ -525,11 +525,7 @@ def visit_EndsWith(self, op, *, arg, end):
def visit_ExtractUrlField(self, op, *, arg):
return self.f.parse_url(arg, type(op).__name__[len("Extract") :].upper())

visit_ExtractAuthority = (
visit_ExtractHost
) = (
visit_ExtractUserInfo
) = (
visit_ExtractAuthority = visit_ExtractHost = visit_ExtractUserInfo = (
visit_ExtractProtocol
) = visit_ExtractFile = visit_ExtractPath = visit_ExtractUrlField

Expand Down
3 changes: 1 addition & 2 deletions ibis/backends/flink/tests/test_udf.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,7 @@

def test_builtin_scalar_udf(con):
@udf.scalar.builtin
def parse_url(string1: str, string2: str) -> str:
...
def parse_url(string1: str, string2: str) -> str: ...

expr = parse_url("http://facebook.com/path1/p.php?k1=v1&k2=v2#Ref1", "HOST")
result = con.execute(expr)
Expand Down
15 changes: 5 additions & 10 deletions ibis/backends/flink/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -104,18 +104,15 @@ def __init__(self, value: int, unit: str) -> None:
@classmethod
@property
@abstractmethod
def units(self):
...
def units(self): ...

@classmethod
@property
@abstractmethod
def factors(self):
...
def factors(self): ...

@abstractmethod
def _convert_to_highest_resolution(self):
...
def _convert_to_highest_resolution(self): ...

def _convert_to_combined_units(self) -> dict:
converted_total = self._convert_to_highest_resolution()
Expand All @@ -132,12 +129,10 @@ def _convert_to_combined_units(self) -> dict:
return interval_segments

@abstractmethod
def _calculate_precisions(self) -> dict:
...
def _calculate_precisions(self) -> dict: ...

@abstractmethod
def format_as_string(self, interval_segments: dict, precisions: dict) -> str:
...
def format_as_string(self, interval_segments: dict, precisions: dict) -> str: ...


class YearsToMonthsInterval(FlinkIntervalType):
Expand Down
6 changes: 1 addition & 5 deletions ibis/backends/impala/compiler.py
Original file line number Diff line number Diff line change
Expand Up @@ -242,11 +242,7 @@ def visit_FindInSet(self, op, *, needle, values):
def visit_ExtractUrlField(self, op, *, arg):
return self.f.parse_url(arg, type(op).__name__[len("Extract") :].upper())

visit_ExtractAuthority = (
visit_ExtractHost
) = (
visit_ExtractUserInfo
) = (
visit_ExtractAuthority = visit_ExtractHost = visit_ExtractUserInfo = (
visit_ExtractProtocol
) = visit_ExtractFile = visit_ExtractPath = visit_ExtractUrlField

Expand Down
3 changes: 1 addition & 2 deletions ibis/backends/impala/udf.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,8 +44,7 @@ def __call__(self, *args):
return self._klass(*args)

def _make_fn(self):
def fn(*args, **kwargs):
...
def fn(*args, **kwargs): ...

fn.__name__ = self.name
fn.__signature__ = inspect.Signature(
Expand Down
14 changes: 2 additions & 12 deletions ibis/backends/mssql/compiler.py
Original file line number Diff line number Diff line change
Expand Up @@ -293,19 +293,9 @@ def visit_ExtractEpochSeconds(self, op, *, arg):
def visit_ExtractTemporalComponent(self, op, *, arg):
return self.f.datepart(self.v[type(op).__name__[len("Extract") :].lower()], arg)

visit_ExtractYear = (
visit_ExtractMonth
) = (
visit_ExtractDay
) = (
visit_ExtractYear = visit_ExtractMonth = visit_ExtractDay = (
visit_ExtractDayOfYear
) = (
visit_ExtractHour
) = (
visit_ExtractMinute
) = (
visit_ExtractSecond
) = (
) = visit_ExtractHour = visit_ExtractMinute = visit_ExtractSecond = (
visit_ExtractMillisecond
) = visit_ExtractMicrosecond = visit_ExtractTemporalComponent

Expand Down
6 changes: 3 additions & 3 deletions ibis/backends/pandas/convert.py
Original file line number Diff line number Diff line change
Expand Up @@ -40,9 +40,9 @@ def convert_Integer(cls, s, dtype, pandas_type):

convert_SignedInteger = convert_UnsignedInteger = convert_Integer
convert_Int64 = convert_Int32 = convert_Int16 = convert_Int8 = convert_SignedInteger
convert_UInt64 = (
convert_UInt32
) = convert_UInt16 = convert_UInt8 = convert_UnsignedInteger
convert_UInt64 = convert_UInt32 = convert_UInt16 = convert_UInt8 = (
convert_UnsignedInteger
)

@classmethod
def convert_Floating(cls, s, dtype, pandas_type):
Expand Down
6 changes: 2 additions & 4 deletions ibis/backends/polars/tests/test_udf.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,8 +56,7 @@ def test_multiple_argument_udf(alltypes):
)
def test_builtin_scalar_udf(con, value, expected):
@udf.scalar.builtin
def cbrt(a: float) -> float:
...
def cbrt(a: float) -> float: ...

expr = cbrt(value)
result = con.execute(expr)
Expand Down Expand Up @@ -109,8 +108,7 @@ def test_multiple_argument_scalar_udf(alltypes, func):

def test_builtin_agg_udf(con):
@udf.agg.builtin
def approx_n_unique(a, where: bool = True) -> int:
...
def approx_n_unique(a, where: bool = True) -> int: ...

ft = con.tables.functional_alltypes
expr = approx_n_unique(ft.string_col)
Expand Down
3 changes: 1 addition & 2 deletions ibis/backends/postgres/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -381,8 +381,7 @@ def split_name_type(arg: str) -> tuple[str, dt.DataType]:
signature = list(map(split_name_type, signature))

# dummy callable
def fake_func(*args, **kwargs):
...
def fake_func(*args, **kwargs): ...

fake_func.__name__ = name
fake_func.__signature__ = inspect.Signature(
Expand Down
Loading

0 comments on commit ef39aab

Please sign in to comment.