|
50 | 50 | from pyiceberg.schema import Schema |
51 | 51 | from pyiceberg.table import TableProperties |
52 | 52 | from pyiceberg.table.sorting import SortDirection, SortField, SortOrder |
53 | | -from pyiceberg.transforms import BucketTransform, DayTransform, HourTransform, IdentityTransform, Transform |
| 53 | +from pyiceberg.transforms import DayTransform, HourTransform, IdentityTransform, Transform |
54 | 54 | from pyiceberg.types import ( |
55 | 55 | DateType, |
56 | 56 | DecimalType, |
@@ -1274,7 +1274,7 @@ def test_table_write_schema_with_valid_upcast( |
1274 | 1274 | pa.field("list", pa.list_(pa.int64()), nullable=False), |
1275 | 1275 | pa.field("map", pa.map_(pa.string(), pa.int64()), nullable=False), |
1276 | 1276 | pa.field("double", pa.float64(), nullable=True), # can support upcasting float to double |
1277 | | - pa.field("uuid", pa.binary(length=16), nullable=True), # can UUID is read as fixed length binary of length 16 |
| 1277 | + pa.field("uuid", pa.uuid(), nullable=True), |
1278 | 1278 | ) |
1279 | 1279 | ) |
1280 | 1280 | ) |
@@ -1847,7 +1847,15 @@ def test_read_write_decimals(session_catalog: Catalog) -> None: |
1847 | 1847 |
|
1848 | 1848 |
|
1849 | 1849 | @pytest.mark.integration |
1850 | | -@pytest.mark.parametrize("transform", [IdentityTransform(), BucketTransform(32)]) |
| 1850 | +@pytest.mark.parametrize( |
| 1851 | + "transform", |
| 1852 | + [ |
| 1853 | + IdentityTransform(), |
| 1854 | + # Bucket is disabled because of an issue in Iceberg Java: |
| 1855 | + # https://github.com/apache/iceberg/pull/13324 |
| 1856 | + # BucketTransform(32) |
| 1857 | + ], |
| 1858 | +) |
1851 | 1859 | def test_uuid_partitioning(session_catalog: Catalog, spark: SparkSession, transform: Transform) -> None: # type: ignore |
1852 | 1860 | identifier = f"default.test_uuid_partitioning_{str(transform).replace('[32]', '')}" |
1853 | 1861 |
|
|
0 commit comments