mirror of
https://github.com/lancedb/lancedb.git
synced 2026-05-05 14:10:41 +00:00
Compare commits
3 Commits
python-v0.
...
jack/memta
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
01c6b9dcb8 | ||
|
|
33a13f0738 | ||
|
|
cabc75f167 |
@@ -1,5 +1,5 @@
|
|||||||
[tool.bumpversion]
|
[tool.bumpversion]
|
||||||
current_version = "0.27.0"
|
current_version = "0.27.0-beta.3"
|
||||||
parse = """(?x)
|
parse = """(?x)
|
||||||
(?P<major>0|[1-9]\\d*)\\.
|
(?P<major>0|[1-9]\\d*)\\.
|
||||||
(?P<minor>0|[1-9]\\d*)\\.
|
(?P<minor>0|[1-9]\\d*)\\.
|
||||||
|
|||||||
6
.github/workflows/dev.yml
vendored
6
.github/workflows/dev.yml
vendored
@@ -15,7 +15,7 @@ jobs:
|
|||||||
name: Label PR
|
name: Label PR
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: srvaroa/labeler@v1
|
- uses: srvaroa/labeler@master
|
||||||
env:
|
env:
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
commitlint:
|
commitlint:
|
||||||
@@ -24,7 +24,7 @@ jobs:
|
|||||||
name: Verify PR title / description conforms to semantic-release
|
name: Verify PR title / description conforms to semantic-release
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: "18"
|
node-version: "18"
|
||||||
# These rules are disabled because Github will always ensure there
|
# These rules are disabled because Github will always ensure there
|
||||||
@@ -47,7 +47,7 @@ jobs:
|
|||||||
|
|
||||||
${{ github.event.pull_request.body }}
|
${{ github.event.pull_request.body }}
|
||||||
- if: failure()
|
- if: failure()
|
||||||
uses: actions/github-script@v7
|
uses: actions/github-script@v6
|
||||||
with:
|
with:
|
||||||
script: |
|
script: |
|
||||||
const message = `**ACTION NEEDED**
|
const message = `**ACTION NEEDED**
|
||||||
|
|||||||
4
.github/workflows/docs.yml
vendored
4
.github/workflows/docs.yml
vendored
@@ -53,7 +53,7 @@ jobs:
|
|||||||
python -m pip install --extra-index-url https://pypi.fury.io/lance-format/ --extra-index-url https://pypi.fury.io/lancedb/ -e .
|
python -m pip install --extra-index-url https://pypi.fury.io/lance-format/ --extra-index-url https://pypi.fury.io/lancedb/ -e .
|
||||||
python -m pip install --extra-index-url https://pypi.fury.io/lance-format/ --extra-index-url https://pypi.fury.io/lancedb/ -r ../docs/requirements.txt
|
python -m pip install --extra-index-url https://pypi.fury.io/lance-format/ --extra-index-url https://pypi.fury.io/lancedb/ -r ../docs/requirements.txt
|
||||||
- name: Set up node
|
- name: Set up node
|
||||||
uses: actions/setup-node@v4
|
uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 20
|
node-version: 20
|
||||||
cache: 'npm'
|
cache: 'npm'
|
||||||
@@ -68,7 +68,7 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
PYTHONPATH=. mkdocs build
|
PYTHONPATH=. mkdocs build
|
||||||
- name: Setup Pages
|
- name: Setup Pages
|
||||||
uses: actions/configure-pages@v5
|
uses: actions/configure-pages@v2
|
||||||
- name: Upload artifact
|
- name: Upload artifact
|
||||||
uses: actions/upload-pages-artifact@v3
|
uses: actions/upload-pages-artifact@v3
|
||||||
with:
|
with:
|
||||||
|
|||||||
9
.github/workflows/nodejs.yml
vendored
9
.github/workflows/nodejs.yml
vendored
@@ -7,7 +7,6 @@ on:
|
|||||||
pull_request:
|
pull_request:
|
||||||
paths:
|
paths:
|
||||||
- Cargo.toml
|
- Cargo.toml
|
||||||
- Cargo.lock
|
|
||||||
- nodejs/**
|
- nodejs/**
|
||||||
- rust/**
|
- rust/**
|
||||||
- docs/src/js/**
|
- docs/src/js/**
|
||||||
@@ -38,7 +37,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
lfs: true
|
lfs: true
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 20
|
node-version: 20
|
||||||
cache: 'npm'
|
cache: 'npm'
|
||||||
@@ -78,7 +77,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
lfs: true
|
lfs: true
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v3
|
||||||
name: Setup Node.js 20 for build
|
name: Setup Node.js 20 for build
|
||||||
with:
|
with:
|
||||||
# @napi-rs/cli v3 requires Node >= 20.12 (via @inquirer/prompts@8).
|
# @napi-rs/cli v3 requires Node >= 20.12 (via @inquirer/prompts@8).
|
||||||
@@ -95,7 +94,7 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
npm ci --include=optional
|
npm ci --include=optional
|
||||||
npm run build:debug -- --profile ci
|
npm run build:debug -- --profile ci
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v3
|
||||||
name: Setup Node.js ${{ matrix.node-version }} for test
|
name: Setup Node.js ${{ matrix.node-version }} for test
|
||||||
with:
|
with:
|
||||||
node-version: ${{ matrix.node-version }}
|
node-version: ${{ matrix.node-version }}
|
||||||
@@ -144,7 +143,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
lfs: true
|
lfs: true
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 20
|
node-version: 20
|
||||||
cache: 'npm'
|
cache: 'npm'
|
||||||
|
|||||||
17
.github/workflows/npm-publish.yml
vendored
17
.github/workflows/npm-publish.yml
vendored
@@ -19,7 +19,6 @@ on:
|
|||||||
paths:
|
paths:
|
||||||
- .github/workflows/npm-publish.yml
|
- .github/workflows/npm-publish.yml
|
||||||
- Cargo.toml # Change in dependency frequently breaks builds
|
- Cargo.toml # Change in dependency frequently breaks builds
|
||||||
- Cargo.lock
|
|
||||||
|
|
||||||
concurrency:
|
concurrency:
|
||||||
group: ${{ github.workflow }}-${{ github.ref }}
|
group: ${{ github.workflow }}-${{ github.ref }}
|
||||||
@@ -125,12 +124,7 @@ jobs:
|
|||||||
pre_build: |-
|
pre_build: |-
|
||||||
set -e &&
|
set -e &&
|
||||||
apt-get update &&
|
apt-get update &&
|
||||||
apt-get install -y protobuf-compiler pkg-config &&
|
apt-get install -y protobuf-compiler pkg-config
|
||||||
# The base image (manylinux2014-cross) sets TARGET_CC to the old
|
|
||||||
# GCC 4.8 cross-compiler. aws-lc-sys checks TARGET_CC before CC,
|
|
||||||
# so it picks up GCC even though the napi-rs image sets CC=clang.
|
|
||||||
# Override to use the image's clang-18 which supports -fuse-ld=lld.
|
|
||||||
export TARGET_CC=clang TARGET_CXX=clang++
|
|
||||||
- target: x86_64-unknown-linux-musl
|
- target: x86_64-unknown-linux-musl
|
||||||
# This one seems to need some extra memory
|
# This one seems to need some extra memory
|
||||||
host: ubuntu-2404-8x-x64
|
host: ubuntu-2404-8x-x64
|
||||||
@@ -150,10 +144,9 @@ jobs:
|
|||||||
set -e &&
|
set -e &&
|
||||||
apt-get update &&
|
apt-get update &&
|
||||||
apt-get install -y protobuf-compiler pkg-config &&
|
apt-get install -y protobuf-compiler pkg-config &&
|
||||||
export TARGET_CC=clang TARGET_CXX=clang++ &&
|
# https://github.com/aws/aws-lc-rs/issues/737#issuecomment-2725918627
|
||||||
# The manylinux2014 sysroot has glibc 2.17 headers which lack
|
ln -s /usr/aarch64-unknown-linux-gnu/lib/gcc/aarch64-unknown-linux-gnu/4.8.5/crtbeginS.o /usr/aarch64-unknown-linux-gnu/aarch64-unknown-linux-gnu/sysroot/usr/lib/crtbeginS.o &&
|
||||||
# AT_HWCAP2 (added in Linux 3.17). Define it for aws-lc-sys.
|
ln -s /usr/aarch64-unknown-linux-gnu/lib/gcc /usr/aarch64-unknown-linux-gnu/aarch64-unknown-linux-gnu/sysroot/usr/lib/gcc &&
|
||||||
export CFLAGS="$CFLAGS -DAT_HWCAP2=26" &&
|
|
||||||
rustup target add aarch64-unknown-linux-gnu
|
rustup target add aarch64-unknown-linux-gnu
|
||||||
- target: aarch64-unknown-linux-musl
|
- target: aarch64-unknown-linux-musl
|
||||||
host: ubuntu-2404-8x-x64
|
host: ubuntu-2404-8x-x64
|
||||||
@@ -273,7 +266,7 @@ jobs:
|
|||||||
- target: x86_64-unknown-linux-gnu
|
- target: x86_64-unknown-linux-gnu
|
||||||
host: ubuntu-latest
|
host: ubuntu-latest
|
||||||
- target: aarch64-unknown-linux-gnu
|
- target: aarch64-unknown-linux-gnu
|
||||||
host: ubuntu-2404-8x-arm64
|
host: buildjet-16vcpu-ubuntu-2204-arm
|
||||||
node:
|
node:
|
||||||
- '20'
|
- '20'
|
||||||
runs-on: ${{ matrix.settings.host }}
|
runs-on: ${{ matrix.settings.host }}
|
||||||
|
|||||||
1
.github/workflows/pypi-publish.yml
vendored
1
.github/workflows/pypi-publish.yml
vendored
@@ -9,7 +9,6 @@ on:
|
|||||||
paths:
|
paths:
|
||||||
- .github/workflows/pypi-publish.yml
|
- .github/workflows/pypi-publish.yml
|
||||||
- Cargo.toml # Change in dependency frequently breaks builds
|
- Cargo.toml # Change in dependency frequently breaks builds
|
||||||
- Cargo.lock
|
|
||||||
|
|
||||||
env:
|
env:
|
||||||
PIP_EXTRA_INDEX_URL: "https://pypi.fury.io/lance-format/ https://pypi.fury.io/lancedb/"
|
PIP_EXTRA_INDEX_URL: "https://pypi.fury.io/lance-format/ https://pypi.fury.io/lancedb/"
|
||||||
|
|||||||
1
.github/workflows/python.yml
vendored
1
.github/workflows/python.yml
vendored
@@ -7,7 +7,6 @@ on:
|
|||||||
pull_request:
|
pull_request:
|
||||||
paths:
|
paths:
|
||||||
- Cargo.toml
|
- Cargo.toml
|
||||||
- Cargo.lock
|
|
||||||
- python/**
|
- python/**
|
||||||
- rust/**
|
- rust/**
|
||||||
- .github/workflows/python.yml
|
- .github/workflows/python.yml
|
||||||
|
|||||||
1
.github/workflows/rust.yml
vendored
1
.github/workflows/rust.yml
vendored
@@ -7,7 +7,6 @@ on:
|
|||||||
pull_request:
|
pull_request:
|
||||||
paths:
|
paths:
|
||||||
- Cargo.toml
|
- Cargo.toml
|
||||||
- Cargo.lock
|
|
||||||
- rust/**
|
- rust/**
|
||||||
- .github/workflows/rust.yml
|
- .github/workflows/rust.yml
|
||||||
|
|
||||||
|
|||||||
377
Cargo.lock
generated
377
Cargo.lock
generated
@@ -547,7 +547,7 @@ dependencies = [
|
|||||||
"aws-sdk-ssooidc",
|
"aws-sdk-ssooidc",
|
||||||
"aws-sdk-sts",
|
"aws-sdk-sts",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -567,9 +567,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-credential-types"
|
name = "aws-credential-types"
|
||||||
version = "1.2.12"
|
version = "1.2.8"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "e26bbf46abc608f2dc61fd6cb3b7b0665497cc259a21520151ed98f8b37d2c79"
|
checksum = "faf26925f4a5b59eb76722b63c2892b1d70d06fa053c72e4a100ec308c1d47bc"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -579,9 +579,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-lc-rs"
|
name = "aws-lc-rs"
|
||||||
version = "1.16.1"
|
version = "1.13.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "94bffc006df10ac2a68c83692d734a465f8ee6c5b384d8545a636f81d858f4bf"
|
checksum = "19b756939cb2f8dc900aa6dcd505e6e2428e9cae7ff7b028c49e3946efa70878"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-lc-sys",
|
"aws-lc-sys",
|
||||||
"untrusted 0.7.1",
|
"untrusted 0.7.1",
|
||||||
@@ -590,10 +590,11 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-lc-sys"
|
name = "aws-lc-sys"
|
||||||
version = "0.38.0"
|
version = "0.28.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "4321e568ed89bb5a7d291a7f37997c2c0df89809d7b6d12062c81ddb54aa782e"
|
checksum = "b9f7720b74ed28ca77f90769a71fd8c637a0137f6fae4ae947e1050229cff57f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
|
"bindgen",
|
||||||
"cc",
|
"cc",
|
||||||
"cmake",
|
"cmake",
|
||||||
"dunce",
|
"dunce",
|
||||||
@@ -602,26 +603,23 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-runtime"
|
name = "aws-runtime"
|
||||||
version = "1.7.0"
|
version = "1.5.12"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "b0f92058d22a46adf53ec57a6a96f34447daf02bff52e8fb956c66bcd5c6ac12"
|
checksum = "bfa006bb32360ed90ac51203feafb9d02e3d21046e1fd3a450a404b90ea73e5d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-sigv4",
|
"aws-sigv4",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-eventstream",
|
"aws-smithy-eventstream",
|
||||||
"aws-smithy-http 0.63.4",
|
"aws-smithy-http",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
"aws-types",
|
"aws-types",
|
||||||
"bytes",
|
"bytes",
|
||||||
"bytes-utils",
|
|
||||||
"fastrand",
|
"fastrand",
|
||||||
"http 0.2.12",
|
"http 0.2.12",
|
||||||
"http 1.3.1",
|
|
||||||
"http-body 0.4.6",
|
"http-body 0.4.6",
|
||||||
"http-body 1.0.1",
|
|
||||||
"percent-encoding",
|
"percent-encoding",
|
||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"tracing",
|
"tracing",
|
||||||
@@ -630,16 +628,16 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-bedrockruntime"
|
name = "aws-sdk-bedrockruntime"
|
||||||
version = "1.113.0"
|
version = "1.109.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "d5d2b8f081b9e8ff455b8dd7387b6b02263c3dac73172d188d2b523ff1e775e9"
|
checksum = "fbfdfd941dcb253c17bf70baddbf1e5b22f19e29d313d2e049bad4b1dadb2011"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-sigv4",
|
"aws-sigv4",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-eventstream",
|
"aws-smithy-eventstream",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -655,14 +653,14 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-dynamodb"
|
name = "aws-sdk-dynamodb"
|
||||||
version = "1.96.0"
|
version = "1.95.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "a2e1f3871da847c7ff682ddd2c2b802fa10a562db34eaf16dd863fc2d3f72ad6"
|
checksum = "699bab2e1f9da570071889b674a43df1184b6ec42bf4770ef8758a8f3e02f38f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -677,14 +675,14 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-kms"
|
name = "aws-sdk-kms"
|
||||||
version = "1.91.0"
|
version = "1.89.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "30662f9c5a2be685ee921f92d5ad88d5bc6ac971e8c0e49c5907bd52b2b069df"
|
checksum = "fcf2158ad0759016eb2d36b6eae2365f5c93af47270403b92ad58b75dee5e4df"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -699,9 +697,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-s3"
|
name = "aws-sdk-s3"
|
||||||
version = "1.109.0"
|
version = "1.108.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "3c6d81b75f8ff78882e70c5909804b44553d56136899fb4015a0a68ecc870e0e"
|
checksum = "200be4aed61e3c0669f7268bacb768f283f1c32a7014ce57225e1160be2f6ccb"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
@@ -709,7 +707,7 @@ dependencies = [
|
|||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-checksums",
|
"aws-smithy-checksums",
|
||||||
"aws-smithy-eventstream",
|
"aws-smithy-eventstream",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -733,14 +731,14 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-sso"
|
name = "aws-sdk-sso"
|
||||||
version = "1.87.0"
|
version = "1.86.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "f4af747ffcb5aa8da8be8f0679ef6940f1afdb8c2e10c36738c9ebeb8d17b95e"
|
checksum = "4a0abbfab841446cce6e87af853a3ba2cc1bc9afcd3f3550dd556c43d434c86d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -755,14 +753,14 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-ssooidc"
|
name = "aws-sdk-ssooidc"
|
||||||
version = "1.89.0"
|
version = "1.88.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "695dc67bb861ccb8426c9129b91c30e266a0e3d85650cafdf62fcca14c8fd338"
|
checksum = "9a68d675582afea0e94d38b6ca9c5aaae4ca14f1d36faa6edb19b42e687e70d7"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -777,14 +775,14 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sdk-sts"
|
name = "aws-sdk-sts"
|
||||||
version = "1.89.0"
|
version = "1.88.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "928e87698cd916cf1efd5268148347269e6d2911028742c0061ff6261e639e3c"
|
checksum = "d30990923f4f675523c51eb1c0dec9b752fb267b36a61e83cbc219c9d86da715"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-runtime",
|
"aws-runtime",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-json",
|
"aws-smithy-json",
|
||||||
"aws-smithy-query",
|
"aws-smithy-query",
|
||||||
"aws-smithy-runtime",
|
"aws-smithy-runtime",
|
||||||
@@ -800,13 +798,13 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-sigv4"
|
name = "aws-sigv4"
|
||||||
version = "1.4.0"
|
version = "1.3.5"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "68f6ae9b71597dc5fd115d52849d7a5556ad9265885ad3492ea8d73b93bbc46e"
|
checksum = "bffc03068fbb9c8dd5ce1c6fb240678a5cffb86fb2b7b1985c999c4b83c8df68"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-smithy-eventstream",
|
"aws-smithy-eventstream",
|
||||||
"aws-smithy-http 0.63.4",
|
"aws-smithy-http",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
"bytes",
|
"bytes",
|
||||||
@@ -828,9 +826,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-async"
|
name = "aws-smithy-async"
|
||||||
version = "1.2.12"
|
version = "1.2.6"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "3cba48474f1d6807384d06fec085b909f5807e16653c5af5c45dfe89539f0b70"
|
checksum = "127fcfad33b7dfc531141fda7e1c402ac65f88aca5511a4d31e2e3d2cd01ce9c"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"futures-util",
|
"futures-util",
|
||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
@@ -843,7 +841,7 @@ version = "0.63.9"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "165d8583d8d906e2fb5511d29201d447cc710864f075debcdd9c31c265412806"
|
checksum = "165d8583d8d906e2fb5511d29201d447cc710864f075debcdd9c31c265412806"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-http 0.62.6",
|
"aws-smithy-http",
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
"bytes",
|
"bytes",
|
||||||
"crc-fast",
|
"crc-fast",
|
||||||
@@ -859,9 +857,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-eventstream"
|
name = "aws-smithy-eventstream"
|
||||||
version = "0.60.19"
|
version = "0.60.12"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "1c0b3e587fbaa5d7f7e870544508af8ce82ea47cd30376e69e1e37c4ac746f79"
|
checksum = "9656b85088f8d9dc7ad40f9a6c7228e1e8447cdf4b046c87e152e0805dea02fa"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
"bytes",
|
"bytes",
|
||||||
@@ -870,9 +868,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-http"
|
name = "aws-smithy-http"
|
||||||
version = "0.62.6"
|
version = "0.62.4"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "826141069295752372f8203c17f28e30c464d22899a43a0c9fd9c458d469c88b"
|
checksum = "3feafd437c763db26aa04e0cc7591185d0961e64c61885bece0fb9d50ceac671"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-eventstream",
|
"aws-smithy-eventstream",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -880,7 +878,6 @@ dependencies = [
|
|||||||
"bytes",
|
"bytes",
|
||||||
"bytes-utils",
|
"bytes-utils",
|
||||||
"futures-core",
|
"futures-core",
|
||||||
"futures-util",
|
|
||||||
"http 0.2.12",
|
"http 0.2.12",
|
||||||
"http 1.3.1",
|
"http 1.3.1",
|
||||||
"http-body 0.4.6",
|
"http-body 0.4.6",
|
||||||
@@ -890,32 +887,11 @@ dependencies = [
|
|||||||
"tracing",
|
"tracing",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "aws-smithy-http"
|
|
||||||
version = "0.63.4"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "af4a8a5fe3e4ac7ee871237c340bbce13e982d37543b65700f4419e039f5d78e"
|
|
||||||
dependencies = [
|
|
||||||
"aws-smithy-runtime-api",
|
|
||||||
"aws-smithy-types",
|
|
||||||
"bytes",
|
|
||||||
"bytes-utils",
|
|
||||||
"futures-core",
|
|
||||||
"futures-util",
|
|
||||||
"http 1.3.1",
|
|
||||||
"http-body 1.0.1",
|
|
||||||
"http-body-util",
|
|
||||||
"percent-encoding",
|
|
||||||
"pin-project-lite",
|
|
||||||
"pin-utils",
|
|
||||||
"tracing",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-http-client"
|
name = "aws-smithy-http-client"
|
||||||
version = "1.1.10"
|
version = "1.1.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "0709f0083aa19b704132684bc26d3c868e06bd428ccc4373b0b55c3e8748a58b"
|
checksum = "1053b5e587e6fa40ce5a79ea27957b04ba660baa02b28b7436f64850152234f1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -933,7 +909,7 @@ dependencies = [
|
|||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"rustls 0.21.12",
|
"rustls 0.21.12",
|
||||||
"rustls 0.23.31",
|
"rustls 0.23.31",
|
||||||
"rustls-native-certs",
|
"rustls-native-certs 0.8.1",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-rustls 0.26.4",
|
"tokio-rustls 0.26.4",
|
||||||
@@ -943,18 +919,18 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-json"
|
name = "aws-smithy-json"
|
||||||
version = "0.61.9"
|
version = "0.61.6"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "49fa1213db31ac95288d981476f78d05d9cbb0353d22cdf3472cc05bb02f6551"
|
checksum = "cff418fc8ec5cadf8173b10125f05c2e7e1d46771406187b2c878557d4503390"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-observability"
|
name = "aws-smithy-observability"
|
||||||
version = "0.2.5"
|
version = "0.1.4"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "4d3f39d5bb871aaf461d59144557f16d5927a5248a983a40654d9cf3b9ba183b"
|
checksum = "2d1881b1ea6d313f9890710d65c158bdab6fb08c91ea825f74c1c8c357baf4cc"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
]
|
]
|
||||||
@@ -971,12 +947,12 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-runtime"
|
name = "aws-smithy-runtime"
|
||||||
version = "1.10.1"
|
version = "1.9.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "8fd3dfc18c1ce097cf81fced7192731e63809829c6cbf933c1ec47452d08e1aa"
|
checksum = "40ab99739082da5347660c556689256438defae3bcefd66c52b095905730e404"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-http 0.63.4",
|
"aws-smithy-http",
|
||||||
"aws-smithy-http-client",
|
"aws-smithy-http-client",
|
||||||
"aws-smithy-observability",
|
"aws-smithy-observability",
|
||||||
"aws-smithy-runtime-api",
|
"aws-smithy-runtime-api",
|
||||||
@@ -987,7 +963,6 @@ dependencies = [
|
|||||||
"http 1.3.1",
|
"http 1.3.1",
|
||||||
"http-body 0.4.6",
|
"http-body 0.4.6",
|
||||||
"http-body 1.0.1",
|
"http-body 1.0.1",
|
||||||
"http-body-util",
|
|
||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"pin-utils",
|
"pin-utils",
|
||||||
"tokio",
|
"tokio",
|
||||||
@@ -996,9 +971,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-runtime-api"
|
name = "aws-smithy-runtime-api"
|
||||||
version = "1.11.4"
|
version = "1.9.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "8c55e0837e9b8526f49e0b9bfa9ee18ddee70e853f5bc09c5d11ebceddcb0fec"
|
checksum = "3683c5b152d2ad753607179ed71988e8cfd52964443b4f74fd8e552d0bbfeb46"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
"aws-smithy-types",
|
"aws-smithy-types",
|
||||||
@@ -1013,9 +988,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-smithy-types"
|
name = "aws-smithy-types"
|
||||||
version = "1.4.4"
|
version = "1.3.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "576b0d6991c9c32bc14fc340582ef148311f924d41815f641a308b5d11e8e7cd"
|
checksum = "9f5b3a7486f6690ba25952cabf1e7d75e34d69eaff5081904a47bc79074d6457"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"base64-simd",
|
"base64-simd",
|
||||||
"bytes",
|
"bytes",
|
||||||
@@ -1048,9 +1023,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "aws-types"
|
name = "aws-types"
|
||||||
version = "1.3.12"
|
version = "1.3.9"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "6c50f3cdf47caa8d01f2be4a6663ea02418e892f9bbfd82c7b9a3a37eaccdd3a"
|
checksum = "e2fd329bf0e901ff3f60425691410c69094dc2a1f34b331f37bfc4e9ac1565a1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"aws-credential-types",
|
"aws-credential-types",
|
||||||
"aws-smithy-async",
|
"aws-smithy-async",
|
||||||
@@ -1193,6 +1168,29 @@ dependencies = [
|
|||||||
"num-traits",
|
"num-traits",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "bindgen"
|
||||||
|
version = "0.69.5"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "271383c67ccabffb7381723dea0672a673f292304fcb45c01cc648c7a8d58088"
|
||||||
|
dependencies = [
|
||||||
|
"bitflags 2.9.4",
|
||||||
|
"cexpr",
|
||||||
|
"clang-sys",
|
||||||
|
"itertools 0.12.1",
|
||||||
|
"lazy_static",
|
||||||
|
"lazycell",
|
||||||
|
"log",
|
||||||
|
"prettyplease",
|
||||||
|
"proc-macro2",
|
||||||
|
"quote",
|
||||||
|
"regex",
|
||||||
|
"rustc-hash 1.1.0",
|
||||||
|
"shlex",
|
||||||
|
"syn 2.0.114",
|
||||||
|
"which",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "bit-set"
|
name = "bit-set"
|
||||||
version = "0.5.3"
|
version = "0.5.3"
|
||||||
@@ -1479,6 +1477,15 @@ version = "0.4.2"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "4f4c707c6a209cbe82d10abd08e1ea8995e9ea937d2550646e02798948992be0"
|
checksum = "4f4c707c6a209cbe82d10abd08e1ea8995e9ea937d2550646e02798948992be0"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "cexpr"
|
||||||
|
version = "0.6.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "6fac387a98bb7c37292057cffc56d62ecb629900026402633ae9160df93a8766"
|
||||||
|
dependencies = [
|
||||||
|
"nom 7.1.3",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cfg-if"
|
name = "cfg-if"
|
||||||
version = "1.0.3"
|
version = "1.0.3"
|
||||||
@@ -1547,6 +1554,17 @@ dependencies = [
|
|||||||
"inout",
|
"inout",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "clang-sys"
|
||||||
|
version = "1.8.1"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "0b023947811758c97c59bf9d1c188fd619ad4718dcaa767947df1cadb14f39f4"
|
||||||
|
dependencies = [
|
||||||
|
"glob",
|
||||||
|
"libc",
|
||||||
|
"libloading 0.8.9",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cmake"
|
name = "cmake"
|
||||||
version = "0.1.54"
|
version = "0.1.54"
|
||||||
@@ -3070,9 +3088,8 @@ checksum = "42703706b716c37f96a77aea830392ad231f44c9e9a67872fa5548707e11b11c"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "fsst"
|
name = "fsst"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "a32ddfc5478379cd1782bdd9d7d1411063f563e5b338fc73bafe5916451a5b9d"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"rand 0.9.2",
|
"rand 0.9.2",
|
||||||
@@ -3813,6 +3830,7 @@ dependencies = [
|
|||||||
"hyper 0.14.32",
|
"hyper 0.14.32",
|
||||||
"log",
|
"log",
|
||||||
"rustls 0.21.12",
|
"rustls 0.21.12",
|
||||||
|
"rustls-native-certs 0.6.3",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-rustls 0.24.1",
|
"tokio-rustls 0.24.1",
|
||||||
]
|
]
|
||||||
@@ -3827,7 +3845,7 @@ dependencies = [
|
|||||||
"hyper 1.7.0",
|
"hyper 1.7.0",
|
||||||
"hyper-util",
|
"hyper-util",
|
||||||
"rustls 0.23.31",
|
"rustls 0.23.31",
|
||||||
"rustls-native-certs",
|
"rustls-native-certs 0.8.1",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-rustls 0.26.4",
|
"tokio-rustls 0.26.4",
|
||||||
@@ -4242,9 +4260,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance"
|
name = "lance"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "95c5ce428fda0721f5c48bfde17a1921c4da2d2142b2f46a16c89abf5fce8003"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-arith",
|
"arrow-arith",
|
||||||
@@ -4310,9 +4327,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-arrow"
|
name = "lance-arrow"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "c9fdaf99863fa0d631e422881e88be4837d8b82f36a87143d723a9d285acec4b"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"arrow-buffer",
|
"arrow-buffer",
|
||||||
@@ -4332,9 +4348,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-bitpacking"
|
name = "lance-bitpacking"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "866b1634d38d94e8ab86fbcf238ac82dc8a5f72a4a6a90525f29899772e7cc7f"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrayref",
|
"arrayref",
|
||||||
"paste",
|
"paste",
|
||||||
@@ -4343,9 +4358,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-core"
|
name = "lance-core"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "977c29f4e48c201c2806fe6ae117b65d0287eda236acd07357b556a54b0d5c5a"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"arrow-buffer",
|
"arrow-buffer",
|
||||||
@@ -4382,9 +4396,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-datafusion"
|
name = "lance-datafusion"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "0ccc72695473f4207df4c6df3b347a63e84c32c0bc36bf42a7d86e8a7c0c67e2"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
@@ -4414,9 +4427,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-datagen"
|
name = "lance-datagen"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "8fe84d76944acd834ded14d7562663af995556e0c6594f4b4ac69b0183f99c1a"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
@@ -4434,9 +4446,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-encoding"
|
name = "lance-encoding"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "be1007242188e5d53c98717e7f2cb340dc80eb9c94c2b935587598919b3a36bd"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-arith",
|
"arrow-arith",
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
@@ -4473,9 +4484,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-file"
|
name = "lance-file"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "f80088e418941f39cf5599d166ae1a6ef498cc2d967652a0692477d4871a9277"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-arith",
|
"arrow-arith",
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
@@ -4507,9 +4517,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-index"
|
name = "lance-index"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "e0011daf1ddde99becffd2ae235ad324576736a526c54ffbc4d7e583872f1215"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-arith",
|
"arrow-arith",
|
||||||
@@ -4572,9 +4581,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-io"
|
name = "lance-io"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "cfa8a74e93753d19a27ce3adaeb99e31227df13ad5926dd43572be76b43dd284"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-arith",
|
"arrow-arith",
|
||||||
@@ -4615,9 +4623,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-linalg"
|
name = "lance-linalg"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "6e2d8da8f6b8dd37ab3b8199896ee265817f86232e3727c0b0eeb3c9093b64d9"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"arrow-buffer",
|
"arrow-buffer",
|
||||||
@@ -4633,9 +4640,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-namespace"
|
name = "lance-namespace"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "f176e427d9c35938d8a7097876114bc35dfd280b06077779753f2effe3e86aab"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"async-trait",
|
"async-trait",
|
||||||
@@ -4647,9 +4653,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-namespace-impls"
|
name = "lance-namespace-impls"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "663c32086ecfab311acb0813c65a4bb352a5b648ccf8b513c24697ce8d412039"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-ipc",
|
"arrow-ipc",
|
||||||
@@ -4693,9 +4698,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-table"
|
name = "lance-table"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "aa189b3081481a97b64cf1161297947a63b8adb941b1950989d0269858703a43"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
@@ -4734,9 +4738,8 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lance-testing"
|
name = "lance-testing"
|
||||||
version = "3.0.1"
|
version = "3.0.0-rc.3"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "git+https://github.com/lance-format/lance.git?tag=v3.0.0-rc.3#de393a26a068dd297929ca7d798e43dc31c57337"
|
||||||
checksum = "79a6f4ab0788ee82893bac5de4ff0d0d88bba96de87db4b6e18b1883616d4dbe"
|
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"arrow-schema",
|
"arrow-schema",
|
||||||
@@ -4747,7 +4750,7 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lancedb"
|
name = "lancedb"
|
||||||
version = "0.27.0"
|
version = "0.27.0-beta.3"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"ahash",
|
"ahash",
|
||||||
"anyhow",
|
"anyhow",
|
||||||
@@ -4829,7 +4832,7 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lancedb-nodejs"
|
name = "lancedb-nodejs"
|
||||||
version = "0.27.0"
|
version = "0.27.0-beta.3"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow-array",
|
"arrow-array",
|
||||||
"arrow-ipc",
|
"arrow-ipc",
|
||||||
@@ -4849,24 +4852,20 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lancedb-python"
|
name = "lancedb-python"
|
||||||
version = "0.30.0"
|
version = "0.30.0-beta.3"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"arrow",
|
"arrow",
|
||||||
"async-trait",
|
"async-trait",
|
||||||
"bytes",
|
|
||||||
"env_logger",
|
"env_logger",
|
||||||
"futures",
|
"futures",
|
||||||
"lance-core",
|
"lance-core",
|
||||||
"lance-io",
|
"lance-io",
|
||||||
"lance-namespace",
|
"lance-namespace",
|
||||||
"lance-namespace-impls",
|
|
||||||
"lancedb",
|
"lancedb",
|
||||||
"pin-project",
|
"pin-project",
|
||||||
"pyo3",
|
"pyo3",
|
||||||
"pyo3-async-runtimes",
|
"pyo3-async-runtimes",
|
||||||
"pyo3-build-config",
|
"pyo3-build-config",
|
||||||
"serde",
|
|
||||||
"serde_json",
|
|
||||||
"snafu 0.8.9",
|
"snafu 0.8.9",
|
||||||
"tokio",
|
"tokio",
|
||||||
]
|
]
|
||||||
@@ -4880,6 +4879,12 @@ dependencies = [
|
|||||||
"spin",
|
"spin",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "lazycell"
|
||||||
|
version = "1.3.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "830d08ce1d1d941e6b30645f1a0eb5643013d835ce3779a5fc208261dbe10f55"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "leb128fmt"
|
name = "leb128fmt"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
@@ -5332,7 +5337,7 @@ dependencies = [
|
|||||||
"napi-build",
|
"napi-build",
|
||||||
"napi-sys",
|
"napi-sys",
|
||||||
"nohash-hasher",
|
"nohash-hasher",
|
||||||
"rustc-hash",
|
"rustc-hash 2.1.1",
|
||||||
"tokio",
|
"tokio",
|
||||||
]
|
]
|
||||||
|
|
||||||
@@ -5497,9 +5502,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "num-conv"
|
name = "num-conv"
|
||||||
version = "0.2.0"
|
version = "0.1.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "cf97ec579c3c42f953ef76dbf8d55ac91fb219dde70e49aa4a6b7d74e9919050"
|
checksum = "51d515d32fb182ee37cda2ccdcb92950d6a3c2893aa280e540671c2cd0f3b1d9"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "num-integer"
|
name = "num-integer"
|
||||||
@@ -5605,7 +5610,7 @@ dependencies = [
|
|||||||
"rand 0.9.2",
|
"rand 0.9.2",
|
||||||
"reqwest",
|
"reqwest",
|
||||||
"ring",
|
"ring",
|
||||||
"rustls-pemfile",
|
"rustls-pemfile 2.2.0",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"serde_urlencoded",
|
"serde_urlencoded",
|
||||||
@@ -5648,9 +5653,9 @@ checksum = "a4895175b425cb1f87721b59f0f286c2092bd4af812243672510e1ac53e2e0ad"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "oneshot"
|
name = "oneshot"
|
||||||
version = "0.1.13"
|
version = "0.1.11"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "269bca4c2591a28585d6bf10d9ed0332b7d76900a1b02bec41bdc3a2cdcda107"
|
checksum = "b4ce411919553d3f9fa53a0880544cda985a112117a0444d5ff1e870a893d6ea"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "onig"
|
name = "onig"
|
||||||
@@ -6646,7 +6651,7 @@ dependencies = [
|
|||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"quinn-proto",
|
"quinn-proto",
|
||||||
"quinn-udp",
|
"quinn-udp",
|
||||||
"rustc-hash",
|
"rustc-hash 2.1.1",
|
||||||
"rustls 0.23.31",
|
"rustls 0.23.31",
|
||||||
"socket2 0.6.0",
|
"socket2 0.6.0",
|
||||||
"thiserror 2.0.17",
|
"thiserror 2.0.17",
|
||||||
@@ -6666,7 +6671,7 @@ dependencies = [
|
|||||||
"lru-slab",
|
"lru-slab",
|
||||||
"rand 0.9.2",
|
"rand 0.9.2",
|
||||||
"ring",
|
"ring",
|
||||||
"rustc-hash",
|
"rustc-hash 2.1.1",
|
||||||
"rustls 0.23.31",
|
"rustls 0.23.31",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"slab",
|
"slab",
|
||||||
@@ -7053,7 +7058,7 @@ dependencies = [
|
|||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"quinn",
|
"quinn",
|
||||||
"rustls 0.23.31",
|
"rustls 0.23.31",
|
||||||
"rustls-native-certs",
|
"rustls-native-certs 0.8.1",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
@@ -7126,9 +7131,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rsa"
|
name = "rsa"
|
||||||
version = "0.9.10"
|
version = "0.9.8"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "b8573f03f5883dcaebdfcf4725caa1ecb9c15b2ef50c43a07b816e06799bb12d"
|
checksum = "78928ac1ed176a5ca1d17e578a1825f3d81ca54cf41053a592584b020cfd691b"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"const-oid",
|
"const-oid",
|
||||||
"digest",
|
"digest",
|
||||||
@@ -7195,6 +7200,12 @@ dependencies = [
|
|||||||
"serde_derive",
|
"serde_derive",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "rustc-hash"
|
||||||
|
version = "1.1.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "08d43f7aa6b08d49f382cde6a7982047c3426db949b1424bc4b7ec9ae12c6ce2"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rustc-hash"
|
name = "rustc-hash"
|
||||||
version = "2.1.1"
|
version = "2.1.1"
|
||||||
@@ -7264,6 +7275,18 @@ dependencies = [
|
|||||||
"zeroize",
|
"zeroize",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "rustls-native-certs"
|
||||||
|
version = "0.6.3"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "a9aace74cb666635c918e9c12bc0d348266037aa8eb599b5cba565709a8dff00"
|
||||||
|
dependencies = [
|
||||||
|
"openssl-probe",
|
||||||
|
"rustls-pemfile 1.0.4",
|
||||||
|
"schannel",
|
||||||
|
"security-framework 2.11.1",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rustls-native-certs"
|
name = "rustls-native-certs"
|
||||||
version = "0.8.1"
|
version = "0.8.1"
|
||||||
@@ -7273,7 +7296,16 @@ dependencies = [
|
|||||||
"openssl-probe",
|
"openssl-probe",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"schannel",
|
"schannel",
|
||||||
"security-framework",
|
"security-framework 3.5.1",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "rustls-pemfile"
|
||||||
|
version = "1.0.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "1c74cae0a4cf6ccbbf5f359f08efdf8ee7e1dc532573bf0db71968cb56b1448c"
|
||||||
|
dependencies = [
|
||||||
|
"base64 0.21.7",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
@@ -7447,6 +7479,19 @@ dependencies = [
|
|||||||
"zeroize",
|
"zeroize",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "security-framework"
|
||||||
|
version = "2.11.1"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "897b2245f0b511c87893af39b033e5ca9cce68824c4d7e7630b5a1d339658d02"
|
||||||
|
dependencies = [
|
||||||
|
"bitflags 2.9.4",
|
||||||
|
"core-foundation 0.9.4",
|
||||||
|
"core-foundation-sys",
|
||||||
|
"libc",
|
||||||
|
"security-framework-sys",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "security-framework"
|
name = "security-framework"
|
||||||
version = "3.5.1"
|
version = "3.5.1"
|
||||||
@@ -7763,7 +7808,7 @@ version = "0.9.0"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "54254b8531cafa275c5e096f62d48c81435d1015405a91198ddb11e967301d40"
|
checksum = "54254b8531cafa275c5e096f62d48c81435d1015405a91198ddb11e967301d40"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"heck 0.5.0",
|
"heck 0.4.1",
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
"syn 2.0.114",
|
"syn 2.0.114",
|
||||||
@@ -8117,7 +8162,7 @@ dependencies = [
|
|||||||
"rayon",
|
"rayon",
|
||||||
"regex",
|
"regex",
|
||||||
"rust-stemmers",
|
"rust-stemmers",
|
||||||
"rustc-hash",
|
"rustc-hash 2.1.1",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"sketches-ddsketch",
|
"sketches-ddsketch",
|
||||||
@@ -8343,30 +8388,30 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "time"
|
name = "time"
|
||||||
version = "0.3.47"
|
version = "0.3.44"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "743bd48c283afc0388f9b8827b976905fb217ad9e647fae3a379a9283c4def2c"
|
checksum = "91e7d9e3bb61134e77bde20dd4825b97c010155709965fedf0f49bb138e52a9d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"deranged",
|
"deranged",
|
||||||
"itoa",
|
"itoa",
|
||||||
"num-conv",
|
"num-conv",
|
||||||
"powerfmt",
|
"powerfmt",
|
||||||
"serde_core",
|
"serde",
|
||||||
"time-core",
|
"time-core",
|
||||||
"time-macros",
|
"time-macros",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "time-core"
|
name = "time-core"
|
||||||
version = "0.1.8"
|
version = "0.1.6"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "7694e1cfe791f8d31026952abf09c69ca6f6fa4e1a1229e18988f06a04a12dca"
|
checksum = "40868e7c1d2f0b8d73e4a8c7f0ff63af4f6d19be117e90bd73eb1d62cf831c6b"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "time-macros"
|
name = "time-macros"
|
||||||
version = "0.2.27"
|
version = "0.2.24"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "2e70e4c5a0e0a8a4823ad65dfe1a6930e4f4d756dcd9dd7939022b5e8c501215"
|
checksum = "30cfb0125f12d9c277f35663a0a33f8c30190f4e4574868a330595412d34ebf3"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"num-conv",
|
"num-conv",
|
||||||
"time-core",
|
"time-core",
|
||||||
@@ -9073,6 +9118,18 @@ dependencies = [
|
|||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "which"
|
||||||
|
version = "4.4.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "87ba24419a2078cd2b0f2ede2691b6c66d8e47836da3b6db8265ebad47afbfc7"
|
||||||
|
dependencies = [
|
||||||
|
"either",
|
||||||
|
"home",
|
||||||
|
"once_cell",
|
||||||
|
"rustix 0.38.44",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "winapi"
|
name = "winapi"
|
||||||
version = "0.3.9"
|
version = "0.3.9"
|
||||||
|
|||||||
28
Cargo.toml
28
Cargo.toml
@@ -15,20 +15,20 @@ categories = ["database-implementations"]
|
|||||||
rust-version = "1.91.0"
|
rust-version = "1.91.0"
|
||||||
|
|
||||||
[workspace.dependencies]
|
[workspace.dependencies]
|
||||||
lance = { version = "=3.0.1", default-features = false }
|
lance = { "version" = "=3.0.0-rc.3", default-features = false, "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-core = { version = "=3.0.1" }
|
lance-core = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-datagen = { version = "=3.0.1" }
|
lance-datagen = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-file = { version = "=3.0.1" }
|
lance-file = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-io = { version = "=3.0.1", default-features = false }
|
lance-io = { "version" = "=3.0.0-rc.3", default-features = false, "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-index = { version = "=3.0.1" }
|
lance-index = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-linalg = { version = "=3.0.1" }
|
lance-linalg = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-namespace = { version = "=3.0.1" }
|
lance-namespace = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-namespace-impls = { version = "=3.0.1", default-features = false }
|
lance-namespace-impls = { "version" = "=3.0.0-rc.3", default-features = false, "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-table = { version = "=3.0.1" }
|
lance-table = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-testing = { version = "=3.0.1" }
|
lance-testing = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-datafusion = { version = "=3.0.1" }
|
lance-datafusion = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-encoding = { version = "=3.0.1" }
|
lance-encoding = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
lance-arrow = { version = "=3.0.1" }
|
lance-arrow = { "version" = "=3.0.0-rc.3", "tag" = "v3.0.0-rc.3", "git" = "https://github.com/lance-format/lance.git" }
|
||||||
ahash = "0.8"
|
ahash = "0.8"
|
||||||
# Note that this one does not include pyarrow
|
# Note that this one does not include pyarrow
|
||||||
arrow = { version = "57.2", optional = false }
|
arrow = { version = "57.2", optional = false }
|
||||||
|
|||||||
@@ -3,7 +3,6 @@
|
|||||||
from __future__ import annotations
|
from __future__ import annotations
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
import functools
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
@@ -27,7 +26,6 @@ SEMVER_RE = re.compile(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@functools.total_ordering
|
|
||||||
@dataclass(frozen=True)
|
@dataclass(frozen=True)
|
||||||
class SemVer:
|
class SemVer:
|
||||||
major: int
|
major: int
|
||||||
@@ -158,9 +156,7 @@ def read_current_version(repo_root: Path) -> str:
|
|||||||
|
|
||||||
|
|
||||||
def determine_latest_tag(tags: Iterable[TagInfo]) -> TagInfo:
|
def determine_latest_tag(tags: Iterable[TagInfo]) -> TagInfo:
|
||||||
# Stable releases (no prerelease) are always preferred over pre-releases.
|
return max(tags, key=lambda tag: tag.semver)
|
||||||
# Within each group, standard semver ordering applies.
|
|
||||||
return max(tags, key=lambda tag: (not tag.semver.prerelease, tag.semver))
|
|
||||||
|
|
||||||
|
|
||||||
def write_outputs(args: argparse.Namespace, payload: dict) -> None:
|
def write_outputs(args: argparse.Namespace, payload: dict) -> None:
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
version: "3.9"
|
version: "3.9"
|
||||||
services:
|
services:
|
||||||
localstack:
|
localstack:
|
||||||
image: localstack/localstack:4.0
|
image: localstack/localstack:3.3
|
||||||
ports:
|
ports:
|
||||||
- 4566:4566
|
- 4566:4566
|
||||||
environment:
|
environment:
|
||||||
|
|||||||
@@ -1,27 +1,27 @@
|
|||||||
# Simple base dockerfile that supports basic dependencies required to run lance with FTS and Hybrid Search
|
#Simple base dockerfile that supports basic dependencies required to run lance with FTS and Hybrid Search
|
||||||
# Usage: docker build -t lancedb:latest -f Dockerfile .
|
#Usage docker build -t lancedb:latest -f Dockerfile .
|
||||||
FROM python:3.12-slim-bookworm
|
FROM python:3.10-slim-buster
|
||||||
|
|
||||||
# Install build dependencies in a single layer
|
# Install Rust
|
||||||
RUN apt-get update && \
|
RUN apt-get update && apt-get install -y curl build-essential && \
|
||||||
apt-get install -y --no-install-recommends \
|
curl https://sh.rustup.rs -sSf | sh -s -- -y
|
||||||
curl \
|
|
||||||
build-essential \
|
|
||||||
protobuf-compiler \
|
|
||||||
git \
|
|
||||||
ca-certificates && \
|
|
||||||
apt-get clean && \
|
|
||||||
rm -rf /var/lib/apt/lists/*
|
|
||||||
|
|
||||||
# Install Rust (pinned installer, non-interactive)
|
|
||||||
RUN curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable --profile minimal
|
|
||||||
|
|
||||||
# Set the environment variable for Rust
|
# Set the environment variable for Rust
|
||||||
ENV PATH="/root/.cargo/bin:${PATH}"
|
ENV PATH="/root/.cargo/bin:${PATH}"
|
||||||
|
|
||||||
|
# Install protobuf compiler
|
||||||
|
RUN apt-get install -y protobuf-compiler && \
|
||||||
|
apt-get clean && \
|
||||||
|
rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
|
RUN apt-get -y update &&\
|
||||||
|
apt-get -y upgrade && \
|
||||||
|
apt-get -y install git
|
||||||
|
|
||||||
|
|
||||||
# Verify installations
|
# Verify installations
|
||||||
RUN python --version && \
|
RUN python --version && \
|
||||||
rustc --version && \
|
rustc --version && \
|
||||||
protoc --version
|
protoc --version
|
||||||
|
|
||||||
RUN pip install --no-cache-dir tantivy lancedb
|
RUN pip install tantivy lancedb
|
||||||
|
|||||||
@@ -52,21 +52,14 @@ plugins:
|
|||||||
options:
|
options:
|
||||||
docstring_style: numpy
|
docstring_style: numpy
|
||||||
heading_level: 3
|
heading_level: 3
|
||||||
|
show_source: true
|
||||||
|
show_symbol_type_in_heading: true
|
||||||
show_signature_annotations: true
|
show_signature_annotations: true
|
||||||
show_root_heading: true
|
show_root_heading: true
|
||||||
show_docstring_examples: true
|
|
||||||
show_docstring_attributes: false
|
|
||||||
show_docstring_other_parameters: true
|
|
||||||
show_symbol_type_heading: true
|
|
||||||
show_labels: false
|
|
||||||
show_if_no_docstring: true
|
|
||||||
show_source: false
|
|
||||||
members_order: source
|
members_order: source
|
||||||
docstring_section_style: list
|
docstring_section_style: list
|
||||||
signature_crossrefs: true
|
signature_crossrefs: true
|
||||||
separate_signature: true
|
separate_signature: true
|
||||||
filters:
|
|
||||||
- "!^_"
|
|
||||||
import:
|
import:
|
||||||
# for cross references
|
# for cross references
|
||||||
- https://arrow.apache.org/docs/objects.inv
|
- https://arrow.apache.org/docs/objects.inv
|
||||||
@@ -120,7 +113,7 @@ markdown_extensions:
|
|||||||
emoji_index: !!python/name:material.extensions.emoji.twemoji
|
emoji_index: !!python/name:material.extensions.emoji.twemoji
|
||||||
emoji_generator: !!python/name:material.extensions.emoji.to_svg
|
emoji_generator: !!python/name:material.extensions.emoji.to_svg
|
||||||
- markdown.extensions.toc:
|
- markdown.extensions.toc:
|
||||||
toc_depth: 4
|
toc_depth: 3
|
||||||
permalink: true
|
permalink: true
|
||||||
permalink_title: Anchor link to this section
|
permalink_title: Anchor link to this section
|
||||||
|
|
||||||
|
|||||||
@@ -1,9 +1,9 @@
|
|||||||
mkdocs==1.5.3
|
mkdocs==1.5.3
|
||||||
mkdocs-jupyter==0.24.1
|
mkdocs-jupyter==0.24.1
|
||||||
mkdocs-material==9.5.3
|
mkdocs-material==9.5.3
|
||||||
mkdocs-autorefs>=0.5,<=1.0
|
mkdocs-autorefs<=1.0
|
||||||
mkdocstrings[python]==0.25.2
|
mkdocstrings[python]==0.25.2
|
||||||
griffe>=0.40,<1.0
|
griffe
|
||||||
mkdocs-render-swagger-plugin>=0.1.0
|
mkdocs-render-swagger-plugin
|
||||||
pydantic>=2.0,<3.0
|
pydantic
|
||||||
mkdocs-redirects>=1.2.0
|
mkdocs-redirects
|
||||||
@@ -14,7 +14,7 @@ Add the following dependency to your `pom.xml`:
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>com.lancedb</groupId>
|
<groupId>com.lancedb</groupId>
|
||||||
<artifactId>lancedb-core</artifactId>
|
<artifactId>lancedb-core</artifactId>
|
||||||
<version>0.27.0</version>
|
<version>0.27.0-beta.3</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|||||||
@@ -71,12 +71,11 @@ Add new columns with defined values.
|
|||||||
|
|
||||||
#### Parameters
|
#### Parameters
|
||||||
|
|
||||||
* **newColumnTransforms**: `Field`<`any`> \| `Field`<`any`>[] \| `Schema`<`any`> \| [`AddColumnsSql`](../interfaces/AddColumnsSql.md)[]
|
* **newColumnTransforms**: [`AddColumnsSql`](../interfaces/AddColumnsSql.md)[]
|
||||||
Either:
|
pairs of column names and
|
||||||
- An array of objects with column names and SQL expressions to calculate values
|
the SQL expression to use to calculate the value of the new column. These
|
||||||
- A single Arrow Field defining one column with its data type (column will be initialized with null values)
|
expressions will be evaluated for each row in the table, and can
|
||||||
- An array of Arrow Fields defining columns with their data types (columns will be initialized with null values)
|
reference existing columns in the table.
|
||||||
- An Arrow Schema defining columns with their data types (columns will be initialized with null values)
|
|
||||||
|
|
||||||
#### Returns
|
#### Returns
|
||||||
|
|
||||||
@@ -485,7 +484,19 @@ Modeled after ``VACUUM`` in PostgreSQL.
|
|||||||
- Prune: Removes old versions of the dataset
|
- Prune: Removes old versions of the dataset
|
||||||
- Index: Optimizes the indices, adding new data to existing indices
|
- Index: Optimizes the indices, adding new data to existing indices
|
||||||
|
|
||||||
The frequency an application should call optimize is based on the frequency of
|
Experimental API
|
||||||
|
----------------
|
||||||
|
|
||||||
|
The optimization process is undergoing active development and may change.
|
||||||
|
Our goal with these changes is to improve the performance of optimization and
|
||||||
|
reduce the complexity.
|
||||||
|
|
||||||
|
That being said, it is essential today to run optimize if you want the best
|
||||||
|
performance. It should be stable and safe to use in production, but it our
|
||||||
|
hope that the API may be simplified (or not even need to be called) in the
|
||||||
|
future.
|
||||||
|
|
||||||
|
The frequency an application shoudl call optimize is based on the frequency of
|
||||||
data modifications. If data is frequently added, deleted, or updated then
|
data modifications. If data is frequently added, deleted, or updated then
|
||||||
optimize should be run frequently. A good rule of thumb is to run optimize if
|
optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
you have added or modified 100,000 or more records or run more than 20 data
|
you have added or modified 100,000 or more records or run more than 20 data
|
||||||
|
|||||||
@@ -37,12 +37,3 @@ tbl.optimize({cleanupOlderThan: new Date()});
|
|||||||
```ts
|
```ts
|
||||||
deleteUnverified: boolean;
|
deleteUnverified: boolean;
|
||||||
```
|
```
|
||||||
|
|
||||||
Because they may be part of an in-progress transaction, files newer than
|
|
||||||
7 days old are not deleted by default. If you are sure that there are no
|
|
||||||
in-progress transactions, then you can set this to true to delete all
|
|
||||||
files older than `cleanupOlderThan`.
|
|
||||||
|
|
||||||
**WARNING**: This should only be set to true if you can guarantee that
|
|
||||||
no other process is currently working on this dataset. Otherwise the
|
|
||||||
dataset could be put into a corrupted state.
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# LanceDB Java Enterprise Client
|
# LanceDB Java SDK
|
||||||
|
|
||||||
## Configuration and Initialization
|
## Configuration and Initialization
|
||||||
|
|
||||||
|
|||||||
@@ -8,7 +8,7 @@
|
|||||||
<parent>
|
<parent>
|
||||||
<groupId>com.lancedb</groupId>
|
<groupId>com.lancedb</groupId>
|
||||||
<artifactId>lancedb-parent</artifactId>
|
<artifactId>lancedb-parent</artifactId>
|
||||||
<version>0.27.0-final.0</version>
|
<version>0.27.0-beta.3</version>
|
||||||
<relativePath>../pom.xml</relativePath>
|
<relativePath>../pom.xml</relativePath>
|
||||||
</parent>
|
</parent>
|
||||||
|
|
||||||
@@ -56,21 +56,21 @@
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.logging.log4j</groupId>
|
<groupId>org.apache.logging.log4j</groupId>
|
||||||
<artifactId>log4j-slf4j2-impl</artifactId>
|
<artifactId>log4j-slf4j2-impl</artifactId>
|
||||||
<version>2.25.3</version>
|
<version>2.24.3</version>
|
||||||
<scope>test</scope>
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.logging.log4j</groupId>
|
<groupId>org.apache.logging.log4j</groupId>
|
||||||
<artifactId>log4j-core</artifactId>
|
<artifactId>log4j-core</artifactId>
|
||||||
<version>2.25.3</version>
|
<version>2.24.3</version>
|
||||||
<scope>test</scope>
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.logging.log4j</groupId>
|
<groupId>org.apache.logging.log4j</groupId>
|
||||||
<artifactId>log4j-api</artifactId>
|
<artifactId>log4j-api</artifactId>
|
||||||
<version>2.25.3</version>
|
<version>2.24.3</version>
|
||||||
<scope>test</scope>
|
<scope>test</scope>
|
||||||
</dependency>
|
</dependency>
|
||||||
</dependencies>
|
</dependencies>
|
||||||
|
|||||||
18
java/pom.xml
18
java/pom.xml
@@ -6,7 +6,7 @@
|
|||||||
|
|
||||||
<groupId>com.lancedb</groupId>
|
<groupId>com.lancedb</groupId>
|
||||||
<artifactId>lancedb-parent</artifactId>
|
<artifactId>lancedb-parent</artifactId>
|
||||||
<version>0.27.0-final.0</version>
|
<version>0.27.0-beta.3</version>
|
||||||
<packaging>pom</packaging>
|
<packaging>pom</packaging>
|
||||||
<name>${project.artifactId}</name>
|
<name>${project.artifactId}</name>
|
||||||
<description>LanceDB Java SDK Parent POM</description>
|
<description>LanceDB Java SDK Parent POM</description>
|
||||||
@@ -111,7 +111,7 @@
|
|||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-source-plugin</artifactId>
|
<artifactId>maven-source-plugin</artifactId>
|
||||||
<version>3.3.1</version>
|
<version>2.2.1</version>
|
||||||
<executions>
|
<executions>
|
||||||
<execution>
|
<execution>
|
||||||
<id>attach-sources</id>
|
<id>attach-sources</id>
|
||||||
@@ -124,7 +124,7 @@
|
|||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-javadoc-plugin</artifactId>
|
<artifactId>maven-javadoc-plugin</artifactId>
|
||||||
<version>3.11.2</version>
|
<version>2.9.1</version>
|
||||||
<executions>
|
<executions>
|
||||||
<execution>
|
<execution>
|
||||||
<id>attach-javadocs</id>
|
<id>attach-javadocs</id>
|
||||||
@@ -178,15 +178,15 @@
|
|||||||
<plugins>
|
<plugins>
|
||||||
<plugin>
|
<plugin>
|
||||||
<artifactId>maven-clean-plugin</artifactId>
|
<artifactId>maven-clean-plugin</artifactId>
|
||||||
<version>3.4.1</version>
|
<version>3.1.0</version>
|
||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<artifactId>maven-resources-plugin</artifactId>
|
<artifactId>maven-resources-plugin</artifactId>
|
||||||
<version>3.3.1</version>
|
<version>3.0.2</version>
|
||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<artifactId>maven-compiler-plugin</artifactId>
|
<artifactId>maven-compiler-plugin</artifactId>
|
||||||
<version>3.14.0</version>
|
<version>3.8.1</version>
|
||||||
<configuration>
|
<configuration>
|
||||||
<compilerArgs>
|
<compilerArgs>
|
||||||
<arg>-h</arg>
|
<arg>-h</arg>
|
||||||
@@ -205,11 +205,11 @@
|
|||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<artifactId>maven-jar-plugin</artifactId>
|
<artifactId>maven-jar-plugin</artifactId>
|
||||||
<version>3.4.2</version>
|
<version>3.0.2</version>
|
||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<artifactId>maven-install-plugin</artifactId>
|
<artifactId>maven-install-plugin</artifactId>
|
||||||
<version>3.1.3</version>
|
<version>2.5.2</version>
|
||||||
</plugin>
|
</plugin>
|
||||||
<plugin>
|
<plugin>
|
||||||
<groupId>com.diffplug.spotless</groupId>
|
<groupId>com.diffplug.spotless</groupId>
|
||||||
@@ -327,7 +327,7 @@
|
|||||||
<plugin>
|
<plugin>
|
||||||
<groupId>org.apache.maven.plugins</groupId>
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
<artifactId>maven-gpg-plugin</artifactId>
|
<artifactId>maven-gpg-plugin</artifactId>
|
||||||
<version>3.2.7</version>
|
<version>1.5</version>
|
||||||
<executions>
|
<executions>
|
||||||
<execution>
|
<execution>
|
||||||
<id>sign-artifacts</id>
|
<id>sign-artifacts</id>
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "lancedb-nodejs"
|
name = "lancedb-nodejs"
|
||||||
edition.workspace = true
|
edition.workspace = true
|
||||||
version = "0.27.0"
|
version = "0.27.0-beta.3"
|
||||||
license.workspace = true
|
license.workspace = true
|
||||||
description.workspace = true
|
description.workspace = true
|
||||||
repository.workspace = true
|
repository.workspace = true
|
||||||
@@ -25,12 +25,12 @@ napi = { version = "3.8.3", default-features = false, features = [
|
|||||||
] }
|
] }
|
||||||
napi-derive = "3.5.2"
|
napi-derive = "3.5.2"
|
||||||
# Prevent dynamic linking of lzma, which comes from datafusion
|
# Prevent dynamic linking of lzma, which comes from datafusion
|
||||||
lzma-sys = { version = "0.1", features = ["static"] }
|
lzma-sys = { version = "*", features = ["static"] }
|
||||||
log.workspace = true
|
log.workspace = true
|
||||||
|
|
||||||
# Pin to resolve build failures; update periodically for security patches.
|
# Workaround for build failure until we can fix it.
|
||||||
aws-lc-sys = "=0.38.0"
|
aws-lc-sys = "=0.28.0"
|
||||||
aws-lc-rs = "=1.16.1"
|
aws-lc-rs = "=1.13.0"
|
||||||
|
|
||||||
[build-dependencies]
|
[build-dependencies]
|
||||||
napi-build = "2.3.1"
|
napi-build = "2.3.1"
|
||||||
|
|||||||
@@ -63,7 +63,6 @@ describe.each([arrow15, arrow16, arrow17, arrow18])(
|
|||||||
tableFromIPC,
|
tableFromIPC,
|
||||||
DataType,
|
DataType,
|
||||||
Dictionary,
|
Dictionary,
|
||||||
Uint8: ArrowUint8,
|
|
||||||
// biome-ignore lint/suspicious/noExplicitAny: <explanation>
|
// biome-ignore lint/suspicious/noExplicitAny: <explanation>
|
||||||
} = <any>arrow;
|
} = <any>arrow;
|
||||||
type Schema = ApacheArrow["Schema"];
|
type Schema = ApacheArrow["Schema"];
|
||||||
@@ -363,38 +362,6 @@ describe.each([arrow15, arrow16, arrow17, arrow18])(
|
|||||||
).toEqual(new Float64().toString());
|
).toEqual(new Float64().toString());
|
||||||
});
|
});
|
||||||
|
|
||||||
it("will infer FixedSizeList<Float32> from Float32Array values", async function () {
|
|
||||||
const table = makeArrowTable([
|
|
||||||
{ id: "a", vector: new Float32Array([0.1, 0.2, 0.3]) },
|
|
||||||
{ id: "b", vector: new Float32Array([0.4, 0.5, 0.6]) },
|
|
||||||
]);
|
|
||||||
|
|
||||||
expect(DataType.isFixedSizeList(table.getChild("vector")?.type)).toBe(
|
|
||||||
true,
|
|
||||||
);
|
|
||||||
const vectorType = table.getChild("vector")?.type;
|
|
||||||
expect(vectorType.listSize).toBe(3);
|
|
||||||
expect(vectorType.children[0].type.toString()).toEqual(
|
|
||||||
new Float32().toString(),
|
|
||||||
);
|
|
||||||
});
|
|
||||||
|
|
||||||
it("will infer FixedSizeList<Uint8> from Uint8Array values", async function () {
|
|
||||||
const table = makeArrowTable([
|
|
||||||
{ id: "a", vector: new Uint8Array([1, 2, 3]) },
|
|
||||||
{ id: "b", vector: new Uint8Array([4, 5, 6]) },
|
|
||||||
]);
|
|
||||||
|
|
||||||
expect(DataType.isFixedSizeList(table.getChild("vector")?.type)).toBe(
|
|
||||||
true,
|
|
||||||
);
|
|
||||||
const vectorType = table.getChild("vector")?.type;
|
|
||||||
expect(vectorType.listSize).toBe(3);
|
|
||||||
expect(vectorType.children[0].type.toString()).toEqual(
|
|
||||||
new ArrowUint8().toString(),
|
|
||||||
);
|
|
||||||
});
|
|
||||||
|
|
||||||
it("will use dictionary encoded strings if asked", async function () {
|
it("will use dictionary encoded strings if asked", async function () {
|
||||||
const table = makeArrowTable([{ str: "hello" }]);
|
const table = makeArrowTable([{ str: "hello" }]);
|
||||||
expect(DataType.isUtf8(table.getChild("str")?.type)).toBe(true);
|
expect(DataType.isUtf8(table.getChild("str")?.type)).toBe(true);
|
||||||
|
|||||||
@@ -1259,98 +1259,6 @@ describe("schema evolution", function () {
|
|||||||
expect(await table.schema()).toEqual(expectedSchema);
|
expect(await table.schema()).toEqual(expectedSchema);
|
||||||
});
|
});
|
||||||
|
|
||||||
it("can add columns with schema for explicit data types", async function () {
|
|
||||||
const con = await connect(tmpDir.name);
|
|
||||||
const table = await con.createTable("vectors", [
|
|
||||||
{ id: 1n, vector: [0.1, 0.2] },
|
|
||||||
]);
|
|
||||||
|
|
||||||
// Define schema for new columns with explicit data types
|
|
||||||
// Note: All columns must be nullable when using addColumns with Schema
|
|
||||||
// because they are initially populated with null values
|
|
||||||
const newColumnsSchema = new Schema([
|
|
||||||
new Field("price", new Float64(), true),
|
|
||||||
new Field("category", new Utf8(), true),
|
|
||||||
new Field("rating", new Int32(), true),
|
|
||||||
]);
|
|
||||||
|
|
||||||
const result = await table.addColumns(newColumnsSchema);
|
|
||||||
expect(result).toHaveProperty("version");
|
|
||||||
expect(result.version).toBe(2);
|
|
||||||
|
|
||||||
const expectedSchema = new Schema([
|
|
||||||
new Field("id", new Int64(), true),
|
|
||||||
new Field(
|
|
||||||
"vector",
|
|
||||||
new FixedSizeList(2, new Field("item", new Float32(), true)),
|
|
||||||
true,
|
|
||||||
),
|
|
||||||
new Field("price", new Float64(), true),
|
|
||||||
new Field("category", new Utf8(), true),
|
|
||||||
new Field("rating", new Int32(), true),
|
|
||||||
]);
|
|
||||||
expect(await table.schema()).toEqual(expectedSchema);
|
|
||||||
|
|
||||||
// Verify that new columns are populated with null values
|
|
||||||
const results = await table.query().toArray();
|
|
||||||
expect(results).toHaveLength(1);
|
|
||||||
expect(results[0].price).toBeNull();
|
|
||||||
expect(results[0].category).toBeNull();
|
|
||||||
expect(results[0].rating).toBeNull();
|
|
||||||
});
|
|
||||||
|
|
||||||
it("can add a single column using Field", async function () {
|
|
||||||
const con = await connect(tmpDir.name);
|
|
||||||
const table = await con.createTable("vectors", [
|
|
||||||
{ id: 1n, vector: [0.1, 0.2] },
|
|
||||||
]);
|
|
||||||
|
|
||||||
// Add a single field
|
|
||||||
const priceField = new Field("price", new Float64(), true);
|
|
||||||
const result = await table.addColumns(priceField);
|
|
||||||
expect(result).toHaveProperty("version");
|
|
||||||
expect(result.version).toBe(2);
|
|
||||||
|
|
||||||
const expectedSchema = new Schema([
|
|
||||||
new Field("id", new Int64(), true),
|
|
||||||
new Field(
|
|
||||||
"vector",
|
|
||||||
new FixedSizeList(2, new Field("item", new Float32(), true)),
|
|
||||||
true,
|
|
||||||
),
|
|
||||||
new Field("price", new Float64(), true),
|
|
||||||
]);
|
|
||||||
expect(await table.schema()).toEqual(expectedSchema);
|
|
||||||
});
|
|
||||||
|
|
||||||
it("can add multiple columns using array of Fields", async function () {
|
|
||||||
const con = await connect(tmpDir.name);
|
|
||||||
const table = await con.createTable("vectors", [
|
|
||||||
{ id: 1n, vector: [0.1, 0.2] },
|
|
||||||
]);
|
|
||||||
|
|
||||||
// Add multiple fields as array
|
|
||||||
const fields = [
|
|
||||||
new Field("price", new Float64(), true),
|
|
||||||
new Field("category", new Utf8(), true),
|
|
||||||
];
|
|
||||||
const result = await table.addColumns(fields);
|
|
||||||
expect(result).toHaveProperty("version");
|
|
||||||
expect(result.version).toBe(2);
|
|
||||||
|
|
||||||
const expectedSchema = new Schema([
|
|
||||||
new Field("id", new Int64(), true),
|
|
||||||
new Field(
|
|
||||||
"vector",
|
|
||||||
new FixedSizeList(2, new Field("item", new Float32(), true)),
|
|
||||||
true,
|
|
||||||
),
|
|
||||||
new Field("price", new Float64(), true),
|
|
||||||
new Field("category", new Utf8(), true),
|
|
||||||
]);
|
|
||||||
expect(await table.schema()).toEqual(expectedSchema);
|
|
||||||
});
|
|
||||||
|
|
||||||
it("can alter the columns in the schema", async function () {
|
it("can alter the columns in the schema", async function () {
|
||||||
const con = await connect(tmpDir.name);
|
const con = await connect(tmpDir.name);
|
||||||
const schema = new Schema([
|
const schema = new Schema([
|
||||||
@@ -2296,36 +2204,3 @@ describe("when creating an empty table", () => {
|
|||||||
expect((actualSchema.fields[1].type as Float64).precision).toBe(2);
|
expect((actualSchema.fields[1].type as Float64).precision).toBe(2);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
// Ensure we can create float32 arrays without using Arrow
|
|
||||||
// by utilizing native JS TypedArray support
|
|
||||||
//
|
|
||||||
// https://github.com/lancedb/lancedb/issues/3115
|
|
||||||
describe("when creating a table with Float32Array vectors", () => {
|
|
||||||
let tmpDir: tmp.DirResult;
|
|
||||||
beforeEach(() => {
|
|
||||||
tmpDir = tmp.dirSync({ unsafeCleanup: true });
|
|
||||||
});
|
|
||||||
afterEach(() => {
|
|
||||||
tmpDir.removeCallback();
|
|
||||||
});
|
|
||||||
|
|
||||||
it("should persist Float32Array as FixedSizeList<Float32> in the LanceDB schema", async () => {
|
|
||||||
const db = await connect(tmpDir.name);
|
|
||||||
const table = await db.createTable("test", [
|
|
||||||
{ id: "a", vector: new Float32Array([0.1, 0.2, 0.3]) },
|
|
||||||
{ id: "b", vector: new Float32Array([0.4, 0.5, 0.6]) },
|
|
||||||
]);
|
|
||||||
|
|
||||||
const schema = await table.schema();
|
|
||||||
const vectorField = schema.fields.find((f) => f.name === "vector");
|
|
||||||
expect(vectorField).toBeDefined();
|
|
||||||
expect(vectorField!.type).toBeInstanceOf(FixedSizeList);
|
|
||||||
|
|
||||||
const fsl = vectorField!.type as FixedSizeList;
|
|
||||||
expect(fsl.listSize).toBe(3);
|
|
||||||
expect(fsl.children[0].type.typeId).toBe(Type.Float);
|
|
||||||
// precision: HALF=0, SINGLE=1, DOUBLE=2
|
|
||||||
expect((fsl.children[0].type as Float32).precision).toBe(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|||||||
599
nodejs/examples/package-lock.json
generated
599
nodejs/examples/package-lock.json
generated
@@ -30,15 +30,12 @@
|
|||||||
"x64",
|
"x64",
|
||||||
"arm64"
|
"arm64"
|
||||||
],
|
],
|
||||||
"dev": true,
|
|
||||||
"license": "Apache-2.0",
|
"license": "Apache-2.0",
|
||||||
"optional": true,
|
|
||||||
"os": [
|
"os": [
|
||||||
"darwin",
|
"darwin",
|
||||||
"linux",
|
"linux",
|
||||||
"win32"
|
"win32"
|
||||||
],
|
],
|
||||||
"peer": true,
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"reflect-metadata": "^0.2.2"
|
"reflect-metadata": "^0.2.2"
|
||||||
},
|
},
|
||||||
@@ -94,15 +91,14 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/code-frame": {
|
"node_modules/@babel/code-frame": {
|
||||||
"version": "7.29.0",
|
"version": "7.26.2",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/code-frame/-/code-frame-7.29.0.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/code-frame/-/code-frame-7.26.2.tgz",
|
||||||
"integrity": "sha512-9NhCeYjq9+3uxgdtp20LSiJXJvN0FeCtNGpJxuMFZ1Kv3cWUNb6DOhJwUvcVCzKGR66cw4njwM6hrJLqgOwbcw==",
|
"integrity": "sha512-RJlIHRueQgwWitWgF8OdFYGZX328Ax5BCemNGlqHfplnRT9ESi8JkFlvaVYbS+UubVY6dpv87Fs2u5M29iNFVQ==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@babel/helper-validator-identifier": "^7.28.5",
|
"@babel/helper-validator-identifier": "^7.25.9",
|
||||||
"js-tokens": "^4.0.0",
|
"js-tokens": "^4.0.0",
|
||||||
"picocolors": "^1.1.1"
|
"picocolors": "^1.0.0"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
@@ -237,21 +233,19 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/helper-string-parser": {
|
"node_modules/@babel/helper-string-parser": {
|
||||||
"version": "7.27.1",
|
"version": "7.25.9",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/helper-string-parser/-/helper-string-parser-7.27.1.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/helper-string-parser/-/helper-string-parser-7.25.9.tgz",
|
||||||
"integrity": "sha512-qMlSxKbpRlAridDExk92nSobyDdpPijUq2DW6oDnUqd0iOGxmQjyqhMIihI9+zv4LPyZdRje2cavWPbCbWm3eA==",
|
"integrity": "sha512-4A/SCr/2KLd5jrtOMFzaKjVtAei3+2r/NChoBNoZ3EyP/+GlhoaEGoWOZUmFmoITP7zOJyHIMm+DYRd8o3PvHA==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/helper-validator-identifier": {
|
"node_modules/@babel/helper-validator-identifier": {
|
||||||
"version": "7.28.5",
|
"version": "7.25.9",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/helper-validator-identifier/-/helper-validator-identifier-7.28.5.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/helper-validator-identifier/-/helper-validator-identifier-7.25.9.tgz",
|
||||||
"integrity": "sha512-qSs4ifwzKJSV39ucNjsvc6WVHs6b7S03sOh2OcHF9UHfVPqWWALUsNUVzhSBiItjRZoLHx7nIarVjqKVusUZ1Q==",
|
"integrity": "sha512-Ed61U6XJc3CVRfkERJWDz4dJwKe7iLmmJsbOGu9wSloNSFttHV0I8g6UAgb7qnK5ly5bGLPd4oXZlxCdANBOWQ==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
}
|
}
|
||||||
@@ -266,27 +260,25 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/helpers": {
|
"node_modules/@babel/helpers": {
|
||||||
"version": "7.28.6",
|
"version": "7.26.0",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/helpers/-/helpers-7.28.6.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/helpers/-/helpers-7.26.0.tgz",
|
||||||
"integrity": "sha512-xOBvwq86HHdB7WUDTfKfT/Vuxh7gElQ+Sfti2Cy6yIWNW05P8iUslOVcZ4/sKbE+/jQaukQAdz/gf3724kYdqw==",
|
"integrity": "sha512-tbhNuIxNcVb21pInl3ZSjksLCvgdZy9KwJ8brv993QtIVKJBBkYXz4q4ZbAv31GdnC+R90np23L5FbEBlthAEw==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@babel/template": "^7.28.6",
|
"@babel/template": "^7.25.9",
|
||||||
"@babel/types": "^7.28.6"
|
"@babel/types": "^7.26.0"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/parser": {
|
"node_modules/@babel/parser": {
|
||||||
"version": "7.29.0",
|
"version": "7.26.2",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/parser/-/parser-7.29.0.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/parser/-/parser-7.26.2.tgz",
|
||||||
"integrity": "sha512-IyDgFV5GeDUVX4YdF/3CPULtVGSXXMLh1xVIgdCgxApktqnQV0r7/8Nqthg+8YLGaAtdyIlo2qIdZrbCv4+7ww==",
|
"integrity": "sha512-DWMCZH9WA4Maitz2q21SRKHo9QXZxkDsbNZoVD62gusNtNBBqDg9i7uOhASfTfIGNzW+O+r7+jAlM8dwphcJKQ==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@babel/types": "^7.29.0"
|
"@babel/types": "^7.26.0"
|
||||||
},
|
},
|
||||||
"bin": {
|
"bin": {
|
||||||
"parser": "bin/babel-parser.js"
|
"parser": "bin/babel-parser.js"
|
||||||
@@ -518,15 +510,14 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/template": {
|
"node_modules/@babel/template": {
|
||||||
"version": "7.28.6",
|
"version": "7.25.9",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/template/-/template-7.28.6.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/template/-/template-7.25.9.tgz",
|
||||||
"integrity": "sha512-YA6Ma2KsCdGb+WC6UpBVFJGXL58MDA6oyONbjyF/+5sBgxY/dwkhLogbMT2GXXyU84/IhRw/2D1Os1B/giz+BQ==",
|
"integrity": "sha512-9DGttpmPvIxBb/2uwpVo3dqJ+O6RooAFOS+lB+xDqoE2PVCE8nfoHMdZLpfCQRLwvohzXISPZcgxt80xLfsuwg==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@babel/code-frame": "^7.28.6",
|
"@babel/code-frame": "^7.25.9",
|
||||||
"@babel/parser": "^7.28.6",
|
"@babel/parser": "^7.25.9",
|
||||||
"@babel/types": "^7.28.6"
|
"@babel/types": "^7.25.9"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
@@ -551,14 +542,13 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@babel/types": {
|
"node_modules/@babel/types": {
|
||||||
"version": "7.29.0",
|
"version": "7.26.0",
|
||||||
"resolved": "https://registry.npmjs.org/@babel/types/-/types-7.29.0.tgz",
|
"resolved": "https://registry.npmjs.org/@babel/types/-/types-7.26.0.tgz",
|
||||||
"integrity": "sha512-LwdZHpScM4Qz8Xw2iKSzS+cfglZzJGvofQICy7W7v4caru4EaAmyUuO6BGrbyQ2mYV11W0U8j5mBhd14dd3B0A==",
|
"integrity": "sha512-Z/yiTPj+lDVnF7lWeKCIJzaIkI0vYO87dMpZ4bg4TDrFe4XXLFWL1TbXU27gBP3QccxV9mZICCrnjnYlJjXHOA==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@babel/helper-string-parser": "^7.27.1",
|
"@babel/helper-string-parser": "^7.25.9",
|
||||||
"@babel/helper-validator-identifier": "^7.28.5"
|
"@babel/helper-validator-identifier": "^7.25.9"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=6.9.0"
|
"node": ">=6.9.0"
|
||||||
@@ -1161,6 +1151,95 @@
|
|||||||
"url": "https://opencollective.com/libvips"
|
"url": "https://opencollective.com/libvips"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/@isaacs/cliui": {
|
||||||
|
"version": "8.0.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/@isaacs/cliui/-/cliui-8.0.2.tgz",
|
||||||
|
"integrity": "sha512-O8jcjabXaleOG9DQ0+ARXWZBTfnP4WNAqzuiJK7ll44AmxGKv/J2M4TPjxjY3znBCfvBXFzucm1twdyFybFqEA==",
|
||||||
|
"dependencies": {
|
||||||
|
"string-width": "^5.1.2",
|
||||||
|
"string-width-cjs": "npm:string-width@^4.2.0",
|
||||||
|
"strip-ansi": "^7.0.1",
|
||||||
|
"strip-ansi-cjs": "npm:strip-ansi@^6.0.1",
|
||||||
|
"wrap-ansi": "^8.1.0",
|
||||||
|
"wrap-ansi-cjs": "npm:wrap-ansi@^7.0.0"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/ansi-regex": {
|
||||||
|
"version": "6.1.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.1.0.tgz",
|
||||||
|
"integrity": "sha512-7HSX4QQb4CspciLpVFwyRe79O3xsIZDDLER21kERQ71oaPodF8jL725AgJMFAYbooIqolJoRLuM81SpeUkpkvA==",
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/chalk/ansi-regex?sponsor=1"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/ansi-styles": {
|
||||||
|
"version": "6.2.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz",
|
||||||
|
"integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==",
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/chalk/ansi-styles?sponsor=1"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/emoji-regex": {
|
||||||
|
"version": "9.2.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz",
|
||||||
|
"integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg=="
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/string-width": {
|
||||||
|
"version": "5.1.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz",
|
||||||
|
"integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==",
|
||||||
|
"dependencies": {
|
||||||
|
"eastasianwidth": "^0.2.0",
|
||||||
|
"emoji-regex": "^9.2.2",
|
||||||
|
"strip-ansi": "^7.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/sindresorhus"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/strip-ansi": {
|
||||||
|
"version": "7.1.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.1.0.tgz",
|
||||||
|
"integrity": "sha512-iq6eVVI64nQQTRYq2KtEg2d2uU7LElhTJwsH4YzIHZshxlgZms/wIc4VoDQTlG/IvVIrBKG06CrZnp0qv7hkcQ==",
|
||||||
|
"dependencies": {
|
||||||
|
"ansi-regex": "^6.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/chalk/strip-ansi?sponsor=1"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@isaacs/cliui/node_modules/wrap-ansi": {
|
||||||
|
"version": "8.1.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz",
|
||||||
|
"integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==",
|
||||||
|
"dependencies": {
|
||||||
|
"ansi-styles": "^6.1.0",
|
||||||
|
"string-width": "^5.0.1",
|
||||||
|
"strip-ansi": "^7.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=12"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/chalk/wrap-ansi?sponsor=1"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/@isaacs/fs-minipass": {
|
"node_modules/@isaacs/fs-minipass": {
|
||||||
"version": "4.0.1",
|
"version": "4.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz",
|
"resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz",
|
||||||
@@ -1527,6 +1606,15 @@
|
|||||||
"resolved": "../dist",
|
"resolved": "../dist",
|
||||||
"link": true
|
"link": true
|
||||||
},
|
},
|
||||||
|
"node_modules/@pkgjs/parseargs": {
|
||||||
|
"version": "0.11.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/@pkgjs/parseargs/-/parseargs-0.11.0.tgz",
|
||||||
|
"integrity": "sha512-+1VkjdD0QBLPodGrJUeqarH8VAIvQODIbwh9XpP5Syisf7YoQgsJKPNFoqqLQlu+VQ/tVSshMR6loPMn8U+dPg==",
|
||||||
|
"optional": true,
|
||||||
|
"engines": {
|
||||||
|
"node": ">=14"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/@protobufjs/aspromise": {
|
"node_modules/@protobufjs/aspromise": {
|
||||||
"version": "1.1.2",
|
"version": "1.1.2",
|
||||||
"resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz",
|
"resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz",
|
||||||
@@ -1758,7 +1846,6 @@
|
|||||||
"version": "5.0.1",
|
"version": "5.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz",
|
"resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz",
|
||||||
"integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==",
|
"integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==",
|
||||||
"dev": true,
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
@@ -1767,7 +1854,6 @@
|
|||||||
"version": "4.3.0",
|
"version": "4.3.0",
|
||||||
"resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz",
|
"resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz",
|
||||||
"integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==",
|
"integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==",
|
||||||
"dev": true,
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"color-convert": "^2.0.1"
|
"color-convert": "^2.0.1"
|
||||||
},
|
},
|
||||||
@@ -1933,15 +2019,13 @@
|
|||||||
"node_modules/balanced-match": {
|
"node_modules/balanced-match": {
|
||||||
"version": "1.0.2",
|
"version": "1.0.2",
|
||||||
"resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz",
|
"resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz",
|
||||||
"integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw==",
|
"integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw=="
|
||||||
"dev": true
|
|
||||||
},
|
},
|
||||||
"node_modules/brace-expansion": {
|
"node_modules/brace-expansion": {
|
||||||
"version": "1.1.12",
|
"version": "1.1.11",
|
||||||
"resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-1.1.12.tgz",
|
"resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-1.1.11.tgz",
|
||||||
"integrity": "sha512-9T9UjW3r0UW5c1Q7GTwllptXwhvYmEzFhzMfZ9H7FQWt+uZePjZPjBP/W1ZEyZ1twGWom5/56TF4lPcqjnDHcg==",
|
"integrity": "sha512-iCuPHDFgrHX7H2vEI/5xpz07zSHB00TpugqhmYtVmMO6518mCuRMoOYFldEBl0g187ufozdaHgWKcYFb61qGiA==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"balanced-match": "^1.0.0",
|
"balanced-match": "^1.0.0",
|
||||||
"concat-map": "0.0.1"
|
"concat-map": "0.0.1"
|
||||||
@@ -2018,19 +2102,6 @@
|
|||||||
"integrity": "sha512-E+XQCRwSbaaiChtv6k6Dwgc+bx+Bs6vuKJHHl5kox/BaKbhiXzqQOwK4cO22yElGp2OCmjwVhT3HmxgyPGnJfQ==",
|
"integrity": "sha512-E+XQCRwSbaaiChtv6k6Dwgc+bx+Bs6vuKJHHl5kox/BaKbhiXzqQOwK4cO22yElGp2OCmjwVhT3HmxgyPGnJfQ==",
|
||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
"node_modules/call-bind-apply-helpers": {
|
|
||||||
"version": "1.0.2",
|
|
||||||
"resolved": "https://registry.npmjs.org/call-bind-apply-helpers/-/call-bind-apply-helpers-1.0.2.tgz",
|
|
||||||
"integrity": "sha512-Sp1ablJ0ivDkSzjcaJdxEunN5/XvksFJ2sMBFfq6x0ryhQV/2b/KwFe21cMpmHtPOSij8K99/wSfoEuTObmuMQ==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"es-errors": "^1.3.0",
|
|
||||||
"function-bind": "^1.1.2"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/callsites": {
|
"node_modules/callsites": {
|
||||||
"version": "3.1.0",
|
"version": "3.1.0",
|
||||||
"resolved": "https://registry.npmjs.org/callsites/-/callsites-3.1.0.tgz",
|
"resolved": "https://registry.npmjs.org/callsites/-/callsites-3.1.0.tgz",
|
||||||
@@ -2227,11 +2298,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/cross-spawn": {
|
"node_modules/cross-spawn": {
|
||||||
"version": "7.0.6",
|
"version": "7.0.3",
|
||||||
"resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.6.tgz",
|
"resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.3.tgz",
|
||||||
"integrity": "sha512-uV2QOWP2nWzsy2aMp8aRibhi9dlzF5Hgh5SHaB9OiTGEyDTiJJyx0uy51QXdyWbtAHNua4XJzUKca3OzKUd3vA==",
|
"integrity": "sha512-iRDPJKUPVEND7dHPO8rkbOnPpyDygcDFtWjpeWNCgy8WP2rXcxXL8TskReQl6OrB2G7+UJrags1q15Fudc7G6w==",
|
||||||
"dev": true,
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"path-key": "^3.1.0",
|
"path-key": "^3.1.0",
|
||||||
"shebang-command": "^2.0.0",
|
"shebang-command": "^2.0.0",
|
||||||
@@ -2315,19 +2384,10 @@
|
|||||||
"node": "^14.15.0 || ^16.10.0 || >=18.0.0"
|
"node": "^14.15.0 || ^16.10.0 || >=18.0.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/dunder-proto": {
|
"node_modules/eastasianwidth": {
|
||||||
"version": "1.0.1",
|
"version": "0.2.0",
|
||||||
"resolved": "https://registry.npmjs.org/dunder-proto/-/dunder-proto-1.0.1.tgz",
|
"resolved": "https://registry.npmjs.org/eastasianwidth/-/eastasianwidth-0.2.0.tgz",
|
||||||
"integrity": "sha512-KIN/nDJBQRcXw0MLVhZE9iQHmG68qAVIBg9CqmUYjmQIhgij9U5MFvrqkUL5FbtyyzZuOeOt0zdeRe4UY7ct+A==",
|
"integrity": "sha512-I88TYZWc9XiYHRQ4/3c5rjjfgkjhLyW2luGIheGERbNQ6OY7yTybanSpDXZa8y7VUP9YmDcYa+eyq4ca7iLqWA=="
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"call-bind-apply-helpers": "^1.0.1",
|
|
||||||
"es-errors": "^1.3.0",
|
|
||||||
"gopd": "^1.2.0"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"node_modules/ejs": {
|
"node_modules/ejs": {
|
||||||
"version": "3.1.10",
|
"version": "3.1.10",
|
||||||
@@ -2365,8 +2425,7 @@
|
|||||||
"node_modules/emoji-regex": {
|
"node_modules/emoji-regex": {
|
||||||
"version": "8.0.0",
|
"version": "8.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz",
|
||||||
"integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==",
|
"integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A=="
|
||||||
"dev": true
|
|
||||||
},
|
},
|
||||||
"node_modules/error-ex": {
|
"node_modules/error-ex": {
|
||||||
"version": "1.3.2",
|
"version": "1.3.2",
|
||||||
@@ -2383,51 +2442,6 @@
|
|||||||
"integrity": "sha512-zz06S8t0ozoDXMG+ube26zeCTNXcKIPJZJi8hBrF4idCLms4CG9QtK7qBl1boi5ODzFpjswb5JPmHCbMpjaYzg==",
|
"integrity": "sha512-zz06S8t0ozoDXMG+ube26zeCTNXcKIPJZJi8hBrF4idCLms4CG9QtK7qBl1boi5ODzFpjswb5JPmHCbMpjaYzg==",
|
||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
"node_modules/es-define-property": {
|
|
||||||
"version": "1.0.1",
|
|
||||||
"resolved": "https://registry.npmjs.org/es-define-property/-/es-define-property-1.0.1.tgz",
|
|
||||||
"integrity": "sha512-e3nRfgfUZ4rNGL232gUgX06QNyyez04KdjFrF+LTRoOXmrOgFKDg4BCdsjW8EnT69eqdYGmRpJwiPVYNrCaW3g==",
|
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/es-errors": {
|
|
||||||
"version": "1.3.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/es-errors/-/es-errors-1.3.0.tgz",
|
|
||||||
"integrity": "sha512-Zf5H2Kxt2xjTvbJvP2ZWLEICxA6j+hAmMzIlypy4xcBg1vKVnx89Wy0GbS+kf5cwCVFFzdCFh2XSCFNULS6csw==",
|
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/es-object-atoms": {
|
|
||||||
"version": "1.1.1",
|
|
||||||
"resolved": "https://registry.npmjs.org/es-object-atoms/-/es-object-atoms-1.1.1.tgz",
|
|
||||||
"integrity": "sha512-FGgH2h8zKNim9ljj7dankFPcICIK9Cp5bm+c2gQSYePhpaG5+esrLODihIorn+Pe6FGJzWhXQotPv73jTaldXA==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"es-errors": "^1.3.0"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/es-set-tostringtag": {
|
|
||||||
"version": "2.1.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/es-set-tostringtag/-/es-set-tostringtag-2.1.0.tgz",
|
|
||||||
"integrity": "sha512-j6vWzfrGVfyXxge+O0x5sh6cvxAog0a/4Rdd2K36zCMV5eJ+/+tOAngRO8cODMNWbVRdVlmGZQL2YS3yR8bIUA==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"es-errors": "^1.3.0",
|
|
||||||
"get-intrinsic": "^1.2.6",
|
|
||||||
"has-tostringtag": "^1.0.2",
|
|
||||||
"hasown": "^2.0.2"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/escalade": {
|
"node_modules/escalade": {
|
||||||
"version": "3.2.0",
|
"version": "3.2.0",
|
||||||
"resolved": "https://registry.npmjs.org/escalade/-/escalade-3.2.0.tgz",
|
"resolved": "https://registry.npmjs.org/escalade/-/escalade-3.2.0.tgz",
|
||||||
@@ -2540,21 +2554,19 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/filelist/node_modules/brace-expansion": {
|
"node_modules/filelist/node_modules/brace-expansion": {
|
||||||
"version": "2.0.2",
|
"version": "2.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.2.tgz",
|
"resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.1.tgz",
|
||||||
"integrity": "sha512-Jt0vHyM+jmUBqojB7E1NIYadt0vI0Qxjxd2TErW94wDz+E2LAm5vKMXXwg6ZZBTHPuUlDgQHKXvjGBdfcF1ZDQ==",
|
"integrity": "sha512-XnAIvQ8eM+kC6aULx6wuQiwVsnzsi9d3WxzV3FpWTGA19F621kwdbsAcFKXgKUHZWsy+mY6iL1sHTxWEFCytDA==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"balanced-match": "^1.0.0"
|
"balanced-match": "^1.0.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/filelist/node_modules/minimatch": {
|
"node_modules/filelist/node_modules/minimatch": {
|
||||||
"version": "5.1.9",
|
"version": "5.1.6",
|
||||||
"resolved": "https://registry.npmjs.org/minimatch/-/minimatch-5.1.9.tgz",
|
"resolved": "https://registry.npmjs.org/minimatch/-/minimatch-5.1.6.tgz",
|
||||||
"integrity": "sha512-7o1wEA2RyMP7Iu7GNba9vc0RWWGACJOCZBJX2GJWip0ikV+wcOsgVuY9uE8CPiyQhkGFSlhuSkZPavN7u1c2Fw==",
|
"integrity": "sha512-lKwV/1brpG6mBUFHtb7NUmtABCb2WZZmm2wNiOA5hAb8VdCS4B3dtMWyvcoViccwAW/COERjXLt0zP1zXUN26g==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "ISC",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"brace-expansion": "^2.0.1"
|
"brace-expansion": "^2.0.1"
|
||||||
},
|
},
|
||||||
@@ -2592,16 +2604,39 @@
|
|||||||
"resolved": "https://registry.npmjs.org/flatbuffers/-/flatbuffers-1.12.0.tgz",
|
"resolved": "https://registry.npmjs.org/flatbuffers/-/flatbuffers-1.12.0.tgz",
|
||||||
"integrity": "sha512-c7CZADjRcl6j0PlvFy0ZqXQ67qSEZfrVPynmnL+2zPc+NtMvrF8Y0QceMo7QqnSPc7+uWjUIAbvCQ5WIKlMVdQ=="
|
"integrity": "sha512-c7CZADjRcl6j0PlvFy0ZqXQ67qSEZfrVPynmnL+2zPc+NtMvrF8Y0QceMo7QqnSPc7+uWjUIAbvCQ5WIKlMVdQ=="
|
||||||
},
|
},
|
||||||
|
"node_modules/foreground-child": {
|
||||||
|
"version": "3.3.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/foreground-child/-/foreground-child-3.3.0.tgz",
|
||||||
|
"integrity": "sha512-Ld2g8rrAyMYFXBhEqMz8ZAHBi4J4uS1i/CxGMDnjyFWddMXLVcDp051DZfu+t7+ab7Wv6SMqpWmyFIj5UbfFvg==",
|
||||||
|
"dependencies": {
|
||||||
|
"cross-spawn": "^7.0.0",
|
||||||
|
"signal-exit": "^4.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=14"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/foreground-child/node_modules/signal-exit": {
|
||||||
|
"version": "4.1.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-4.1.0.tgz",
|
||||||
|
"integrity": "sha512-bzyZ1e88w9O1iNJbKnOlvYTrWPDl46O1bG0D3XInv+9tkPrxrN8jUUTiFlDkkmKWgn1M6CfIA13SuGqOa9Korw==",
|
||||||
|
"engines": {
|
||||||
|
"node": ">=14"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/form-data": {
|
"node_modules/form-data": {
|
||||||
"version": "4.0.5",
|
"version": "4.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/form-data/-/form-data-4.0.5.tgz",
|
"resolved": "https://registry.npmjs.org/form-data/-/form-data-4.0.1.tgz",
|
||||||
"integrity": "sha512-8RipRLol37bNs2bhoV67fiTEvdTrbMUYcFTiy3+wuuOnUog2QBHCZWXDRijWQfAkhBj2Uf5UnVaiWwA5vdd82w==",
|
"integrity": "sha512-tzN8e4TX8+kkxGPK8D5u0FNmjPUjw3lwC9lSLxxoB/+GtsJG91CO8bSWy73APlgAZzZbXEYZJuxjkHH2w+Ezhw==",
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"asynckit": "^0.4.0",
|
"asynckit": "^0.4.0",
|
||||||
"combined-stream": "^1.0.8",
|
"combined-stream": "^1.0.8",
|
||||||
"es-set-tostringtag": "^2.1.0",
|
|
||||||
"hasown": "^2.0.2",
|
|
||||||
"mime-types": "^2.1.12"
|
"mime-types": "^2.1.12"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
@@ -2649,6 +2684,7 @@
|
|||||||
"version": "1.1.2",
|
"version": "1.1.2",
|
||||||
"resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.2.tgz",
|
"resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.2.tgz",
|
||||||
"integrity": "sha512-7XHNxH7qX9xG5mIwxkhumTox/MIRNcOgDrxWsMt2pAr23WHp6MrRlN7FBSFpCpr+oVO0F744iUgR82nJMfG2SA==",
|
"integrity": "sha512-7XHNxH7qX9xG5mIwxkhumTox/MIRNcOgDrxWsMt2pAr23WHp6MrRlN7FBSFpCpr+oVO0F744iUgR82nJMfG2SA==",
|
||||||
|
"dev": true,
|
||||||
"funding": {
|
"funding": {
|
||||||
"url": "https://github.com/sponsors/ljharb"
|
"url": "https://github.com/sponsors/ljharb"
|
||||||
}
|
}
|
||||||
@@ -2671,30 +2707,6 @@
|
|||||||
"node": "6.* || 8.* || >= 10.*"
|
"node": "6.* || 8.* || >= 10.*"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/get-intrinsic": {
|
|
||||||
"version": "1.3.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/get-intrinsic/-/get-intrinsic-1.3.0.tgz",
|
|
||||||
"integrity": "sha512-9fSjSaos/fRIVIp+xSJlE6lfwhES7LNtKaCBIamHsjr2na1BiABJPo0mOjjz8GJDURarmCPGqaiVg5mfjb98CQ==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"call-bind-apply-helpers": "^1.0.2",
|
|
||||||
"es-define-property": "^1.0.1",
|
|
||||||
"es-errors": "^1.3.0",
|
|
||||||
"es-object-atoms": "^1.1.1",
|
|
||||||
"function-bind": "^1.1.2",
|
|
||||||
"get-proto": "^1.0.1",
|
|
||||||
"gopd": "^1.2.0",
|
|
||||||
"has-symbols": "^1.1.0",
|
|
||||||
"hasown": "^2.0.2",
|
|
||||||
"math-intrinsics": "^1.1.0"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
},
|
|
||||||
"funding": {
|
|
||||||
"url": "https://github.com/sponsors/ljharb"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/get-package-type": {
|
"node_modules/get-package-type": {
|
||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"resolved": "https://registry.npmjs.org/get-package-type/-/get-package-type-0.1.0.tgz",
|
"resolved": "https://registry.npmjs.org/get-package-type/-/get-package-type-0.1.0.tgz",
|
||||||
@@ -2704,19 +2716,6 @@
|
|||||||
"node": ">=8.0.0"
|
"node": ">=8.0.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/get-proto": {
|
|
||||||
"version": "1.0.1",
|
|
||||||
"resolved": "https://registry.npmjs.org/get-proto/-/get-proto-1.0.1.tgz",
|
|
||||||
"integrity": "sha512-sTSfBjoXBp89JvIKIefqw7U2CCebsc74kiY6awiGogKtoSGbgjYE/G/+l9sF3MWFPNc9IcoOC4ODfKHfxFmp0g==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"dunder-proto": "^1.0.1",
|
|
||||||
"es-object-atoms": "^1.0.0"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/get-stream": {
|
"node_modules/get-stream": {
|
||||||
"version": "6.0.1",
|
"version": "6.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/get-stream/-/get-stream-6.0.1.tgz",
|
"resolved": "https://registry.npmjs.org/get-stream/-/get-stream-6.0.1.tgz",
|
||||||
@@ -2759,18 +2758,6 @@
|
|||||||
"node": ">=4"
|
"node": ">=4"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/gopd": {
|
|
||||||
"version": "1.2.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/gopd/-/gopd-1.2.0.tgz",
|
|
||||||
"integrity": "sha512-ZUKRh6/kUFoAiTAtTYPZJ3hw9wNxx+BIBOijnlG9PnrJsCcSjs1wyyD6vJpaYtgnzDrKYRSqf3OO6Rfa93xsRg==",
|
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
},
|
|
||||||
"funding": {
|
|
||||||
"url": "https://github.com/sponsors/ljharb"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/graceful-fs": {
|
"node_modules/graceful-fs": {
|
||||||
"version": "4.2.11",
|
"version": "4.2.11",
|
||||||
"resolved": "https://registry.npmjs.org/graceful-fs/-/graceful-fs-4.2.11.tgz",
|
"resolved": "https://registry.npmjs.org/graceful-fs/-/graceful-fs-4.2.11.tgz",
|
||||||
@@ -2791,37 +2778,11 @@
|
|||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/has-symbols": {
|
|
||||||
"version": "1.1.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/has-symbols/-/has-symbols-1.1.0.tgz",
|
|
||||||
"integrity": "sha512-1cDNdwJ2Jaohmb3sg4OmKaMBwuC48sYni5HUw2DvsC8LjGTLK9h+eb1X6RyuOHe4hT0ULCW68iomhjUoKUqlPQ==",
|
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
},
|
|
||||||
"funding": {
|
|
||||||
"url": "https://github.com/sponsors/ljharb"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/has-tostringtag": {
|
|
||||||
"version": "1.0.2",
|
|
||||||
"resolved": "https://registry.npmjs.org/has-tostringtag/-/has-tostringtag-1.0.2.tgz",
|
|
||||||
"integrity": "sha512-NqADB8VjPFLM2V0VvHUewwwsw0ZWBaIdgo+ieHtK3hasLz4qeCRjYcqfB6AQrBggRKppKF8L52/VqdVsO47Dlw==",
|
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
|
||||||
"has-symbols": "^1.0.3"
|
|
||||||
},
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
},
|
|
||||||
"funding": {
|
|
||||||
"url": "https://github.com/sponsors/ljharb"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/hasown": {
|
"node_modules/hasown": {
|
||||||
"version": "2.0.2",
|
"version": "2.0.2",
|
||||||
"resolved": "https://registry.npmjs.org/hasown/-/hasown-2.0.2.tgz",
|
"resolved": "https://registry.npmjs.org/hasown/-/hasown-2.0.2.tgz",
|
||||||
"integrity": "sha512-0hJU9SCPvmMzIBdZFqNPXWa6dqh7WdH0cII9y+CyS8rG3nL48Bclra9HmKhVVUHyPWNH5Y7xDwAB7bfgSjkUMQ==",
|
"integrity": "sha512-0hJU9SCPvmMzIBdZFqNPXWa6dqh7WdH0cII9y+CyS8rG3nL48Bclra9HmKhVVUHyPWNH5Y7xDwAB7bfgSjkUMQ==",
|
||||||
|
"dev": true,
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"function-bind": "^1.1.2"
|
"function-bind": "^1.1.2"
|
||||||
},
|
},
|
||||||
@@ -2921,7 +2882,6 @@
|
|||||||
"version": "3.0.0",
|
"version": "3.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz",
|
||||||
"integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==",
|
"integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==",
|
||||||
"dev": true,
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
@@ -2959,8 +2919,7 @@
|
|||||||
"node_modules/isexe": {
|
"node_modules/isexe": {
|
||||||
"version": "2.0.0",
|
"version": "2.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz",
|
||||||
"integrity": "sha512-RHxMLp9lnKHGHRng9QFhRCMbYAcVpn69smSGcq3f36xjgVVWThj4qqLbTLlq7Ssj8B+fIQ1EuCEGI2lKsyQeIw==",
|
"integrity": "sha512-RHxMLp9lnKHGHRng9QFhRCMbYAcVpn69smSGcq3f36xjgVVWThj4qqLbTLlq7Ssj8B+fIQ1EuCEGI2lKsyQeIw=="
|
||||||
"dev": true
|
|
||||||
},
|
},
|
||||||
"node_modules/istanbul-lib-coverage": {
|
"node_modules/istanbul-lib-coverage": {
|
||||||
"version": "3.2.2",
|
"version": "3.2.2",
|
||||||
@@ -3028,6 +2987,20 @@
|
|||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/jackspeak": {
|
||||||
|
"version": "3.4.3",
|
||||||
|
"resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-3.4.3.tgz",
|
||||||
|
"integrity": "sha512-OGlZQpz2yfahA/Rd1Y8Cd9SIEsqvXkLVoSw/cgwhnhFMDbsQFeZYoJJ7bIZBS9BcamUW96asq/npPWugM+RQBw==",
|
||||||
|
"dependencies": {
|
||||||
|
"@isaacs/cliui": "^8.0.2"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
},
|
||||||
|
"optionalDependencies": {
|
||||||
|
"@pkgjs/parseargs": "^0.11.0"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/jake": {
|
"node_modules/jake": {
|
||||||
"version": "10.9.2",
|
"version": "10.9.2",
|
||||||
"resolved": "https://registry.npmjs.org/jake/-/jake-10.9.2.tgz",
|
"resolved": "https://registry.npmjs.org/jake/-/jake-10.9.2.tgz",
|
||||||
@@ -3632,11 +3605,10 @@
|
|||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
"node_modules/js-yaml": {
|
"node_modules/js-yaml": {
|
||||||
"version": "3.14.2",
|
"version": "3.14.1",
|
||||||
"resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-3.14.2.tgz",
|
"resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-3.14.1.tgz",
|
||||||
"integrity": "sha512-PMSmkqxr106Xa156c2M265Z+FTrPl+oxd/rgOQy2tijQeK5TxQ43psO1ZCwhVOSdnn+RzkzlRz/eY4BgJBYVpg==",
|
"integrity": "sha512-okMH7OXXJ7YrN9Ok3/SXrnu4iX9yOk+25nqX4imS2npuvTYDmo/QEZoqwZkYaIDk3jVvBOTOIEgEhaLOynBS9g==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"argparse": "^1.0.7",
|
"argparse": "^1.0.7",
|
||||||
"esprima": "^4.0.0"
|
"esprima": "^4.0.0"
|
||||||
@@ -3756,15 +3728,6 @@
|
|||||||
"tmpl": "1.0.5"
|
"tmpl": "1.0.5"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/math-intrinsics": {
|
|
||||||
"version": "1.1.0",
|
|
||||||
"resolved": "https://registry.npmjs.org/math-intrinsics/-/math-intrinsics-1.1.0.tgz",
|
|
||||||
"integrity": "sha512-/IXtbwEk5HTPyEwyKX6hGkYXxM9nbj64B+ilVJnC/R6B0pH5G4V3b0pVbL7DBj4tkhBAppbQUlf6F6Xl9LHu1g==",
|
|
||||||
"license": "MIT",
|
|
||||||
"engines": {
|
|
||||||
"node": ">= 0.4"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"node_modules/merge-stream": {
|
"node_modules/merge-stream": {
|
||||||
"version": "2.0.0",
|
"version": "2.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/merge-stream/-/merge-stream-2.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/merge-stream/-/merge-stream-2.0.0.tgz",
|
||||||
@@ -3813,11 +3776,10 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/minimatch": {
|
"node_modules/minimatch": {
|
||||||
"version": "3.1.5",
|
"version": "3.1.2",
|
||||||
"resolved": "https://registry.npmjs.org/minimatch/-/minimatch-3.1.5.tgz",
|
"resolved": "https://registry.npmjs.org/minimatch/-/minimatch-3.1.2.tgz",
|
||||||
"integrity": "sha512-VgjWUsnnT6n+NUk6eZq77zeFdpW2LWDzP6zFGrCbHXiYNul5Dzqk2HHQ5uFH2DNW5Xbp8+jVzaeNt94ssEEl4w==",
|
"integrity": "sha512-J7p63hRiAjw1NDEww1W7i37+ByIrOWO5XQQAzZ3VOcL0PNybwpfmV/N05zFAzwQ9USyEcX6t3UO+K5aqBQOIHw==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"license": "ISC",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"brace-expansion": "^1.1.7"
|
"brace-expansion": "^1.1.7"
|
||||||
},
|
},
|
||||||
@@ -3834,17 +3796,31 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/minizlib": {
|
"node_modules/minizlib": {
|
||||||
"version": "3.1.0",
|
"version": "3.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.1.0.tgz",
|
"resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.0.1.tgz",
|
||||||
"integrity": "sha512-KZxYo1BUkWD2TVFLr0MQoM8vUUigWD3LlD83a/75BqC+4qE0Hb1Vo5v1FgcfaNXvfXzr+5EhQ6ing/CaBijTlw==",
|
"integrity": "sha512-umcy022ILvb5/3Djuu8LWeqUa8D68JaBzlttKeMWen48SjabqS3iY5w/vzeMzMUNhLDifyhbOwKDSznB1vvrwg==",
|
||||||
"license": "MIT",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"minipass": "^7.1.2"
|
"minipass": "^7.0.4",
|
||||||
|
"rimraf": "^5.0.5"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">= 18"
|
"node": ">= 18"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/mkdirp": {
|
||||||
|
"version": "3.0.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-3.0.1.tgz",
|
||||||
|
"integrity": "sha512-+NsyUUAZDmo6YVHzL/stxSu3t9YS1iljliy3BSDrXJ/dkn1KYdmtZODGGjLcc9XLgVVpH4KshHB8XmZgMhaBXg==",
|
||||||
|
"bin": {
|
||||||
|
"mkdirp": "dist/cjs/src/bin.js"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=10"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/ms": {
|
"node_modules/ms": {
|
||||||
"version": "2.1.3",
|
"version": "2.1.3",
|
||||||
"resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz",
|
"resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz",
|
||||||
@@ -4034,6 +4010,11 @@
|
|||||||
"node": ">=6"
|
"node": ">=6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/package-json-from-dist": {
|
||||||
|
"version": "1.0.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/package-json-from-dist/-/package-json-from-dist-1.0.1.tgz",
|
||||||
|
"integrity": "sha512-UEZIS3/by4OC8vL3P2dTXRETpebLI2NiI5vIrjaD/5UtrkFX/tNbwjTSRAGC/+7CAo2pIcBaRgWmcBBHcsaCIw=="
|
||||||
|
},
|
||||||
"node_modules/parse-json": {
|
"node_modules/parse-json": {
|
||||||
"version": "5.2.0",
|
"version": "5.2.0",
|
||||||
"resolved": "https://registry.npmjs.org/parse-json/-/parse-json-5.2.0.tgz",
|
"resolved": "https://registry.npmjs.org/parse-json/-/parse-json-5.2.0.tgz",
|
||||||
@@ -4074,7 +4055,6 @@
|
|||||||
"version": "3.1.1",
|
"version": "3.1.1",
|
||||||
"resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz",
|
"resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz",
|
||||||
"integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==",
|
"integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==",
|
||||||
"dev": true,
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
@@ -4085,6 +4065,26 @@
|
|||||||
"integrity": "sha512-LDJzPVEEEPR+y48z93A0Ed0yXb8pAByGWo/k5YYdYgpY2/2EsOsksJrq7lOHxryrVOn1ejG6oAp8ahvOIQD8sw==",
|
"integrity": "sha512-LDJzPVEEEPR+y48z93A0Ed0yXb8pAByGWo/k5YYdYgpY2/2EsOsksJrq7lOHxryrVOn1ejG6oAp8ahvOIQD8sw==",
|
||||||
"dev": true
|
"dev": true
|
||||||
},
|
},
|
||||||
|
"node_modules/path-scurry": {
|
||||||
|
"version": "1.11.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.11.1.tgz",
|
||||||
|
"integrity": "sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA==",
|
||||||
|
"dependencies": {
|
||||||
|
"lru-cache": "^10.2.0",
|
||||||
|
"minipass": "^5.0.0 || ^6.0.2 || ^7.0.0"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=16 || 14 >=14.18"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/path-scurry/node_modules/lru-cache": {
|
||||||
|
"version": "10.4.3",
|
||||||
|
"resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.4.3.tgz",
|
||||||
|
"integrity": "sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ=="
|
||||||
|
},
|
||||||
"node_modules/picocolors": {
|
"node_modules/picocolors": {
|
||||||
"version": "1.1.1",
|
"version": "1.1.1",
|
||||||
"resolved": "https://registry.npmjs.org/picocolors/-/picocolors-1.1.1.tgz",
|
"resolved": "https://registry.npmjs.org/picocolors/-/picocolors-1.1.1.tgz",
|
||||||
@@ -4246,6 +4246,61 @@
|
|||||||
"node": ">=10"
|
"node": ">=10"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/rimraf": {
|
||||||
|
"version": "5.0.10",
|
||||||
|
"resolved": "https://registry.npmjs.org/rimraf/-/rimraf-5.0.10.tgz",
|
||||||
|
"integrity": "sha512-l0OE8wL34P4nJH/H2ffoaniAokM2qSmrtXHmlpvYr5AVVX8msAyW0l8NVJFDxlSK4u3Uh/f41cQheDVdnYijwQ==",
|
||||||
|
"dependencies": {
|
||||||
|
"glob": "^10.3.7"
|
||||||
|
},
|
||||||
|
"bin": {
|
||||||
|
"rimraf": "dist/esm/bin.mjs"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/rimraf/node_modules/brace-expansion": {
|
||||||
|
"version": "2.0.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.1.tgz",
|
||||||
|
"integrity": "sha512-XnAIvQ8eM+kC6aULx6wuQiwVsnzsi9d3WxzV3FpWTGA19F621kwdbsAcFKXgKUHZWsy+mY6iL1sHTxWEFCytDA==",
|
||||||
|
"dependencies": {
|
||||||
|
"balanced-match": "^1.0.0"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/rimraf/node_modules/glob": {
|
||||||
|
"version": "10.4.5",
|
||||||
|
"resolved": "https://registry.npmjs.org/glob/-/glob-10.4.5.tgz",
|
||||||
|
"integrity": "sha512-7Bv8RF0k6xjo7d4A/PxYLbUCfb6c+Vpd2/mB2yRDlew7Jb5hEXiCD9ibfO7wpk8i4sevK6DFny9h7EYbM3/sHg==",
|
||||||
|
"dependencies": {
|
||||||
|
"foreground-child": "^3.1.0",
|
||||||
|
"jackspeak": "^3.1.2",
|
||||||
|
"minimatch": "^9.0.4",
|
||||||
|
"minipass": "^7.1.2",
|
||||||
|
"package-json-from-dist": "^1.0.0",
|
||||||
|
"path-scurry": "^1.11.1"
|
||||||
|
},
|
||||||
|
"bin": {
|
||||||
|
"glob": "dist/esm/bin.mjs"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/rimraf/node_modules/minimatch": {
|
||||||
|
"version": "9.0.5",
|
||||||
|
"resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.5.tgz",
|
||||||
|
"integrity": "sha512-G6T0ZX48xgozx7587koeX9Ys2NYy6Gmv//P89sEte9V9whIapMNF4idKxnW2QtCcLiTWlb/wfCabAtAFWhhBow==",
|
||||||
|
"dependencies": {
|
||||||
|
"brace-expansion": "^2.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=16 || 14 >=14.17"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/sponsors/isaacs"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/semver": {
|
"node_modules/semver": {
|
||||||
"version": "7.6.3",
|
"version": "7.6.3",
|
||||||
"resolved": "https://registry.npmjs.org/semver/-/semver-7.6.3.tgz",
|
"resolved": "https://registry.npmjs.org/semver/-/semver-7.6.3.tgz",
|
||||||
@@ -4299,7 +4354,6 @@
|
|||||||
"version": "2.0.0",
|
"version": "2.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz",
|
||||||
"integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==",
|
"integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==",
|
||||||
"dev": true,
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"shebang-regex": "^3.0.0"
|
"shebang-regex": "^3.0.0"
|
||||||
},
|
},
|
||||||
@@ -4311,7 +4365,6 @@
|
|||||||
"version": "3.0.0",
|
"version": "3.0.0",
|
||||||
"resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz",
|
"resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz",
|
||||||
"integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==",
|
"integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==",
|
||||||
"dev": true,
|
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=8"
|
"node": ">=8"
|
||||||
}
|
}
|
||||||
@@ -4399,7 +4452,20 @@
|
|||||||
"version": "4.2.3",
|
"version": "4.2.3",
|
||||||
"resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz",
|
"resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz",
|
||||||
"integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==",
|
"integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==",
|
||||||
"dev": true,
|
"dependencies": {
|
||||||
|
"emoji-regex": "^8.0.0",
|
||||||
|
"is-fullwidth-code-point": "^3.0.0",
|
||||||
|
"strip-ansi": "^6.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=8"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/string-width-cjs": {
|
||||||
|
"name": "string-width",
|
||||||
|
"version": "4.2.3",
|
||||||
|
"resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz",
|
||||||
|
"integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"emoji-regex": "^8.0.0",
|
"emoji-regex": "^8.0.0",
|
||||||
"is-fullwidth-code-point": "^3.0.0",
|
"is-fullwidth-code-point": "^3.0.0",
|
||||||
@@ -4413,7 +4479,18 @@
|
|||||||
"version": "6.0.1",
|
"version": "6.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz",
|
"resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz",
|
||||||
"integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==",
|
"integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==",
|
||||||
"dev": true,
|
"dependencies": {
|
||||||
|
"ansi-regex": "^5.0.1"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=8"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/strip-ansi-cjs": {
|
||||||
|
"name": "strip-ansi",
|
||||||
|
"version": "6.0.1",
|
||||||
|
"resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz",
|
||||||
|
"integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"ansi-regex": "^5.0.1"
|
"ansi-regex": "^5.0.1"
|
||||||
},
|
},
|
||||||
@@ -4464,15 +4541,15 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/tar": {
|
"node_modules/tar": {
|
||||||
"version": "7.5.10",
|
"version": "7.4.3",
|
||||||
"resolved": "https://registry.npmjs.org/tar/-/tar-7.5.10.tgz",
|
"resolved": "https://registry.npmjs.org/tar/-/tar-7.4.3.tgz",
|
||||||
"integrity": "sha512-8mOPs1//5q/rlkNSPcCegA6hiHJYDmSLEI8aMH/CdSQJNWztHC9WHNam5zdQlfpTwB9Xp7IBEsHfV5LKMJGVAw==",
|
"integrity": "sha512-5S7Va8hKfV7W5U6g3aYxXmlPoZVAwUMy9AOKyF2fVuZa2UD3qZjg578OrLRt8PcNN1PleVaL/5/yYATNL0ICUw==",
|
||||||
"license": "BlueOak-1.0.0",
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@isaacs/fs-minipass": "^4.0.0",
|
"@isaacs/fs-minipass": "^4.0.0",
|
||||||
"chownr": "^3.0.0",
|
"chownr": "^3.0.0",
|
||||||
"minipass": "^7.1.2",
|
"minipass": "^7.1.2",
|
||||||
"minizlib": "^3.1.0",
|
"minizlib": "^3.0.1",
|
||||||
|
"mkdirp": "^3.0.1",
|
||||||
"yallist": "^5.0.0"
|
"yallist": "^5.0.0"
|
||||||
},
|
},
|
||||||
"engines": {
|
"engines": {
|
||||||
@@ -4705,7 +4782,6 @@
|
|||||||
"version": "2.0.2",
|
"version": "2.0.2",
|
||||||
"resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz",
|
"resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz",
|
||||||
"integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==",
|
"integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==",
|
||||||
"dev": true,
|
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"isexe": "^2.0.0"
|
"isexe": "^2.0.0"
|
||||||
},
|
},
|
||||||
@@ -4733,6 +4809,23 @@
|
|||||||
"url": "https://github.com/chalk/wrap-ansi?sponsor=1"
|
"url": "https://github.com/chalk/wrap-ansi?sponsor=1"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/wrap-ansi-cjs": {
|
||||||
|
"name": "wrap-ansi",
|
||||||
|
"version": "7.0.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz",
|
||||||
|
"integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==",
|
||||||
|
"dependencies": {
|
||||||
|
"ansi-styles": "^4.0.0",
|
||||||
|
"string-width": "^4.1.0",
|
||||||
|
"strip-ansi": "^6.0.0"
|
||||||
|
},
|
||||||
|
"engines": {
|
||||||
|
"node": ">=10"
|
||||||
|
},
|
||||||
|
"funding": {
|
||||||
|
"url": "https://github.com/chalk/wrap-ansi?sponsor=1"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/wrappy": {
|
"node_modules/wrappy": {
|
||||||
"version": "1.0.2",
|
"version": "1.0.2",
|
||||||
"resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz",
|
"resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz",
|
||||||
|
|||||||
@@ -20,8 +20,6 @@ import {
|
|||||||
Float32,
|
Float32,
|
||||||
Float64,
|
Float64,
|
||||||
Int,
|
Int,
|
||||||
Int8,
|
|
||||||
Int16,
|
|
||||||
Int32,
|
Int32,
|
||||||
Int64,
|
Int64,
|
||||||
LargeBinary,
|
LargeBinary,
|
||||||
@@ -37,8 +35,6 @@ import {
|
|||||||
Timestamp,
|
Timestamp,
|
||||||
Type,
|
Type,
|
||||||
Uint8,
|
Uint8,
|
||||||
Uint16,
|
|
||||||
Uint32,
|
|
||||||
Utf8,
|
Utf8,
|
||||||
Vector,
|
Vector,
|
||||||
makeVector as arrowMakeVector,
|
makeVector as arrowMakeVector,
|
||||||
@@ -533,8 +529,7 @@ function isObject(value: unknown): value is Record<string, unknown> {
|
|||||||
!(value instanceof Date) &&
|
!(value instanceof Date) &&
|
||||||
!(value instanceof Set) &&
|
!(value instanceof Set) &&
|
||||||
!(value instanceof Map) &&
|
!(value instanceof Map) &&
|
||||||
!(value instanceof Buffer) &&
|
!(value instanceof Buffer)
|
||||||
!ArrayBuffer.isView(value)
|
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -593,13 +588,6 @@ function inferType(
|
|||||||
return new Bool();
|
return new Bool();
|
||||||
} else if (value instanceof Buffer) {
|
} else if (value instanceof Buffer) {
|
||||||
return new Binary();
|
return new Binary();
|
||||||
} else if (ArrayBuffer.isView(value) && !(value instanceof DataView)) {
|
|
||||||
const info = typedArrayToArrowType(value);
|
|
||||||
if (info !== undefined) {
|
|
||||||
const child = new Field("item", info.elementType, true);
|
|
||||||
return new FixedSizeList(info.length, child);
|
|
||||||
}
|
|
||||||
return undefined;
|
|
||||||
} else if (Array.isArray(value)) {
|
} else if (Array.isArray(value)) {
|
||||||
if (value.length === 0) {
|
if (value.length === 0) {
|
||||||
return undefined; // Without any values we can't infer the type
|
return undefined; // Without any values we can't infer the type
|
||||||
@@ -758,32 +746,6 @@ function makeListVector(lists: unknown[][]): Vector<unknown> {
|
|||||||
return listBuilder.finish().toVector();
|
return listBuilder.finish().toVector();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Map a JS TypedArray instance to the corresponding Arrow element DataType
|
|
||||||
* and its length. Returns undefined if the value is not a recognized TypedArray.
|
|
||||||
*/
|
|
||||||
function typedArrayToArrowType(
|
|
||||||
value: ArrayBufferView,
|
|
||||||
): { elementType: DataType; length: number } | undefined {
|
|
||||||
if (value instanceof Float32Array)
|
|
||||||
return { elementType: new Float32(), length: value.length };
|
|
||||||
if (value instanceof Float64Array)
|
|
||||||
return { elementType: new Float64(), length: value.length };
|
|
||||||
if (value instanceof Uint8Array)
|
|
||||||
return { elementType: new Uint8(), length: value.length };
|
|
||||||
if (value instanceof Uint16Array)
|
|
||||||
return { elementType: new Uint16(), length: value.length };
|
|
||||||
if (value instanceof Uint32Array)
|
|
||||||
return { elementType: new Uint32(), length: value.length };
|
|
||||||
if (value instanceof Int8Array)
|
|
||||||
return { elementType: new Int8(), length: value.length };
|
|
||||||
if (value instanceof Int16Array)
|
|
||||||
return { elementType: new Int16(), length: value.length };
|
|
||||||
if (value instanceof Int32Array)
|
|
||||||
return { elementType: new Int32(), length: value.length };
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Helper function to convert an Array of JS values to an Arrow Vector */
|
/** Helper function to convert an Array of JS values to an Arrow Vector */
|
||||||
function makeVector(
|
function makeVector(
|
||||||
values: unknown[],
|
values: unknown[],
|
||||||
@@ -852,16 +814,6 @@ function makeVector(
|
|||||||
"makeVector cannot infer the type if all values are null or undefined",
|
"makeVector cannot infer the type if all values are null or undefined",
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
if (ArrayBuffer.isView(sampleValue) && !(sampleValue instanceof DataView)) {
|
|
||||||
const info = typedArrayToArrowType(sampleValue);
|
|
||||||
if (info !== undefined) {
|
|
||||||
const fslType = new FixedSizeList(
|
|
||||||
info.length,
|
|
||||||
new Field("item", info.elementType, true),
|
|
||||||
);
|
|
||||||
return vectorFromArray(values, fslType);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (Array.isArray(sampleValue)) {
|
if (Array.isArray(sampleValue)) {
|
||||||
// Default Arrow inference doesn't handle list types
|
// Default Arrow inference doesn't handle list types
|
||||||
return makeListVector(values as unknown[][]);
|
return makeListVector(values as unknown[][]);
|
||||||
|
|||||||
@@ -5,15 +5,12 @@ import {
|
|||||||
Table as ArrowTable,
|
Table as ArrowTable,
|
||||||
Data,
|
Data,
|
||||||
DataType,
|
DataType,
|
||||||
Field,
|
|
||||||
IntoVector,
|
IntoVector,
|
||||||
MultiVector,
|
MultiVector,
|
||||||
Schema,
|
Schema,
|
||||||
dataTypeToJson,
|
dataTypeToJson,
|
||||||
fromDataToBuffer,
|
fromDataToBuffer,
|
||||||
fromTableToBuffer,
|
|
||||||
isMultiVector,
|
isMultiVector,
|
||||||
makeEmptyTable,
|
|
||||||
tableFromIPC,
|
tableFromIPC,
|
||||||
} from "./arrow";
|
} from "./arrow";
|
||||||
|
|
||||||
@@ -87,16 +84,6 @@ export interface OptimizeOptions {
|
|||||||
* tbl.optimize({cleanupOlderThan: new Date()});
|
* tbl.optimize({cleanupOlderThan: new Date()});
|
||||||
*/
|
*/
|
||||||
cleanupOlderThan: Date;
|
cleanupOlderThan: Date;
|
||||||
/**
|
|
||||||
* Because they may be part of an in-progress transaction, files newer than
|
|
||||||
* 7 days old are not deleted by default. If you are sure that there are no
|
|
||||||
* in-progress transactions, then you can set this to true to delete all
|
|
||||||
* files older than `cleanupOlderThan`.
|
|
||||||
*
|
|
||||||
* **WARNING**: This should only be set to true if you can guarantee that
|
|
||||||
* no other process is currently working on this dataset. Otherwise the
|
|
||||||
* dataset could be put into a corrupted state.
|
|
||||||
*/
|
|
||||||
deleteUnverified: boolean;
|
deleteUnverified: boolean;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -394,16 +381,15 @@ export abstract class Table {
|
|||||||
abstract vectorSearch(vector: IntoVector | MultiVector): VectorQuery;
|
abstract vectorSearch(vector: IntoVector | MultiVector): VectorQuery;
|
||||||
/**
|
/**
|
||||||
* Add new columns with defined values.
|
* Add new columns with defined values.
|
||||||
* @param {AddColumnsSql[] | Field | Field[] | Schema} newColumnTransforms Either:
|
* @param {AddColumnsSql[]} newColumnTransforms pairs of column names and
|
||||||
* - An array of objects with column names and SQL expressions to calculate values
|
* the SQL expression to use to calculate the value of the new column. These
|
||||||
* - A single Arrow Field defining one column with its data type (column will be initialized with null values)
|
* expressions will be evaluated for each row in the table, and can
|
||||||
* - An array of Arrow Fields defining columns with their data types (columns will be initialized with null values)
|
* reference existing columns in the table.
|
||||||
* - An Arrow Schema defining columns with their data types (columns will be initialized with null values)
|
|
||||||
* @returns {Promise<AddColumnsResult>} A promise that resolves to an object
|
* @returns {Promise<AddColumnsResult>} A promise that resolves to an object
|
||||||
* containing the new version number of the table after adding the columns.
|
* containing the new version number of the table after adding the columns.
|
||||||
*/
|
*/
|
||||||
abstract addColumns(
|
abstract addColumns(
|
||||||
newColumnTransforms: AddColumnsSql[] | Field | Field[] | Schema,
|
newColumnTransforms: AddColumnsSql[],
|
||||||
): Promise<AddColumnsResult>;
|
): Promise<AddColumnsResult>;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -515,7 +501,19 @@ export abstract class Table {
|
|||||||
* - Index: Optimizes the indices, adding new data to existing indices
|
* - Index: Optimizes the indices, adding new data to existing indices
|
||||||
*
|
*
|
||||||
*
|
*
|
||||||
* The frequency an application should call optimize is based on the frequency of
|
* Experimental API
|
||||||
|
* ----------------
|
||||||
|
*
|
||||||
|
* The optimization process is undergoing active development and may change.
|
||||||
|
* Our goal with these changes is to improve the performance of optimization and
|
||||||
|
* reduce the complexity.
|
||||||
|
*
|
||||||
|
* That being said, it is essential today to run optimize if you want the best
|
||||||
|
* performance. It should be stable and safe to use in production, but it our
|
||||||
|
* hope that the API may be simplified (or not even need to be called) in the
|
||||||
|
* future.
|
||||||
|
*
|
||||||
|
* The frequency an application shoudl call optimize is based on the frequency of
|
||||||
* data modifications. If data is frequently added, deleted, or updated then
|
* data modifications. If data is frequently added, deleted, or updated then
|
||||||
* optimize should be run frequently. A good rule of thumb is to run optimize if
|
* optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
* you have added or modified 100,000 or more records or run more than 20 data
|
* you have added or modified 100,000 or more records or run more than 20 data
|
||||||
@@ -808,40 +806,9 @@ export class LocalTable extends Table {
|
|||||||
// TODO: Support BatchUDF
|
// TODO: Support BatchUDF
|
||||||
|
|
||||||
async addColumns(
|
async addColumns(
|
||||||
newColumnTransforms: AddColumnsSql[] | Field | Field[] | Schema,
|
newColumnTransforms: AddColumnsSql[],
|
||||||
): Promise<AddColumnsResult> {
|
): Promise<AddColumnsResult> {
|
||||||
// Handle single Field -> convert to array of Fields
|
return await this.inner.addColumns(newColumnTransforms);
|
||||||
if (newColumnTransforms instanceof Field) {
|
|
||||||
newColumnTransforms = [newColumnTransforms];
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle array of Fields -> convert to Schema
|
|
||||||
if (
|
|
||||||
Array.isArray(newColumnTransforms) &&
|
|
||||||
newColumnTransforms.length > 0 &&
|
|
||||||
newColumnTransforms[0] instanceof Field
|
|
||||||
) {
|
|
||||||
const fields = newColumnTransforms as Field[];
|
|
||||||
newColumnTransforms = new Schema(fields);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle Schema -> use schema-based approach
|
|
||||||
if (newColumnTransforms instanceof Schema) {
|
|
||||||
const schema = newColumnTransforms;
|
|
||||||
// Convert schema to buffer using Arrow IPC format
|
|
||||||
const emptyTable = makeEmptyTable(schema);
|
|
||||||
const schemaBuf = await fromTableToBuffer(emptyTable);
|
|
||||||
return await this.inner.addColumnsWithSchema(schemaBuf);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle SQL expressions (existing functionality)
|
|
||||||
if (Array.isArray(newColumnTransforms)) {
|
|
||||||
return await this.inner.addColumns(
|
|
||||||
newColumnTransforms as AddColumnsSql[],
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
throw new Error("Invalid input type for addColumns");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async alterColumns(
|
async alterColumns(
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-darwin-arm64",
|
"name": "@lancedb/lancedb-darwin-arm64",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["darwin"],
|
"os": ["darwin"],
|
||||||
"cpu": ["arm64"],
|
"cpu": ["arm64"],
|
||||||
"main": "lancedb.darwin-arm64.node",
|
"main": "lancedb.darwin-arm64.node",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-linux-arm64-gnu",
|
"name": "@lancedb/lancedb-linux-arm64-gnu",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["linux"],
|
"os": ["linux"],
|
||||||
"cpu": ["arm64"],
|
"cpu": ["arm64"],
|
||||||
"main": "lancedb.linux-arm64-gnu.node",
|
"main": "lancedb.linux-arm64-gnu.node",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-linux-arm64-musl",
|
"name": "@lancedb/lancedb-linux-arm64-musl",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["linux"],
|
"os": ["linux"],
|
||||||
"cpu": ["arm64"],
|
"cpu": ["arm64"],
|
||||||
"main": "lancedb.linux-arm64-musl.node",
|
"main": "lancedb.linux-arm64-musl.node",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-linux-x64-gnu",
|
"name": "@lancedb/lancedb-linux-x64-gnu",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["linux"],
|
"os": ["linux"],
|
||||||
"cpu": ["x64"],
|
"cpu": ["x64"],
|
||||||
"main": "lancedb.linux-x64-gnu.node",
|
"main": "lancedb.linux-x64-gnu.node",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-linux-x64-musl",
|
"name": "@lancedb/lancedb-linux-x64-musl",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["linux"],
|
"os": ["linux"],
|
||||||
"cpu": ["x64"],
|
"cpu": ["x64"],
|
||||||
"main": "lancedb.linux-x64-musl.node",
|
"main": "lancedb.linux-x64-musl.node",
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-win32-arm64-msvc",
|
"name": "@lancedb/lancedb-win32-arm64-msvc",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": [
|
"os": [
|
||||||
"win32"
|
"win32"
|
||||||
],
|
],
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@lancedb/lancedb-win32-x64-msvc",
|
"name": "@lancedb/lancedb-win32-x64-msvc",
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"os": ["win32"],
|
"os": ["win32"],
|
||||||
"cpu": ["x64"],
|
"cpu": ["x64"],
|
||||||
"main": "lancedb.win32-x64-msvc.node",
|
"main": "lancedb.win32-x64-msvc.node",
|
||||||
|
|||||||
4599
nodejs/package-lock.json
generated
4599
nodejs/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@@ -11,7 +11,7 @@
|
|||||||
"ann"
|
"ann"
|
||||||
],
|
],
|
||||||
"private": false,
|
"private": false,
|
||||||
"version": "0.27.0",
|
"version": "0.27.0-beta.3",
|
||||||
"main": "dist/index.js",
|
"main": "dist/index.js",
|
||||||
"exports": {
|
"exports": {
|
||||||
".": "./dist/index.js",
|
".": "./dist/index.js",
|
||||||
|
|||||||
@@ -145,6 +145,7 @@ impl From<ClientConfig> for lancedb::remote::ClientConfig {
|
|||||||
id_delimiter: config.id_delimiter,
|
id_delimiter: config.id_delimiter,
|
||||||
tls_config: config.tls_config.map(Into::into),
|
tls_config: config.tls_config.map(Into::into),
|
||||||
header_provider: None, // the header provider is set separately later
|
header_provider: None, // the header provider is set separately later
|
||||||
|
mem_wal_enabled: None, // mem_wal is set per-operation in merge_insert
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,7 +3,7 @@
|
|||||||
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
use lancedb::ipc::{ipc_file_to_batches, ipc_file_to_schema};
|
use lancedb::ipc::ipc_file_to_batches;
|
||||||
use lancedb::table::{
|
use lancedb::table::{
|
||||||
AddDataMode, ColumnAlteration as LanceColumnAlteration, Duration, NewColumnTransform,
|
AddDataMode, ColumnAlteration as LanceColumnAlteration, Duration, NewColumnTransform,
|
||||||
OptimizeAction, OptimizeOptions, Table as LanceDbTable,
|
OptimizeAction, OptimizeOptions, Table as LanceDbTable,
|
||||||
@@ -279,23 +279,6 @@ impl Table {
|
|||||||
Ok(res.into())
|
Ok(res.into())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[napi(catch_unwind)]
|
|
||||||
pub async fn add_columns_with_schema(
|
|
||||||
&self,
|
|
||||||
schema_buf: Buffer,
|
|
||||||
) -> napi::Result<AddColumnsResult> {
|
|
||||||
let schema = ipc_file_to_schema(schema_buf.to_vec())
|
|
||||||
.map_err(|e| napi::Error::from_reason(format!("Failed to read IPC schema: {}", e)))?;
|
|
||||||
|
|
||||||
let transforms = NewColumnTransform::AllNulls(schema);
|
|
||||||
let res = self
|
|
||||||
.inner_ref()?
|
|
||||||
.add_columns(transforms, None)
|
|
||||||
.await
|
|
||||||
.default_error()?;
|
|
||||||
Ok(res.into())
|
|
||||||
}
|
|
||||||
|
|
||||||
#[napi(catch_unwind)]
|
#[napi(catch_unwind)]
|
||||||
pub async fn alter_columns(
|
pub async fn alter_columns(
|
||||||
&self,
|
&self,
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
[tool.bumpversion]
|
[tool.bumpversion]
|
||||||
current_version = "0.30.1"
|
current_version = "0.30.0-beta.3"
|
||||||
parse = """(?x)
|
parse = """(?x)
|
||||||
(?P<major>0|[1-9]\\d*)\\.
|
(?P<major>0|[1-9]\\d*)\\.
|
||||||
(?P<minor>0|[1-9]\\d*)\\.
|
(?P<minor>0|[1-9]\\d*)\\.
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "lancedb-python"
|
name = "lancedb-python"
|
||||||
version = "0.30.1"
|
version = "0.30.0-beta.3"
|
||||||
edition.workspace = true
|
edition.workspace = true
|
||||||
description = "Python bindings for LanceDB"
|
description = "Python bindings for LanceDB"
|
||||||
license.workspace = true
|
license.workspace = true
|
||||||
@@ -16,11 +16,9 @@ crate-type = ["cdylib"]
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
arrow = { version = "57.2", features = ["pyarrow"] }
|
arrow = { version = "57.2", features = ["pyarrow"] }
|
||||||
async-trait = "0.1"
|
async-trait = "0.1"
|
||||||
bytes = "1"
|
|
||||||
lancedb = { path = "../rust/lancedb", default-features = false }
|
lancedb = { path = "../rust/lancedb", default-features = false }
|
||||||
lance-core.workspace = true
|
lance-core.workspace = true
|
||||||
lance-namespace.workspace = true
|
lance-namespace.workspace = true
|
||||||
lance-namespace-impls.workspace = true
|
|
||||||
lance-io.workspace = true
|
lance-io.workspace = true
|
||||||
env_logger.workspace = true
|
env_logger.workspace = true
|
||||||
pyo3 = { version = "0.26", features = ["extension-module", "abi3-py39"] }
|
pyo3 = { version = "0.26", features = ["extension-module", "abi3-py39"] }
|
||||||
@@ -30,8 +28,6 @@ pyo3-async-runtimes = { version = "0.26", features = [
|
|||||||
] }
|
] }
|
||||||
pin-project = "1.1.5"
|
pin-project = "1.1.5"
|
||||||
futures.workspace = true
|
futures.workspace = true
|
||||||
serde = "1"
|
|
||||||
serde_json = "1"
|
|
||||||
snafu.workspace = true
|
snafu.workspace = true
|
||||||
tokio = { version = "1.40", features = ["sync"] }
|
tokio = { version = "1.40", features = ["sync"] }
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# LanceDB Python SDK
|
# LanceDB
|
||||||
|
|
||||||
A Python library for [LanceDB](https://github.com/lancedb/lancedb).
|
A Python library for [LanceDB](https://github.com/lancedb/lancedb).
|
||||||
|
|
||||||
|
|||||||
@@ -3,10 +3,10 @@ name = "lancedb"
|
|||||||
# version in Cargo.toml
|
# version in Cargo.toml
|
||||||
dynamic = ["version"]
|
dynamic = ["version"]
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"deprecation>=2.1.0",
|
"deprecation",
|
||||||
"numpy>=1.24.0",
|
"numpy",
|
||||||
"overrides>=0.7; python_version<'3.12'",
|
"overrides>=0.7; python_version<'3.12'",
|
||||||
"packaging>=23.0",
|
"packaging",
|
||||||
"pyarrow>=16",
|
"pyarrow>=16",
|
||||||
"pydantic>=1.10",
|
"pydantic>=1.10",
|
||||||
"tqdm>=4.27.0",
|
"tqdm>=4.27.0",
|
||||||
@@ -45,51 +45,51 @@ repository = "https://github.com/lancedb/lancedb"
|
|||||||
|
|
||||||
[project.optional-dependencies]
|
[project.optional-dependencies]
|
||||||
pylance = [
|
pylance = [
|
||||||
"pylance>=4.0.0b7",
|
"pylance>=1.0.0b14",
|
||||||
]
|
]
|
||||||
tests = [
|
tests = [
|
||||||
"aiohttp>=3.9.0",
|
"aiohttp",
|
||||||
"boto3>=1.28.57",
|
"boto3",
|
||||||
"pandas>=1.4",
|
"pandas>=1.4",
|
||||||
"pytest>=7.0",
|
"pytest",
|
||||||
"pytest-mock>=3.10",
|
"pytest-mock",
|
||||||
"pytest-asyncio>=0.21",
|
"pytest-asyncio",
|
||||||
"duckdb>=0.9.0",
|
"duckdb",
|
||||||
"pytz>=2023.3",
|
"pytz",
|
||||||
"polars>=0.19, <=1.3.0",
|
"polars>=0.19, <=1.3.0",
|
||||||
"tantivy>=0.20.0",
|
"tantivy",
|
||||||
"pyarrow-stubs>=16.0",
|
"pyarrow-stubs",
|
||||||
"pylance>=4.0.0b7",
|
"pylance>=1.0.0b14,<3.0.0",
|
||||||
"requests>=2.31.0",
|
"requests",
|
||||||
"datafusion>=52,<53",
|
"datafusion<52",
|
||||||
]
|
]
|
||||||
dev = [
|
dev = [
|
||||||
"ruff>=0.3.0",
|
"ruff",
|
||||||
"pre-commit>=3.5.0",
|
"pre-commit",
|
||||||
"pyright>=1.1.350",
|
"pyright",
|
||||||
'typing-extensions>=4.0.0; python_version < "3.11"',
|
'typing-extensions>=4.0.0; python_version < "3.11"',
|
||||||
]
|
]
|
||||||
docs = ["mkdocs", "mkdocs-jupyter", "mkdocs-material", "mkdocstrings-python"]
|
docs = ["mkdocs", "mkdocs-jupyter", "mkdocs-material", "mkdocstrings-python"]
|
||||||
clip = ["torch", "pillow>=12.1.1", "open-clip-torch"]
|
clip = ["torch", "pillow", "open-clip-torch"]
|
||||||
siglip = ["torch", "pillow>=12.1.1", "transformers>=4.41.0","sentencepiece"]
|
siglip = ["torch", "pillow", "transformers>=4.41.0","sentencepiece"]
|
||||||
embeddings = [
|
embeddings = [
|
||||||
"requests>=2.31.0",
|
"requests>=2.31.0",
|
||||||
"openai>=1.6.1",
|
"openai>=1.6.1",
|
||||||
"sentence-transformers>=2.2.0",
|
"sentence-transformers",
|
||||||
"torch>=2.0.0",
|
"torch",
|
||||||
"pillow>=12.1.1",
|
"pillow",
|
||||||
"open-clip-torch>=2.20.0",
|
"open-clip-torch",
|
||||||
"cohere>=4.0",
|
"cohere",
|
||||||
"colpali-engine>=0.3.10",
|
"colpali-engine>=0.3.10",
|
||||||
"huggingface_hub>=0.19.0",
|
"huggingface_hub",
|
||||||
"InstructorEmbedding>=1.0.1",
|
"InstructorEmbedding",
|
||||||
"google.generativeai>=0.3.0",
|
"google.generativeai",
|
||||||
"boto3>=1.28.57",
|
"boto3>=1.28.57",
|
||||||
"awscli>=1.44.38",
|
"awscli>=1.29.57",
|
||||||
"botocore>=1.31.57",
|
"botocore>=1.31.57",
|
||||||
'ibm-watsonx-ai>=1.1.2; python_version >= "3.10"',
|
'ibm-watsonx-ai>=1.1.2; python_version >= "3.10"',
|
||||||
"ollama>=0.3.0",
|
"ollama>=0.3.0",
|
||||||
"sentencepiece>=0.1.99"
|
"sentencepiece"
|
||||||
]
|
]
|
||||||
azure = ["adlfs>=2024.2.0"]
|
azure = ["adlfs>=2024.2.0"]
|
||||||
|
|
||||||
|
|||||||
@@ -166,8 +166,6 @@ class Table:
|
|||||||
async def checkout(self, version: Union[int, str]): ...
|
async def checkout(self, version: Union[int, str]): ...
|
||||||
async def checkout_latest(self): ...
|
async def checkout_latest(self): ...
|
||||||
async def restore(self, version: Optional[Union[int, str]] = None): ...
|
async def restore(self, version: Optional[Union[int, str]] = None): ...
|
||||||
async def prewarm_index(self, index_name: str) -> None: ...
|
|
||||||
async def prewarm_data(self, columns: Optional[List[str]] = None) -> None: ...
|
|
||||||
async def list_indices(self) -> list[IndexConfig]: ...
|
async def list_indices(self) -> list[IndexConfig]: ...
|
||||||
async def delete(self, filter: str) -> DeleteResult: ...
|
async def delete(self, filter: str) -> DeleteResult: ...
|
||||||
async def add_columns(self, columns: list[tuple[str, str]]) -> AddColumnsResult: ...
|
async def add_columns(self, columns: list[tuple[str, str]]) -> AddColumnsResult: ...
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ from abc import abstractmethod
|
|||||||
from datetime import timedelta
|
from datetime import timedelta
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
import sys
|
import sys
|
||||||
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Literal, Optional, Union
|
from typing import TYPE_CHECKING, Dict, Iterable, List, Literal, Optional, Union
|
||||||
|
|
||||||
if sys.version_info >= (3, 12):
|
if sys.version_info >= (3, 12):
|
||||||
from typing import override
|
from typing import override
|
||||||
@@ -1541,8 +1541,6 @@ class AsyncConnection(object):
|
|||||||
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
||||||
index_cache_size: Optional[int] = None,
|
index_cache_size: Optional[int] = None,
|
||||||
location: Optional[str] = None,
|
location: Optional[str] = None,
|
||||||
namespace_client: Optional[Any] = None,
|
|
||||||
managed_versioning: Optional[bool] = None,
|
|
||||||
) -> AsyncTable:
|
) -> AsyncTable:
|
||||||
"""Open a Lance Table in the database.
|
"""Open a Lance Table in the database.
|
||||||
|
|
||||||
@@ -1575,9 +1573,6 @@ class AsyncConnection(object):
|
|||||||
The explicit location (URI) of the table. If provided, the table will be
|
The explicit location (URI) of the table. If provided, the table will be
|
||||||
opened from this location instead of deriving it from the database URI
|
opened from this location instead of deriving it from the database URI
|
||||||
and table name.
|
and table name.
|
||||||
managed_versioning: bool, optional
|
|
||||||
Whether managed versioning is enabled for this table. If provided,
|
|
||||||
avoids a redundant describe_table call when namespace_client is set.
|
|
||||||
|
|
||||||
Returns
|
Returns
|
||||||
-------
|
-------
|
||||||
@@ -1592,8 +1587,6 @@ class AsyncConnection(object):
|
|||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
location=location,
|
location=location,
|
||||||
namespace_client=namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
return AsyncTable(table)
|
return AsyncTable(table)
|
||||||
|
|
||||||
|
|||||||
@@ -34,6 +34,7 @@ class LanceMergeInsertBuilder(object):
|
|||||||
self._when_not_matched_by_source_condition = None
|
self._when_not_matched_by_source_condition = None
|
||||||
self._timeout = None
|
self._timeout = None
|
||||||
self._use_index = True
|
self._use_index = True
|
||||||
|
self._mem_wal = False
|
||||||
|
|
||||||
def when_matched_update_all(
|
def when_matched_update_all(
|
||||||
self, *, where: Optional[str] = None
|
self, *, where: Optional[str] = None
|
||||||
@@ -96,6 +97,47 @@ class LanceMergeInsertBuilder(object):
|
|||||||
self._use_index = use_index
|
self._use_index = use_index
|
||||||
return self
|
return self
|
||||||
|
|
||||||
|
def mem_wal(self, enabled: bool = True) -> LanceMergeInsertBuilder:
|
||||||
|
"""
|
||||||
|
Enable MemWAL (Memory Write-Ahead Log) mode for this merge insert operation.
|
||||||
|
|
||||||
|
When enabled, the merge insert will route data through a memory node service
|
||||||
|
that buffers writes before flushing to storage. This is only supported for
|
||||||
|
remote (LanceDB Cloud) tables.
|
||||||
|
|
||||||
|
**Important:** MemWAL only supports the upsert pattern. You must use:
|
||||||
|
- `when_matched_update_all()` (without a filter condition)
|
||||||
|
- `when_not_matched_insert_all()`
|
||||||
|
|
||||||
|
MemWAL does NOT support:
|
||||||
|
- `when_matched_update_all(where=...)` with a filter condition
|
||||||
|
- `when_not_matched_by_source_delete()`
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
enabled: bool
|
||||||
|
Whether to enable MemWAL mode. Defaults to `True`.
|
||||||
|
|
||||||
|
Raises
|
||||||
|
------
|
||||||
|
NotImplementedError
|
||||||
|
If used on a native (local) table, as MemWAL is only supported for
|
||||||
|
remote tables.
|
||||||
|
ValueError
|
||||||
|
If the merge insert pattern is not supported by MemWAL.
|
||||||
|
|
||||||
|
Examples
|
||||||
|
--------
|
||||||
|
>>> # Correct usage with MemWAL
|
||||||
|
>>> table.merge_insert(["id"]) \\
|
||||||
|
... .when_matched_update_all() \\
|
||||||
|
... .when_not_matched_insert_all() \\
|
||||||
|
... .mem_wal() \\
|
||||||
|
... .execute(new_data)
|
||||||
|
"""
|
||||||
|
self._mem_wal = enabled
|
||||||
|
return self
|
||||||
|
|
||||||
def execute(
|
def execute(
|
||||||
self,
|
self,
|
||||||
new_data: DATA,
|
new_data: DATA,
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ from __future__ import annotations
|
|||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
import sys
|
import sys
|
||||||
from typing import Any, Dict, Iterable, List, Optional, Union
|
from typing import Dict, Iterable, List, Optional, Union
|
||||||
|
|
||||||
if sys.version_info >= (3, 12):
|
if sys.version_info >= (3, 12):
|
||||||
from typing import override
|
from typing import override
|
||||||
@@ -240,7 +240,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
session : Optional[Session]
|
session : Optional[Session]
|
||||||
A session to use for this connection
|
A session to use for this connection
|
||||||
"""
|
"""
|
||||||
self._namespace_client = namespace
|
self._ns = namespace
|
||||||
self.read_consistency_interval = read_consistency_interval
|
self.read_consistency_interval = read_consistency_interval
|
||||||
self.storage_options = storage_options or {}
|
self.storage_options = storage_options or {}
|
||||||
self.session = session
|
self.session = session
|
||||||
@@ -269,7 +269,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
if namespace is None:
|
if namespace is None:
|
||||||
namespace = []
|
namespace = []
|
||||||
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
||||||
response = self._namespace_client.list_tables(request)
|
response = self._ns.list_tables(request)
|
||||||
return response.tables if response.tables else []
|
return response.tables if response.tables else []
|
||||||
|
|
||||||
@override
|
@override
|
||||||
@@ -309,9 +309,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
# Try to describe the table first to see if it exists
|
# Try to describe the table first to see if it exists
|
||||||
try:
|
try:
|
||||||
describe_request = DescribeTableRequest(id=table_id)
|
describe_request = DescribeTableRequest(id=table_id)
|
||||||
describe_response = self._namespace_client.describe_table(
|
describe_response = self._ns.describe_table(describe_request)
|
||||||
describe_request
|
|
||||||
)
|
|
||||||
location = describe_response.location
|
location = describe_response.location
|
||||||
namespace_storage_options = describe_response.storage_options
|
namespace_storage_options = describe_response.storage_options
|
||||||
except Exception:
|
except Exception:
|
||||||
@@ -325,7 +323,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
location=None,
|
location=None,
|
||||||
properties=self.storage_options if self.storage_options else None,
|
properties=self.storage_options if self.storage_options else None,
|
||||||
)
|
)
|
||||||
declare_response = self._namespace_client.declare_table(declare_request)
|
declare_response = self._ns.declare_table(declare_request)
|
||||||
|
|
||||||
if not declare_response.location:
|
if not declare_response.location:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
@@ -355,7 +353,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
# Only create if namespace returned storage_options (not None)
|
# Only create if namespace returned storage_options (not None)
|
||||||
if storage_options_provider is None and namespace_storage_options is not None:
|
if storage_options_provider is None and namespace_storage_options is not None:
|
||||||
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
||||||
namespace=self._namespace_client,
|
namespace=self._ns,
|
||||||
table_id=table_id,
|
table_id=table_id,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -373,7 +371,6 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
storage_options=merged_storage_options,
|
storage_options=merged_storage_options,
|
||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
location=location,
|
location=location,
|
||||||
namespace_client=self._namespace_client,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
return tbl
|
return tbl
|
||||||
@@ -392,7 +389,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
namespace = []
|
namespace = []
|
||||||
table_id = namespace + [name]
|
table_id = namespace + [name]
|
||||||
request = DescribeTableRequest(id=table_id)
|
request = DescribeTableRequest(id=table_id)
|
||||||
response = self._namespace_client.describe_table(request)
|
response = self._ns.describe_table(request)
|
||||||
|
|
||||||
# Merge storage options: self.storage_options < user options < namespace options
|
# Merge storage options: self.storage_options < user options < namespace options
|
||||||
merged_storage_options = dict(self.storage_options)
|
merged_storage_options = dict(self.storage_options)
|
||||||
@@ -405,14 +402,10 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
# Only create if namespace returned storage_options (not None)
|
# Only create if namespace returned storage_options (not None)
|
||||||
if storage_options_provider is None and response.storage_options is not None:
|
if storage_options_provider is None and response.storage_options is not None:
|
||||||
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
||||||
namespace=self._namespace_client,
|
namespace=self._ns,
|
||||||
table_id=table_id,
|
table_id=table_id,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Pass managed_versioning to avoid redundant describe_table call in Rust.
|
|
||||||
# Convert None to False since we already have the answer from describe_table.
|
|
||||||
managed_versioning = response.managed_versioning is True
|
|
||||||
|
|
||||||
return self._lance_table_from_uri(
|
return self._lance_table_from_uri(
|
||||||
name,
|
name,
|
||||||
response.location,
|
response.location,
|
||||||
@@ -420,8 +413,6 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
storage_options=merged_storage_options,
|
storage_options=merged_storage_options,
|
||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
namespace_client=self._namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
@override
|
@override
|
||||||
@@ -431,7 +422,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
namespace = []
|
namespace = []
|
||||||
table_id = namespace + [name]
|
table_id = namespace + [name]
|
||||||
request = DropTableRequest(id=table_id)
|
request = DropTableRequest(id=table_id)
|
||||||
self._namespace_client.drop_table(request)
|
self._ns.drop_table(request)
|
||||||
|
|
||||||
@override
|
@override
|
||||||
def rename_table(
|
def rename_table(
|
||||||
@@ -493,7 +484,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
request = ListNamespacesRequest(
|
request = ListNamespacesRequest(
|
||||||
id=namespace, page_token=page_token, limit=limit
|
id=namespace, page_token=page_token, limit=limit
|
||||||
)
|
)
|
||||||
response = self._namespace_client.list_namespaces(request)
|
response = self._ns.list_namespaces(request)
|
||||||
return ListNamespacesResponse(
|
return ListNamespacesResponse(
|
||||||
namespaces=response.namespaces if response.namespaces else [],
|
namespaces=response.namespaces if response.namespaces else [],
|
||||||
page_token=response.page_token,
|
page_token=response.page_token,
|
||||||
@@ -529,7 +520,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
mode=_normalize_create_namespace_mode(mode),
|
mode=_normalize_create_namespace_mode(mode),
|
||||||
properties=properties,
|
properties=properties,
|
||||||
)
|
)
|
||||||
response = self._namespace_client.create_namespace(request)
|
response = self._ns.create_namespace(request)
|
||||||
return CreateNamespaceResponse(
|
return CreateNamespaceResponse(
|
||||||
properties=response.properties if hasattr(response, "properties") else None
|
properties=response.properties if hasattr(response, "properties") else None
|
||||||
)
|
)
|
||||||
@@ -564,7 +555,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
mode=_normalize_drop_namespace_mode(mode),
|
mode=_normalize_drop_namespace_mode(mode),
|
||||||
behavior=_normalize_drop_namespace_behavior(behavior),
|
behavior=_normalize_drop_namespace_behavior(behavior),
|
||||||
)
|
)
|
||||||
response = self._namespace_client.drop_namespace(request)
|
response = self._ns.drop_namespace(request)
|
||||||
return DropNamespaceResponse(
|
return DropNamespaceResponse(
|
||||||
properties=(
|
properties=(
|
||||||
response.properties if hasattr(response, "properties") else None
|
response.properties if hasattr(response, "properties") else None
|
||||||
@@ -590,7 +581,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
Response containing the namespace properties.
|
Response containing the namespace properties.
|
||||||
"""
|
"""
|
||||||
request = DescribeNamespaceRequest(id=namespace)
|
request = DescribeNamespaceRequest(id=namespace)
|
||||||
response = self._namespace_client.describe_namespace(request)
|
response = self._ns.describe_namespace(request)
|
||||||
return DescribeNamespaceResponse(
|
return DescribeNamespaceResponse(
|
||||||
properties=response.properties if hasattr(response, "properties") else None
|
properties=response.properties if hasattr(response, "properties") else None
|
||||||
)
|
)
|
||||||
@@ -624,7 +615,7 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
if namespace is None:
|
if namespace is None:
|
||||||
namespace = []
|
namespace = []
|
||||||
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
||||||
response = self._namespace_client.list_tables(request)
|
response = self._ns.list_tables(request)
|
||||||
return ListTablesResponse(
|
return ListTablesResponse(
|
||||||
tables=response.tables if response.tables else [],
|
tables=response.tables if response.tables else [],
|
||||||
page_token=response.page_token,
|
page_token=response.page_token,
|
||||||
@@ -639,8 +630,6 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
storage_options: Optional[Dict[str, str]] = None,
|
storage_options: Optional[Dict[str, str]] = None,
|
||||||
storage_options_provider: Optional[StorageOptionsProvider] = None,
|
storage_options_provider: Optional[StorageOptionsProvider] = None,
|
||||||
index_cache_size: Optional[int] = None,
|
index_cache_size: Optional[int] = None,
|
||||||
namespace_client: Optional[Any] = None,
|
|
||||||
managed_versioning: Optional[bool] = None,
|
|
||||||
) -> LanceTable:
|
) -> LanceTable:
|
||||||
# Open a table directly from a URI using the location parameter
|
# Open a table directly from a URI using the location parameter
|
||||||
# Note: storage_options should already be merged by the caller
|
# Note: storage_options should already be merged by the caller
|
||||||
@@ -654,8 +643,6 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Open the table using the temporary connection with the location parameter
|
# Open the table using the temporary connection with the location parameter
|
||||||
# Pass namespace_client to enable managed versioning support
|
|
||||||
# Pass managed_versioning to avoid redundant describe_table call
|
|
||||||
return LanceTable.open(
|
return LanceTable.open(
|
||||||
temp_conn,
|
temp_conn,
|
||||||
name,
|
name,
|
||||||
@@ -664,8 +651,6 @@ class LanceNamespaceDBConnection(DBConnection):
|
|||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
location=table_uri,
|
location=table_uri,
|
||||||
namespace_client=namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@@ -700,7 +685,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
session : Optional[Session]
|
session : Optional[Session]
|
||||||
A session to use for this connection
|
A session to use for this connection
|
||||||
"""
|
"""
|
||||||
self._namespace_client = namespace
|
self._ns = namespace
|
||||||
self.read_consistency_interval = read_consistency_interval
|
self.read_consistency_interval = read_consistency_interval
|
||||||
self.storage_options = storage_options or {}
|
self.storage_options = storage_options or {}
|
||||||
self.session = session
|
self.session = session
|
||||||
@@ -728,7 +713,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
if namespace is None:
|
if namespace is None:
|
||||||
namespace = []
|
namespace = []
|
||||||
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
||||||
response = self._namespace_client.list_tables(request)
|
response = self._ns.list_tables(request)
|
||||||
return response.tables if response.tables else []
|
return response.tables if response.tables else []
|
||||||
|
|
||||||
async def create_table(
|
async def create_table(
|
||||||
@@ -765,9 +750,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
# Try to describe the table first to see if it exists
|
# Try to describe the table first to see if it exists
|
||||||
try:
|
try:
|
||||||
describe_request = DescribeTableRequest(id=table_id)
|
describe_request = DescribeTableRequest(id=table_id)
|
||||||
describe_response = self._namespace_client.describe_table(
|
describe_response = self._ns.describe_table(describe_request)
|
||||||
describe_request
|
|
||||||
)
|
|
||||||
location = describe_response.location
|
location = describe_response.location
|
||||||
namespace_storage_options = describe_response.storage_options
|
namespace_storage_options = describe_response.storage_options
|
||||||
except Exception:
|
except Exception:
|
||||||
@@ -781,7 +764,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
location=None,
|
location=None,
|
||||||
properties=self.storage_options if self.storage_options else None,
|
properties=self.storage_options if self.storage_options else None,
|
||||||
)
|
)
|
||||||
declare_response = self._namespace_client.declare_table(declare_request)
|
declare_response = self._ns.declare_table(declare_request)
|
||||||
|
|
||||||
if not declare_response.location:
|
if not declare_response.location:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
@@ -814,7 +797,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
and namespace_storage_options is not None
|
and namespace_storage_options is not None
|
||||||
):
|
):
|
||||||
provider = LanceNamespaceStorageOptionsProvider(
|
provider = LanceNamespaceStorageOptionsProvider(
|
||||||
namespace=self._namespace_client,
|
namespace=self._ns,
|
||||||
table_id=table_id,
|
table_id=table_id,
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
@@ -834,7 +817,6 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
storage_options=merged_storage_options,
|
storage_options=merged_storage_options,
|
||||||
storage_options_provider=provider,
|
storage_options_provider=provider,
|
||||||
location=location,
|
location=location,
|
||||||
namespace_client=self._namespace_client,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
lance_table = await asyncio.to_thread(_create_table)
|
lance_table = await asyncio.to_thread(_create_table)
|
||||||
@@ -855,7 +837,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
namespace = []
|
namespace = []
|
||||||
table_id = namespace + [name]
|
table_id = namespace + [name]
|
||||||
request = DescribeTableRequest(id=table_id)
|
request = DescribeTableRequest(id=table_id)
|
||||||
response = self._namespace_client.describe_table(request)
|
response = self._ns.describe_table(request)
|
||||||
|
|
||||||
# Merge storage options: self.storage_options < user options < namespace options
|
# Merge storage options: self.storage_options < user options < namespace options
|
||||||
merged_storage_options = dict(self.storage_options)
|
merged_storage_options = dict(self.storage_options)
|
||||||
@@ -867,14 +849,10 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
# Create a storage options provider if not provided by user
|
# Create a storage options provider if not provided by user
|
||||||
if storage_options_provider is None and response.storage_options is not None:
|
if storage_options_provider is None and response.storage_options is not None:
|
||||||
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
storage_options_provider = LanceNamespaceStorageOptionsProvider(
|
||||||
namespace=self._namespace_client,
|
namespace=self._ns,
|
||||||
table_id=table_id,
|
table_id=table_id,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Capture managed_versioning from describe response.
|
|
||||||
# Convert None to False since we already have the answer from describe_table.
|
|
||||||
managed_versioning = response.managed_versioning is True
|
|
||||||
|
|
||||||
# Open table in a thread
|
# Open table in a thread
|
||||||
def _open_table():
|
def _open_table():
|
||||||
temp_conn = LanceDBConnection(
|
temp_conn = LanceDBConnection(
|
||||||
@@ -892,8 +870,6 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
location=response.location,
|
location=response.location,
|
||||||
namespace_client=self._namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
lance_table = await asyncio.to_thread(_open_table)
|
lance_table = await asyncio.to_thread(_open_table)
|
||||||
@@ -905,7 +881,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
namespace = []
|
namespace = []
|
||||||
table_id = namespace + [name]
|
table_id = namespace + [name]
|
||||||
request = DropTableRequest(id=table_id)
|
request = DropTableRequest(id=table_id)
|
||||||
self._namespace_client.drop_table(request)
|
self._ns.drop_table(request)
|
||||||
|
|
||||||
async def rename_table(
|
async def rename_table(
|
||||||
self,
|
self,
|
||||||
@@ -967,7 +943,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
request = ListNamespacesRequest(
|
request = ListNamespacesRequest(
|
||||||
id=namespace, page_token=page_token, limit=limit
|
id=namespace, page_token=page_token, limit=limit
|
||||||
)
|
)
|
||||||
response = self._namespace_client.list_namespaces(request)
|
response = self._ns.list_namespaces(request)
|
||||||
return ListNamespacesResponse(
|
return ListNamespacesResponse(
|
||||||
namespaces=response.namespaces if response.namespaces else [],
|
namespaces=response.namespaces if response.namespaces else [],
|
||||||
page_token=response.page_token,
|
page_token=response.page_token,
|
||||||
@@ -1002,7 +978,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
mode=_normalize_create_namespace_mode(mode),
|
mode=_normalize_create_namespace_mode(mode),
|
||||||
properties=properties,
|
properties=properties,
|
||||||
)
|
)
|
||||||
response = self._namespace_client.create_namespace(request)
|
response = self._ns.create_namespace(request)
|
||||||
return CreateNamespaceResponse(
|
return CreateNamespaceResponse(
|
||||||
properties=response.properties if hasattr(response, "properties") else None
|
properties=response.properties if hasattr(response, "properties") else None
|
||||||
)
|
)
|
||||||
@@ -1036,7 +1012,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
mode=_normalize_drop_namespace_mode(mode),
|
mode=_normalize_drop_namespace_mode(mode),
|
||||||
behavior=_normalize_drop_namespace_behavior(behavior),
|
behavior=_normalize_drop_namespace_behavior(behavior),
|
||||||
)
|
)
|
||||||
response = self._namespace_client.drop_namespace(request)
|
response = self._ns.drop_namespace(request)
|
||||||
return DropNamespaceResponse(
|
return DropNamespaceResponse(
|
||||||
properties=(
|
properties=(
|
||||||
response.properties if hasattr(response, "properties") else None
|
response.properties if hasattr(response, "properties") else None
|
||||||
@@ -1063,7 +1039,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
Response containing the namespace properties.
|
Response containing the namespace properties.
|
||||||
"""
|
"""
|
||||||
request = DescribeNamespaceRequest(id=namespace)
|
request = DescribeNamespaceRequest(id=namespace)
|
||||||
response = self._namespace_client.describe_namespace(request)
|
response = self._ns.describe_namespace(request)
|
||||||
return DescribeNamespaceResponse(
|
return DescribeNamespaceResponse(
|
||||||
properties=response.properties if hasattr(response, "properties") else None
|
properties=response.properties if hasattr(response, "properties") else None
|
||||||
)
|
)
|
||||||
@@ -1096,7 +1072,7 @@ class AsyncLanceNamespaceDBConnection:
|
|||||||
if namespace is None:
|
if namespace is None:
|
||||||
namespace = []
|
namespace = []
|
||||||
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
request = ListTablesRequest(id=namespace, page_token=page_token, limit=limit)
|
||||||
response = self._namespace_client.list_tables(request)
|
response = self._ns.list_tables(request)
|
||||||
return ListTablesResponse(
|
return ListTablesResponse(
|
||||||
tables=response.tables if response.tables else [],
|
tables=response.tables if response.tables else [],
|
||||||
page_token=response.page_token,
|
page_token=response.page_token,
|
||||||
|
|||||||
@@ -2205,8 +2205,8 @@ class LanceHybridQueryBuilder(LanceQueryBuilder):
|
|||||||
self._vector_query.select(self._columns)
|
self._vector_query.select(self._columns)
|
||||||
self._fts_query.select(self._columns)
|
self._fts_query.select(self._columns)
|
||||||
if self._where:
|
if self._where:
|
||||||
self._vector_query.where(self._where, not self._postfilter)
|
self._vector_query.where(self._where, self._postfilter)
|
||||||
self._fts_query.where(self._where, not self._postfilter)
|
self._fts_query.where(self._where, self._postfilter)
|
||||||
if self._with_row_id:
|
if self._with_row_id:
|
||||||
self._vector_query.with_row_id(True)
|
self._vector_query.with_row_id(True)
|
||||||
self._fts_query.with_row_id(True)
|
self._fts_query.with_row_id(True)
|
||||||
|
|||||||
@@ -640,45 +640,6 @@ class RemoteTable(Table):
|
|||||||
def drop_index(self, index_name: str):
|
def drop_index(self, index_name: str):
|
||||||
return LOOP.run(self._table.drop_index(index_name))
|
return LOOP.run(self._table.drop_index(index_name))
|
||||||
|
|
||||||
def prewarm_index(self, name: str) -> None:
|
|
||||||
"""Prewarm an index in the table.
|
|
||||||
|
|
||||||
This is a hint to the database that the index will be accessed in the
|
|
||||||
future and should be loaded into memory if possible. This can reduce
|
|
||||||
cold-start latency for subsequent queries.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
name: str
|
|
||||||
The name of the index to prewarm
|
|
||||||
"""
|
|
||||||
return LOOP.run(self._table.prewarm_index(name))
|
|
||||||
|
|
||||||
def prewarm_data(self, columns: Optional[List[str]] = None) -> None:
|
|
||||||
"""Prewarm data for the table.
|
|
||||||
|
|
||||||
This is a hint to the database that the given columns will be accessed
|
|
||||||
in the future and the database should prefetch the data if possible.
|
|
||||||
Currently only supported on remote tables.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
|
|
||||||
This operation has a large upfront cost but can speed up future queries
|
|
||||||
that need to fetch the given columns. Large columns such as embeddings
|
|
||||||
or binary data may not be practical to prewarm. This feature is intended
|
|
||||||
for workloads that issue many queries against the same columns.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
columns: list of str, optional
|
|
||||||
The columns to prewarm. If None, all columns are prewarmed.
|
|
||||||
"""
|
|
||||||
return LOOP.run(self._table.prewarm_data(columns))
|
|
||||||
|
|
||||||
def wait_for_index(
|
def wait_for_index(
|
||||||
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
||||||
):
|
):
|
||||||
|
|||||||
@@ -1506,17 +1506,22 @@ class Table(ABC):
|
|||||||
in-progress operation (e.g. appending new data) and these files will not
|
in-progress operation (e.g. appending new data) and these files will not
|
||||||
be deleted unless they are at least 7 days old. If delete_unverified is True
|
be deleted unless they are at least 7 days old. If delete_unverified is True
|
||||||
then these files will be deleted regardless of their age.
|
then these files will be deleted regardless of their age.
|
||||||
|
|
||||||
.. warning::
|
|
||||||
|
|
||||||
This should only be set to True if you can guarantee that no other
|
|
||||||
process is currently working on this dataset. Otherwise the dataset
|
|
||||||
could be put into a corrupted state.
|
|
||||||
|
|
||||||
retrain: bool, default False
|
retrain: bool, default False
|
||||||
This parameter is no longer used and is deprecated.
|
This parameter is no longer used and is deprecated.
|
||||||
|
|
||||||
The frequency an application should call optimize is based on the frequency of
|
Experimental API
|
||||||
|
----------------
|
||||||
|
|
||||||
|
The optimization process is undergoing active development and may change.
|
||||||
|
Our goal with these changes is to improve the performance of optimization and
|
||||||
|
reduce the complexity.
|
||||||
|
|
||||||
|
That being said, it is essential today to run optimize if you want the best
|
||||||
|
performance. It should be stable and safe to use in production, but it our
|
||||||
|
hope that the API may be simplified (or not even need to be called) in the
|
||||||
|
future.
|
||||||
|
|
||||||
|
The frequency an application shoudl call optimize is based on the frequency of
|
||||||
data modifications. If data is frequently added, deleted, or updated then
|
data modifications. If data is frequently added, deleted, or updated then
|
||||||
optimize should be run frequently. A good rule of thumb is to run optimize if
|
optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
you have added or modified 100,000 or more records or run more than 20 data
|
you have added or modified 100,000 or more records or run more than 20 data
|
||||||
@@ -1741,8 +1746,6 @@ class LanceTable(Table):
|
|||||||
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
||||||
index_cache_size: Optional[int] = None,
|
index_cache_size: Optional[int] = None,
|
||||||
location: Optional[str] = None,
|
location: Optional[str] = None,
|
||||||
namespace_client: Optional[Any] = None,
|
|
||||||
managed_versioning: Optional[bool] = None,
|
|
||||||
_async: AsyncTable = None,
|
_async: AsyncTable = None,
|
||||||
):
|
):
|
||||||
if namespace is None:
|
if namespace is None:
|
||||||
@@ -1750,7 +1753,6 @@ class LanceTable(Table):
|
|||||||
self._conn = connection
|
self._conn = connection
|
||||||
self._namespace = namespace
|
self._namespace = namespace
|
||||||
self._location = location # Store location for use in _dataset_path
|
self._location = location # Store location for use in _dataset_path
|
||||||
self._namespace_client = namespace_client
|
|
||||||
if _async is not None:
|
if _async is not None:
|
||||||
self._table = _async
|
self._table = _async
|
||||||
else:
|
else:
|
||||||
@@ -1762,8 +1764,6 @@ class LanceTable(Table):
|
|||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
location=location,
|
location=location,
|
||||||
namespace_client=namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -1806,8 +1806,6 @@ class LanceTable(Table):
|
|||||||
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
storage_options_provider: Optional["StorageOptionsProvider"] = None,
|
||||||
index_cache_size: Optional[int] = None,
|
index_cache_size: Optional[int] = None,
|
||||||
location: Optional[str] = None,
|
location: Optional[str] = None,
|
||||||
namespace_client: Optional[Any] = None,
|
|
||||||
managed_versioning: Optional[bool] = None,
|
|
||||||
):
|
):
|
||||||
if namespace is None:
|
if namespace is None:
|
||||||
namespace = []
|
namespace = []
|
||||||
@@ -1819,8 +1817,6 @@ class LanceTable(Table):
|
|||||||
storage_options_provider=storage_options_provider,
|
storage_options_provider=storage_options_provider,
|
||||||
index_cache_size=index_cache_size,
|
index_cache_size=index_cache_size,
|
||||||
location=location,
|
location=location,
|
||||||
namespace_client=namespace_client,
|
|
||||||
managed_versioning=managed_versioning,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# check the dataset exists
|
# check the dataset exists
|
||||||
@@ -1852,16 +1848,6 @@ class LanceTable(Table):
|
|||||||
"Please install with `pip install pylance`."
|
"Please install with `pip install pylance`."
|
||||||
)
|
)
|
||||||
|
|
||||||
if self._namespace_client is not None:
|
|
||||||
table_id = self._namespace + [self.name]
|
|
||||||
return lance.dataset(
|
|
||||||
version=self.version,
|
|
||||||
storage_options=self._conn.storage_options,
|
|
||||||
namespace=self._namespace_client,
|
|
||||||
table_id=table_id,
|
|
||||||
**kwargs,
|
|
||||||
)
|
|
||||||
|
|
||||||
return lance.dataset(
|
return lance.dataset(
|
||||||
self._dataset_path,
|
self._dataset_path,
|
||||||
version=self.version,
|
version=self.version,
|
||||||
@@ -2214,18 +2200,12 @@ class LanceTable(Table):
|
|||||||
|
|
||||||
def prewarm_index(self, name: str) -> None:
|
def prewarm_index(self, name: str) -> None:
|
||||||
"""
|
"""
|
||||||
Prewarm an index in the table.
|
Prewarms an index in the table
|
||||||
|
|
||||||
This is a hint to the database that the index will be accessed in the
|
This loads the entire index into memory
|
||||||
future and should be loaded into memory if possible. This can reduce
|
|
||||||
cold-start latency for subsequent queries.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
If the index does not fit into the available cache this call
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
may be wasteful
|
||||||
|
|
||||||
It is generally wasteful to call this if the index does not fit into the
|
|
||||||
available cache. Not all index types support prewarming; unsupported
|
|
||||||
indices will silently ignore the request.
|
|
||||||
|
|
||||||
Parameters
|
Parameters
|
||||||
----------
|
----------
|
||||||
@@ -2234,29 +2214,6 @@ class LanceTable(Table):
|
|||||||
"""
|
"""
|
||||||
return LOOP.run(self._table.prewarm_index(name))
|
return LOOP.run(self._table.prewarm_index(name))
|
||||||
|
|
||||||
def prewarm_data(self, columns: Optional[List[str]] = None) -> None:
|
|
||||||
"""
|
|
||||||
Prewarm data for the table.
|
|
||||||
|
|
||||||
This is a hint to the database that the given columns will be accessed
|
|
||||||
in the future and the database should prefetch the data if possible.
|
|
||||||
Currently only supported on remote tables.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
|
|
||||||
This operation has a large upfront cost but can speed up future queries
|
|
||||||
that need to fetch the given columns. Large columns such as embeddings
|
|
||||||
or binary data may not be practical to prewarm. This feature is intended
|
|
||||||
for workloads that issue many queries against the same columns.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
columns: list of str, optional
|
|
||||||
The columns to prewarm. If None, all columns are prewarmed.
|
|
||||||
"""
|
|
||||||
return LOOP.run(self._table.prewarm_data(columns))
|
|
||||||
|
|
||||||
def wait_for_index(
|
def wait_for_index(
|
||||||
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
||||||
) -> None:
|
) -> None:
|
||||||
@@ -2756,7 +2713,6 @@ class LanceTable(Table):
|
|||||||
data_storage_version: Optional[str] = None,
|
data_storage_version: Optional[str] = None,
|
||||||
enable_v2_manifest_paths: Optional[bool] = None,
|
enable_v2_manifest_paths: Optional[bool] = None,
|
||||||
location: Optional[str] = None,
|
location: Optional[str] = None,
|
||||||
namespace_client: Optional[Any] = None,
|
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Create a new table.
|
Create a new table.
|
||||||
@@ -2817,7 +2773,6 @@ class LanceTable(Table):
|
|||||||
self._conn = db
|
self._conn = db
|
||||||
self._namespace = namespace
|
self._namespace = namespace
|
||||||
self._location = location
|
self._location = location
|
||||||
self._namespace_client = namespace_client
|
|
||||||
|
|
||||||
if data_storage_version is not None:
|
if data_storage_version is not None:
|
||||||
warnings.warn(
|
warnings.warn(
|
||||||
@@ -3042,17 +2997,22 @@ class LanceTable(Table):
|
|||||||
in-progress operation (e.g. appending new data) and these files will not
|
in-progress operation (e.g. appending new data) and these files will not
|
||||||
be deleted unless they are at least 7 days old. If delete_unverified is True
|
be deleted unless they are at least 7 days old. If delete_unverified is True
|
||||||
then these files will be deleted regardless of their age.
|
then these files will be deleted regardless of their age.
|
||||||
|
|
||||||
.. warning::
|
|
||||||
|
|
||||||
This should only be set to True if you can guarantee that no other
|
|
||||||
process is currently working on this dataset. Otherwise the dataset
|
|
||||||
could be put into a corrupted state.
|
|
||||||
|
|
||||||
retrain: bool, default False
|
retrain: bool, default False
|
||||||
This parameter is no longer used and is deprecated.
|
This parameter is no longer used and is deprecated.
|
||||||
|
|
||||||
The frequency an application should call optimize is based on the frequency of
|
Experimental API
|
||||||
|
----------------
|
||||||
|
|
||||||
|
The optimization process is undergoing active development and may change.
|
||||||
|
Our goal with these changes is to improve the performance of optimization and
|
||||||
|
reduce the complexity.
|
||||||
|
|
||||||
|
That being said, it is essential today to run optimize if you want the best
|
||||||
|
performance. It should be stable and safe to use in production, but it our
|
||||||
|
hope that the API may be simplified (or not even need to be called) in the
|
||||||
|
future.
|
||||||
|
|
||||||
|
The frequency an application shoudl call optimize is based on the frequency of
|
||||||
data modifications. If data is frequently added, deleted, or updated then
|
data modifications. If data is frequently added, deleted, or updated then
|
||||||
optimize should be run frequently. A good rule of thumb is to run optimize if
|
optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
you have added or modified 100,000 or more records or run more than 20 data
|
you have added or modified 100,000 or more records or run more than 20 data
|
||||||
@@ -3653,47 +3613,19 @@ class AsyncTable:
|
|||||||
"""
|
"""
|
||||||
Prewarm an index in the table.
|
Prewarm an index in the table.
|
||||||
|
|
||||||
This is a hint to the database that the index will be accessed in the
|
|
||||||
future and should be loaded into memory if possible. This can reduce
|
|
||||||
cold-start latency for subsequent queries.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
|
|
||||||
It is generally wasteful to call this if the index does not fit into the
|
|
||||||
available cache. Not all index types support prewarming; unsupported
|
|
||||||
indices will silently ignore the request.
|
|
||||||
|
|
||||||
Parameters
|
Parameters
|
||||||
----------
|
----------
|
||||||
name: str
|
name: str
|
||||||
The name of the index to prewarm
|
The name of the index to prewarm
|
||||||
|
|
||||||
|
Notes
|
||||||
|
-----
|
||||||
|
This will load the index into memory. This may reduce the cold-start time for
|
||||||
|
future queries. If the index does not fit in the cache then this call may be
|
||||||
|
wasteful.
|
||||||
"""
|
"""
|
||||||
await self._inner.prewarm_index(name)
|
await self._inner.prewarm_index(name)
|
||||||
|
|
||||||
async def prewarm_data(self, columns: Optional[List[str]] = None) -> None:
|
|
||||||
"""
|
|
||||||
Prewarm data for the table.
|
|
||||||
|
|
||||||
This is a hint to the database that the given columns will be accessed
|
|
||||||
in the future and the database should prefetch the data if possible.
|
|
||||||
Currently only supported on remote tables.
|
|
||||||
|
|
||||||
This call initiates prewarming and returns once the request is accepted.
|
|
||||||
It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
|
|
||||||
This operation has a large upfront cost but can speed up future queries
|
|
||||||
that need to fetch the given columns. Large columns such as embeddings
|
|
||||||
or binary data may not be practical to prewarm. This feature is intended
|
|
||||||
for workloads that issue many queries against the same columns.
|
|
||||||
|
|
||||||
Parameters
|
|
||||||
----------
|
|
||||||
columns: list of str, optional
|
|
||||||
The columns to prewarm. If None, all columns are prewarmed.
|
|
||||||
"""
|
|
||||||
await self._inner.prewarm_data(columns)
|
|
||||||
|
|
||||||
async def wait_for_index(
|
async def wait_for_index(
|
||||||
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
self, index_names: Iterable[str], timeout: timedelta = timedelta(seconds=300)
|
||||||
) -> None:
|
) -> None:
|
||||||
@@ -4249,6 +4181,7 @@ class AsyncTable:
|
|||||||
when_not_matched_by_source_condition=merge._when_not_matched_by_source_condition,
|
when_not_matched_by_source_condition=merge._when_not_matched_by_source_condition,
|
||||||
timeout=merge._timeout,
|
timeout=merge._timeout,
|
||||||
use_index=merge._use_index,
|
use_index=merge._use_index,
|
||||||
|
mem_wal=merge._mem_wal,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -4620,17 +4553,22 @@ class AsyncTable:
|
|||||||
in-progress operation (e.g. appending new data) and these files will not
|
in-progress operation (e.g. appending new data) and these files will not
|
||||||
be deleted unless they are at least 7 days old. If delete_unverified is True
|
be deleted unless they are at least 7 days old. If delete_unverified is True
|
||||||
then these files will be deleted regardless of their age.
|
then these files will be deleted regardless of their age.
|
||||||
|
|
||||||
.. warning::
|
|
||||||
|
|
||||||
This should only be set to True if you can guarantee that no other
|
|
||||||
process is currently working on this dataset. Otherwise the dataset
|
|
||||||
could be put into a corrupted state.
|
|
||||||
|
|
||||||
retrain: bool, default False
|
retrain: bool, default False
|
||||||
This parameter is no longer used and is deprecated.
|
This parameter is no longer used and is deprecated.
|
||||||
|
|
||||||
The frequency an application should call optimize is based on the frequency of
|
Experimental API
|
||||||
|
----------------
|
||||||
|
|
||||||
|
The optimization process is undergoing active development and may change.
|
||||||
|
Our goal with these changes is to improve the performance of optimization and
|
||||||
|
reduce the complexity.
|
||||||
|
|
||||||
|
That being said, it is essential today to run optimize if you want the best
|
||||||
|
performance. It should be stable and safe to use in production, but it our
|
||||||
|
hope that the API may be simplified (or not even need to be called) in the
|
||||||
|
future.
|
||||||
|
|
||||||
|
The frequency an application shoudl call optimize is based on the frequency of
|
||||||
data modifications. If data is frequently added, deleted, or updated then
|
data modifications. If data is frequently added, deleted, or updated then
|
||||||
optimize should be run frequently. A good rule of thumb is to run optimize if
|
optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
you have added or modified 100,000 or more records or run more than 20 data
|
you have added or modified 100,000 or more records or run more than 20 data
|
||||||
|
|||||||
@@ -177,60 +177,6 @@ async def test_analyze_plan(table: AsyncTable):
|
|||||||
assert "metrics=" in res
|
assert "metrics=" in res
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def table_with_id(tmpdir_factory) -> Table:
|
|
||||||
tmp_path = str(tmpdir_factory.mktemp("data"))
|
|
||||||
db = lancedb.connect(tmp_path)
|
|
||||||
data = pa.table(
|
|
||||||
{
|
|
||||||
"id": pa.array([1, 2, 3, 4], type=pa.int64()),
|
|
||||||
"text": pa.array(["a", "b", "cat", "dog"]),
|
|
||||||
"vector": pa.array(
|
|
||||||
[[0.1, 0.1], [2, 2], [-0.1, -0.1], [0.5, -0.5]],
|
|
||||||
type=pa.list_(pa.float32(), list_size=2),
|
|
||||||
),
|
|
||||||
}
|
|
||||||
)
|
|
||||||
table = db.create_table("test_with_id", data)
|
|
||||||
table.create_fts_index("text", with_position=False, use_tantivy=False)
|
|
||||||
return table
|
|
||||||
|
|
||||||
|
|
||||||
def test_hybrid_prefilter_explain_plan(table_with_id: Table):
|
|
||||||
"""
|
|
||||||
Verify that the prefilter logic is not inverted in LanceHybridQueryBuilder.
|
|
||||||
"""
|
|
||||||
plan_prefilter = (
|
|
||||||
table_with_id.search(query_type="hybrid")
|
|
||||||
.vector([0.0, 0.0])
|
|
||||||
.text("dog")
|
|
||||||
.where("id = 1", prefilter=True)
|
|
||||||
.limit(2)
|
|
||||||
.explain_plan(verbose=True)
|
|
||||||
)
|
|
||||||
|
|
||||||
plan_postfilter = (
|
|
||||||
table_with_id.search(query_type="hybrid")
|
|
||||||
.vector([0.0, 0.0])
|
|
||||||
.text("dog")
|
|
||||||
.where("id = 1", prefilter=False)
|
|
||||||
.limit(2)
|
|
||||||
.explain_plan(verbose=True)
|
|
||||||
)
|
|
||||||
|
|
||||||
# prefilter=True: filter is pushed into the LanceRead scan.
|
|
||||||
# The FTS sub-plan exposes this as "full_filter=id = Int64(1)" inside LanceRead.
|
|
||||||
assert "full_filter=id = Int64(1)" in plan_prefilter, (
|
|
||||||
f"Should push the filter into the scan.\nPlan:\n{plan_prefilter}"
|
|
||||||
)
|
|
||||||
|
|
||||||
# prefilter=False: filter is applied as a separate FilterExec after the search.
|
|
||||||
# The filter must NOT be embedded in the scan.
|
|
||||||
assert "full_filter=id = Int64(1)" not in plan_postfilter, (
|
|
||||||
f"Should NOT push the filter into the scan.\nPlan:\n{plan_postfilter}"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def test_normalize_scores():
|
def test_normalize_scores():
|
||||||
cases = [
|
cases = [
|
||||||
(pa.array([0.1, 0.4]), pa.array([0.0, 1.0])),
|
(pa.array([0.1, 0.4]), pa.array([0.0, 1.0])),
|
||||||
|
|||||||
@@ -147,12 +147,7 @@ class TrackingNamespace(LanceNamespace):
|
|||||||
This simulates a credential rotation system where each call returns
|
This simulates a credential rotation system where each call returns
|
||||||
new credentials that expire after credential_expires_in_seconds.
|
new credentials that expire after credential_expires_in_seconds.
|
||||||
"""
|
"""
|
||||||
# Start from base storage options (endpoint, region, allow_http, etc.)
|
modified = copy.deepcopy(storage_options) if storage_options else {}
|
||||||
# because DirectoryNamespace returns None for storage_options from
|
|
||||||
# describe_table/declare_table when no credential vendor is configured.
|
|
||||||
modified = copy.deepcopy(self.base_storage_options)
|
|
||||||
if storage_options:
|
|
||||||
modified.update(storage_options)
|
|
||||||
|
|
||||||
# Increment credentials to simulate rotation
|
# Increment credentials to simulate rotation
|
||||||
modified["aws_access_key_id"] = f"AKID_{count}"
|
modified["aws_access_key_id"] = f"AKID_{count}"
|
||||||
|
|||||||
@@ -326,24 +326,6 @@ def test_add_struct(mem_db: DBConnection):
|
|||||||
table = mem_db.create_table("test2", schema=schema)
|
table = mem_db.create_table("test2", schema=schema)
|
||||||
table.add(data)
|
table.add(data)
|
||||||
|
|
||||||
struct_type = pa.struct(
|
|
||||||
[
|
|
||||||
("b", pa.int64()),
|
|
||||||
("a", pa.int64()),
|
|
||||||
]
|
|
||||||
)
|
|
||||||
expected = pa.table(
|
|
||||||
{
|
|
||||||
"s_list": [
|
|
||||||
[
|
|
||||||
pa.scalar({"b": 1, "a": 2}, type=struct_type),
|
|
||||||
pa.scalar({"b": 4, "a": None}, type=struct_type),
|
|
||||||
]
|
|
||||||
],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
assert table.to_arrow() == expected
|
|
||||||
|
|
||||||
|
|
||||||
def test_add_subschema(mem_db: DBConnection):
|
def test_add_subschema(mem_db: DBConnection):
|
||||||
schema = pa.schema(
|
schema = pa.schema(
|
||||||
|
|||||||
@@ -17,8 +17,7 @@ use pyo3::{
|
|||||||
use pyo3_async_runtimes::tokio::future_into_py;
|
use pyo3_async_runtimes::tokio::future_into_py;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
error::PythonErrorExt, namespace::extract_namespace_arc,
|
error::PythonErrorExt, storage_options::py_object_to_storage_options_provider, table::Table,
|
||||||
storage_options::py_object_to_storage_options_provider, table::Table,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
#[pyclass]
|
#[pyclass]
|
||||||
@@ -183,8 +182,7 @@ impl Connection {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
#[allow(clippy::too_many_arguments)]
|
#[pyo3(signature = (name, namespace=vec![], storage_options = None, storage_options_provider=None, index_cache_size = None, location=None))]
|
||||||
#[pyo3(signature = (name, namespace=vec![], storage_options = None, storage_options_provider=None, index_cache_size = None, location=None, namespace_client=None, managed_versioning=None))]
|
|
||||||
pub fn open_table(
|
pub fn open_table(
|
||||||
self_: PyRef<'_, Self>,
|
self_: PyRef<'_, Self>,
|
||||||
name: String,
|
name: String,
|
||||||
@@ -193,13 +191,11 @@ impl Connection {
|
|||||||
storage_options_provider: Option<Py<PyAny>>,
|
storage_options_provider: Option<Py<PyAny>>,
|
||||||
index_cache_size: Option<u32>,
|
index_cache_size: Option<u32>,
|
||||||
location: Option<String>,
|
location: Option<String>,
|
||||||
namespace_client: Option<Py<PyAny>>,
|
|
||||||
managed_versioning: Option<bool>,
|
|
||||||
) -> PyResult<Bound<'_, PyAny>> {
|
) -> PyResult<Bound<'_, PyAny>> {
|
||||||
let inner = self_.get_inner()?.clone();
|
let inner = self_.get_inner()?.clone();
|
||||||
|
|
||||||
let mut builder = inner.open_table(name);
|
let mut builder = inner.open_table(name);
|
||||||
builder = builder.namespace(namespace.clone());
|
builder = builder.namespace(namespace);
|
||||||
if let Some(storage_options) = storage_options {
|
if let Some(storage_options) = storage_options {
|
||||||
builder = builder.storage_options(storage_options);
|
builder = builder.storage_options(storage_options);
|
||||||
}
|
}
|
||||||
@@ -213,20 +209,6 @@ impl Connection {
|
|||||||
if let Some(location) = location {
|
if let Some(location) = location {
|
||||||
builder = builder.location(location);
|
builder = builder.location(location);
|
||||||
}
|
}
|
||||||
// Extract namespace client from Python object if provided
|
|
||||||
let ns_client = if let Some(ns_obj) = namespace_client {
|
|
||||||
let py = self_.py();
|
|
||||||
Some(extract_namespace_arc(py, ns_obj)?)
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
if let Some(ns_client) = ns_client {
|
|
||||||
builder = builder.namespace_client(ns_client);
|
|
||||||
}
|
|
||||||
// Pass managed_versioning if provided to avoid redundant describe_table call
|
|
||||||
if let Some(enabled) = managed_versioning {
|
|
||||||
builder = builder.managed_versioning(enabled);
|
|
||||||
}
|
|
||||||
|
|
||||||
future_into_py(self_.py(), async move {
|
future_into_py(self_.py(), async move {
|
||||||
let table = builder.execute().await.infer_error()?;
|
let table = builder.execute().await.infer_error()?;
|
||||||
@@ -524,6 +506,7 @@ pub struct PyClientConfig {
|
|||||||
id_delimiter: Option<String>,
|
id_delimiter: Option<String>,
|
||||||
tls_config: Option<PyClientTlsConfig>,
|
tls_config: Option<PyClientTlsConfig>,
|
||||||
header_provider: Option<Py<PyAny>>,
|
header_provider: Option<Py<PyAny>>,
|
||||||
|
mem_wal_enabled: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(FromPyObject)]
|
#[derive(FromPyObject)]
|
||||||
@@ -608,6 +591,7 @@ impl From<PyClientConfig> for lancedb::remote::ClientConfig {
|
|||||||
id_delimiter: value.id_delimiter,
|
id_delimiter: value.id_delimiter,
|
||||||
tls_config: value.tls_config.map(Into::into),
|
tls_config: value.tls_config.map(Into::into),
|
||||||
header_provider,
|
header_provider,
|
||||||
|
mem_wal_enabled: value.mem_wal_enabled,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -23,7 +23,6 @@ pub mod connection;
|
|||||||
pub mod error;
|
pub mod error;
|
||||||
pub mod header;
|
pub mod header;
|
||||||
pub mod index;
|
pub mod index;
|
||||||
pub mod namespace;
|
|
||||||
pub mod permutation;
|
pub mod permutation;
|
||||||
pub mod query;
|
pub mod query;
|
||||||
pub mod session;
|
pub mod session;
|
||||||
|
|||||||
@@ -1,696 +0,0 @@
|
|||||||
// SPDX-License-Identifier: Apache-2.0
|
|
||||||
// SPDX-FileCopyrightText: Copyright The LanceDB Authors
|
|
||||||
|
|
||||||
//! Namespace utilities for Python bindings
|
|
||||||
|
|
||||||
use std::collections::HashMap;
|
|
||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use async_trait::async_trait;
|
|
||||||
use bytes::Bytes;
|
|
||||||
use lance_namespace::LanceNamespace as LanceNamespaceTrait;
|
|
||||||
use lance_namespace::models::*;
|
|
||||||
use pyo3::prelude::*;
|
|
||||||
use pyo3::types::PyDict;
|
|
||||||
|
|
||||||
/// Wrapper that allows any Python object implementing LanceNamespace protocol
|
|
||||||
/// to be used as a Rust LanceNamespace.
|
|
||||||
///
|
|
||||||
/// This is similar to PyLanceNamespace in lance's Python bindings - it wraps a Python
|
|
||||||
/// object and calls back into Python when namespace methods are invoked.
|
|
||||||
pub struct PyLanceNamespace {
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
namespace_id: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl PyLanceNamespace {
|
|
||||||
/// Create a new PyLanceNamespace wrapper around a Python namespace object.
|
|
||||||
pub fn new(_py: Python<'_>, py_namespace: &Bound<'_, PyAny>) -> PyResult<Self> {
|
|
||||||
let namespace_id = py_namespace
|
|
||||||
.call_method0("namespace_id")?
|
|
||||||
.extract::<String>()?;
|
|
||||||
|
|
||||||
Ok(Self {
|
|
||||||
py_namespace: Arc::new(py_namespace.clone().unbind()),
|
|
||||||
namespace_id,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Create an Arc<dyn LanceNamespace> from a Python namespace object.
|
|
||||||
pub fn create_arc(
|
|
||||||
py: Python<'_>,
|
|
||||||
py_namespace: &Bound<'_, PyAny>,
|
|
||||||
) -> PyResult<Arc<dyn LanceNamespaceTrait>> {
|
|
||||||
let wrapper = Self::new(py, py_namespace)?;
|
|
||||||
Ok(Arc::new(wrapper))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl std::fmt::Debug for PyLanceNamespace {
|
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
|
||||||
write!(f, "PyLanceNamespace {{ id: {} }}", self.namespace_id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get or create the DictWithModelDump class in Python.
|
|
||||||
/// This class acts like a dict but also has model_dump() method.
|
|
||||||
/// This allows it to work with both:
|
|
||||||
/// - depythonize (which expects a dict/Mapping)
|
|
||||||
/// - Python code that calls .model_dump() (like DirectoryNamespace wrapper)
|
|
||||||
fn get_dict_with_model_dump_class(py: Python<'_>) -> PyResult<Bound<'_, PyAny>> {
|
|
||||||
// Use a module-level cache via __builtins__
|
|
||||||
let builtins = py.import("builtins")?;
|
|
||||||
if builtins.hasattr("_DictWithModelDump")? {
|
|
||||||
return builtins.getattr("_DictWithModelDump");
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create the class using exec
|
|
||||||
let locals = PyDict::new(py);
|
|
||||||
py.run(
|
|
||||||
c"class DictWithModelDump(dict):
|
|
||||||
def model_dump(self):
|
|
||||||
return dict(self)",
|
|
||||||
None,
|
|
||||||
Some(&locals),
|
|
||||||
)?;
|
|
||||||
let class = locals.get_item("DictWithModelDump")?.ok_or_else(|| {
|
|
||||||
pyo3::exceptions::PyRuntimeError::new_err("Failed to create DictWithModelDump class")
|
|
||||||
})?;
|
|
||||||
|
|
||||||
// Cache it
|
|
||||||
builtins.setattr("_DictWithModelDump", &class)?;
|
|
||||||
Ok(class)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper to call a Python namespace method with JSON serialization.
|
|
||||||
/// For methods that take a request and return a response.
|
|
||||||
/// Uses DictWithModelDump to pass a dict that also has model_dump() method,
|
|
||||||
/// making it compatible with both depythonize and Python wrappers.
|
|
||||||
async fn call_py_method<Req, Resp>(
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
method_name: &'static str,
|
|
||||||
request: Req,
|
|
||||||
) -> lance_core::Result<Resp>
|
|
||||||
where
|
|
||||||
Req: serde::Serialize + Send + 'static,
|
|
||||||
Resp: serde::de::DeserializeOwned + Send + 'static,
|
|
||||||
{
|
|
||||||
let request_json = serde_json::to_string(&request).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to serialize request for {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})?;
|
|
||||||
|
|
||||||
let response_json = tokio::task::spawn_blocking(move || {
|
|
||||||
Python::attach(|py| {
|
|
||||||
let json_module = py.import("json")?;
|
|
||||||
let request_dict = json_module.call_method1("loads", (&request_json,))?;
|
|
||||||
|
|
||||||
// Wrap dict in DictWithModelDump so it works with both depythonize and .model_dump()
|
|
||||||
let dict_class = get_dict_with_model_dump_class(py)?;
|
|
||||||
let request_arg = dict_class.call1((request_dict,))?;
|
|
||||||
|
|
||||||
// Call the Python method
|
|
||||||
let result = py_namespace.call_method1(py, method_name, (request_arg,))?;
|
|
||||||
|
|
||||||
// Convert response to dict, then to JSON
|
|
||||||
// Pydantic models have model_dump() method
|
|
||||||
let result_dict = if result.bind(py).hasattr("model_dump")? {
|
|
||||||
result.call_method0(py, "model_dump")?
|
|
||||||
} else {
|
|
||||||
result
|
|
||||||
};
|
|
||||||
let response_json: String = json_module
|
|
||||||
.call_method1("dumps", (result_dict,))?
|
|
||||||
.extract()?;
|
|
||||||
Ok::<_, PyErr>(response_json)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.map_err(|e| lance_core::Error::io(format!("Task join error for {}: {}", method_name, e)))?
|
|
||||||
.map_err(|e: PyErr| lance_core::Error::io(format!("Python error in {}: {}", method_name, e)))?;
|
|
||||||
|
|
||||||
serde_json::from_str(&response_json).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to deserialize response from {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper for methods that return () on success
|
|
||||||
async fn call_py_method_unit<Req>(
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
method_name: &'static str,
|
|
||||||
request: Req,
|
|
||||||
) -> lance_core::Result<()>
|
|
||||||
where
|
|
||||||
Req: serde::Serialize + Send + 'static,
|
|
||||||
{
|
|
||||||
let request_json = serde_json::to_string(&request).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to serialize request for {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})?;
|
|
||||||
|
|
||||||
tokio::task::spawn_blocking(move || {
|
|
||||||
Python::attach(|py| {
|
|
||||||
let json_module = py.import("json")?;
|
|
||||||
let request_dict = json_module.call_method1("loads", (&request_json,))?;
|
|
||||||
|
|
||||||
// Wrap dict in DictWithModelDump
|
|
||||||
let dict_class = get_dict_with_model_dump_class(py)?;
|
|
||||||
let request_arg = dict_class.call1((request_dict,))?;
|
|
||||||
|
|
||||||
// Call the Python method
|
|
||||||
py_namespace.call_method1(py, method_name, (request_arg,))?;
|
|
||||||
Ok::<_, PyErr>(())
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.map_err(|e| lance_core::Error::io(format!("Task join error for {}: {}", method_name, e)))?
|
|
||||||
.map_err(|e: PyErr| lance_core::Error::io(format!("Python error in {}: {}", method_name, e)))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper for methods that return a primitive type
|
|
||||||
async fn call_py_method_primitive<Req, Resp>(
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
method_name: &'static str,
|
|
||||||
request: Req,
|
|
||||||
) -> lance_core::Result<Resp>
|
|
||||||
where
|
|
||||||
Req: serde::Serialize + Send + 'static,
|
|
||||||
Resp: for<'py> pyo3::FromPyObject<'py> + Send + 'static,
|
|
||||||
{
|
|
||||||
let request_json = serde_json::to_string(&request).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to serialize request for {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})?;
|
|
||||||
|
|
||||||
tokio::task::spawn_blocking(move || {
|
|
||||||
Python::attach(|py| {
|
|
||||||
let json_module = py.import("json")?;
|
|
||||||
let request_dict = json_module.call_method1("loads", (&request_json,))?;
|
|
||||||
|
|
||||||
// Wrap dict in DictWithModelDump
|
|
||||||
let dict_class = get_dict_with_model_dump_class(py)?;
|
|
||||||
let request_arg = dict_class.call1((request_dict,))?;
|
|
||||||
|
|
||||||
// Call the Python method
|
|
||||||
let result = py_namespace.call_method1(py, method_name, (request_arg,))?;
|
|
||||||
let value: Resp = result.extract(py)?;
|
|
||||||
Ok::<_, PyErr>(value)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.map_err(|e| lance_core::Error::io(format!("Task join error for {}: {}", method_name, e)))?
|
|
||||||
.map_err(|e: PyErr| lance_core::Error::io(format!("Python error in {}: {}", method_name, e)))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper for methods that return Bytes
|
|
||||||
async fn call_py_method_bytes<Req>(
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
method_name: &'static str,
|
|
||||||
request: Req,
|
|
||||||
) -> lance_core::Result<Bytes>
|
|
||||||
where
|
|
||||||
Req: serde::Serialize + Send + 'static,
|
|
||||||
{
|
|
||||||
let request_json = serde_json::to_string(&request).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to serialize request for {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})?;
|
|
||||||
|
|
||||||
tokio::task::spawn_blocking(move || {
|
|
||||||
Python::attach(|py| {
|
|
||||||
let json_module = py.import("json")?;
|
|
||||||
let request_dict = json_module.call_method1("loads", (&request_json,))?;
|
|
||||||
|
|
||||||
// Wrap dict in DictWithModelDump
|
|
||||||
let dict_class = get_dict_with_model_dump_class(py)?;
|
|
||||||
let request_arg = dict_class.call1((request_dict,))?;
|
|
||||||
|
|
||||||
// Call the Python method
|
|
||||||
let result = py_namespace.call_method1(py, method_name, (request_arg,))?;
|
|
||||||
let bytes_data: Vec<u8> = result.extract(py)?;
|
|
||||||
Ok::<_, PyErr>(Bytes::from(bytes_data))
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.map_err(|e| lance_core::Error::io(format!("Task join error for {}: {}", method_name, e)))?
|
|
||||||
.map_err(|e: PyErr| lance_core::Error::io(format!("Python error in {}: {}", method_name, e)))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper for methods that take request + data and return a response
|
|
||||||
async fn call_py_method_with_data<Req, Resp>(
|
|
||||||
py_namespace: Arc<Py<PyAny>>,
|
|
||||||
method_name: &'static str,
|
|
||||||
request: Req,
|
|
||||||
data: Bytes,
|
|
||||||
) -> lance_core::Result<Resp>
|
|
||||||
where
|
|
||||||
Req: serde::Serialize + Send + 'static,
|
|
||||||
Resp: serde::de::DeserializeOwned + Send + 'static,
|
|
||||||
{
|
|
||||||
let request_json = serde_json::to_string(&request).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to serialize request for {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})?;
|
|
||||||
|
|
||||||
let response_json = tokio::task::spawn_blocking(move || {
|
|
||||||
Python::attach(|py| {
|
|
||||||
let json_module = py.import("json")?;
|
|
||||||
let request_dict = json_module.call_method1("loads", (&request_json,))?;
|
|
||||||
|
|
||||||
// Wrap dict in DictWithModelDump
|
|
||||||
let dict_class = get_dict_with_model_dump_class(py)?;
|
|
||||||
let request_arg = dict_class.call1((request_dict,))?;
|
|
||||||
|
|
||||||
// Pass request and bytes to Python method
|
|
||||||
let py_bytes = pyo3::types::PyBytes::new(py, &data);
|
|
||||||
let result = py_namespace.call_method1(py, method_name, (request_arg, py_bytes))?;
|
|
||||||
|
|
||||||
// Convert response dict to JSON
|
|
||||||
let response_json: String = json_module.call_method1("dumps", (result,))?.extract()?;
|
|
||||||
Ok::<_, PyErr>(response_json)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.map_err(|e| lance_core::Error::io(format!("Task join error for {}: {}", method_name, e)))?
|
|
||||||
.map_err(|e: PyErr| lance_core::Error::io(format!("Python error in {}: {}", method_name, e)))?;
|
|
||||||
|
|
||||||
serde_json::from_str(&response_json).map_err(|e| {
|
|
||||||
lance_core::Error::io(format!(
|
|
||||||
"Failed to deserialize response from {}: {}",
|
|
||||||
method_name, e
|
|
||||||
))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
#[async_trait]
|
|
||||||
impl LanceNamespaceTrait for PyLanceNamespace {
|
|
||||||
fn namespace_id(&self) -> String {
|
|
||||||
self.namespace_id.clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_namespaces(
|
|
||||||
&self,
|
|
||||||
request: ListNamespacesRequest,
|
|
||||||
) -> lance_core::Result<ListNamespacesResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_namespaces", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn describe_namespace(
|
|
||||||
&self,
|
|
||||||
request: DescribeNamespaceRequest,
|
|
||||||
) -> lance_core::Result<DescribeNamespaceResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "describe_namespace", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_namespace(
|
|
||||||
&self,
|
|
||||||
request: CreateNamespaceRequest,
|
|
||||||
) -> lance_core::Result<CreateNamespaceResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "create_namespace", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn drop_namespace(
|
|
||||||
&self,
|
|
||||||
request: DropNamespaceRequest,
|
|
||||||
) -> lance_core::Result<DropNamespaceResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "drop_namespace", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn namespace_exists(&self, request: NamespaceExistsRequest) -> lance_core::Result<()> {
|
|
||||||
call_py_method_unit(self.py_namespace.clone(), "namespace_exists", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_tables(
|
|
||||||
&self,
|
|
||||||
request: ListTablesRequest,
|
|
||||||
) -> lance_core::Result<ListTablesResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_tables", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn describe_table(
|
|
||||||
&self,
|
|
||||||
request: DescribeTableRequest,
|
|
||||||
) -> lance_core::Result<DescribeTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "describe_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn register_table(
|
|
||||||
&self,
|
|
||||||
request: RegisterTableRequest,
|
|
||||||
) -> lance_core::Result<RegisterTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "register_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn table_exists(&self, request: TableExistsRequest) -> lance_core::Result<()> {
|
|
||||||
call_py_method_unit(self.py_namespace.clone(), "table_exists", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn drop_table(&self, request: DropTableRequest) -> lance_core::Result<DropTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "drop_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn deregister_table(
|
|
||||||
&self,
|
|
||||||
request: DeregisterTableRequest,
|
|
||||||
) -> lance_core::Result<DeregisterTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "deregister_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn count_table_rows(&self, request: CountTableRowsRequest) -> lance_core::Result<i64> {
|
|
||||||
call_py_method_primitive(self.py_namespace.clone(), "count_table_rows", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_table(
|
|
||||||
&self,
|
|
||||||
request: CreateTableRequest,
|
|
||||||
request_data: Bytes,
|
|
||||||
) -> lance_core::Result<CreateTableResponse> {
|
|
||||||
call_py_method_with_data(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"create_table",
|
|
||||||
request,
|
|
||||||
request_data,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn declare_table(
|
|
||||||
&self,
|
|
||||||
request: DeclareTableRequest,
|
|
||||||
) -> lance_core::Result<DeclareTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "declare_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn insert_into_table(
|
|
||||||
&self,
|
|
||||||
request: InsertIntoTableRequest,
|
|
||||||
request_data: Bytes,
|
|
||||||
) -> lance_core::Result<InsertIntoTableResponse> {
|
|
||||||
call_py_method_with_data(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"insert_into_table",
|
|
||||||
request,
|
|
||||||
request_data,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn merge_insert_into_table(
|
|
||||||
&self,
|
|
||||||
request: MergeInsertIntoTableRequest,
|
|
||||||
request_data: Bytes,
|
|
||||||
) -> lance_core::Result<MergeInsertIntoTableResponse> {
|
|
||||||
call_py_method_with_data(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"merge_insert_into_table",
|
|
||||||
request,
|
|
||||||
request_data,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn update_table(
|
|
||||||
&self,
|
|
||||||
request: UpdateTableRequest,
|
|
||||||
) -> lance_core::Result<UpdateTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "update_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn delete_from_table(
|
|
||||||
&self,
|
|
||||||
request: DeleteFromTableRequest,
|
|
||||||
) -> lance_core::Result<DeleteFromTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "delete_from_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn query_table(&self, request: QueryTableRequest) -> lance_core::Result<Bytes> {
|
|
||||||
call_py_method_bytes(self.py_namespace.clone(), "query_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_table_index(
|
|
||||||
&self,
|
|
||||||
request: CreateTableIndexRequest,
|
|
||||||
) -> lance_core::Result<CreateTableIndexResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "create_table_index", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_table_indices(
|
|
||||||
&self,
|
|
||||||
request: ListTableIndicesRequest,
|
|
||||||
) -> lance_core::Result<ListTableIndicesResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_table_indices", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn describe_table_index_stats(
|
|
||||||
&self,
|
|
||||||
request: DescribeTableIndexStatsRequest,
|
|
||||||
) -> lance_core::Result<DescribeTableIndexStatsResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"describe_table_index_stats",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn describe_transaction(
|
|
||||||
&self,
|
|
||||||
request: DescribeTransactionRequest,
|
|
||||||
) -> lance_core::Result<DescribeTransactionResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "describe_transaction", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn alter_transaction(
|
|
||||||
&self,
|
|
||||||
request: AlterTransactionRequest,
|
|
||||||
) -> lance_core::Result<AlterTransactionResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "alter_transaction", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_table_scalar_index(
|
|
||||||
&self,
|
|
||||||
request: CreateTableIndexRequest,
|
|
||||||
) -> lance_core::Result<CreateTableScalarIndexResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"create_table_scalar_index",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn drop_table_index(
|
|
||||||
&self,
|
|
||||||
request: DropTableIndexRequest,
|
|
||||||
) -> lance_core::Result<DropTableIndexResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "drop_table_index", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_all_tables(
|
|
||||||
&self,
|
|
||||||
request: ListTablesRequest,
|
|
||||||
) -> lance_core::Result<ListTablesResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_all_tables", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn restore_table(
|
|
||||||
&self,
|
|
||||||
request: RestoreTableRequest,
|
|
||||||
) -> lance_core::Result<RestoreTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "restore_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn rename_table(
|
|
||||||
&self,
|
|
||||||
request: RenameTableRequest,
|
|
||||||
) -> lance_core::Result<RenameTableResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "rename_table", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_table_versions(
|
|
||||||
&self,
|
|
||||||
request: ListTableVersionsRequest,
|
|
||||||
) -> lance_core::Result<ListTableVersionsResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_table_versions", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_table_version(
|
|
||||||
&self,
|
|
||||||
request: CreateTableVersionRequest,
|
|
||||||
) -> lance_core::Result<CreateTableVersionResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "create_table_version", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn describe_table_version(
|
|
||||||
&self,
|
|
||||||
request: DescribeTableVersionRequest,
|
|
||||||
) -> lance_core::Result<DescribeTableVersionResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "describe_table_version", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn batch_delete_table_versions(
|
|
||||||
&self,
|
|
||||||
request: BatchDeleteTableVersionsRequest,
|
|
||||||
) -> lance_core::Result<BatchDeleteTableVersionsResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"batch_delete_table_versions",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn update_table_schema_metadata(
|
|
||||||
&self,
|
|
||||||
request: UpdateTableSchemaMetadataRequest,
|
|
||||||
) -> lance_core::Result<UpdateTableSchemaMetadataResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"update_table_schema_metadata",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn get_table_stats(
|
|
||||||
&self,
|
|
||||||
request: GetTableStatsRequest,
|
|
||||||
) -> lance_core::Result<GetTableStatsResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "get_table_stats", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn explain_table_query_plan(
|
|
||||||
&self,
|
|
||||||
request: ExplainTableQueryPlanRequest,
|
|
||||||
) -> lance_core::Result<String> {
|
|
||||||
call_py_method_primitive(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"explain_table_query_plan",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn analyze_table_query_plan(
|
|
||||||
&self,
|
|
||||||
request: AnalyzeTableQueryPlanRequest,
|
|
||||||
) -> lance_core::Result<String> {
|
|
||||||
call_py_method_primitive(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"analyze_table_query_plan",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn alter_table_add_columns(
|
|
||||||
&self,
|
|
||||||
request: AlterTableAddColumnsRequest,
|
|
||||||
) -> lance_core::Result<AlterTableAddColumnsResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"alter_table_add_columns",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn alter_table_alter_columns(
|
|
||||||
&self,
|
|
||||||
request: AlterTableAlterColumnsRequest,
|
|
||||||
) -> lance_core::Result<AlterTableAlterColumnsResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"alter_table_alter_columns",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn alter_table_drop_columns(
|
|
||||||
&self,
|
|
||||||
request: AlterTableDropColumnsRequest,
|
|
||||||
) -> lance_core::Result<AlterTableDropColumnsResponse> {
|
|
||||||
call_py_method(
|
|
||||||
self.py_namespace.clone(),
|
|
||||||
"alter_table_drop_columns",
|
|
||||||
request,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn list_table_tags(
|
|
||||||
&self,
|
|
||||||
request: ListTableTagsRequest,
|
|
||||||
) -> lance_core::Result<ListTableTagsResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "list_table_tags", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn create_table_tag(
|
|
||||||
&self,
|
|
||||||
request: CreateTableTagRequest,
|
|
||||||
) -> lance_core::Result<CreateTableTagResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "create_table_tag", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn delete_table_tag(
|
|
||||||
&self,
|
|
||||||
request: DeleteTableTagRequest,
|
|
||||||
) -> lance_core::Result<DeleteTableTagResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "delete_table_tag", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn update_table_tag(
|
|
||||||
&self,
|
|
||||||
request: UpdateTableTagRequest,
|
|
||||||
) -> lance_core::Result<UpdateTableTagResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "update_table_tag", request).await
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn get_table_tag_version(
|
|
||||||
&self,
|
|
||||||
request: GetTableTagVersionRequest,
|
|
||||||
) -> lance_core::Result<GetTableTagVersionResponse> {
|
|
||||||
call_py_method(self.py_namespace.clone(), "get_table_tag_version", request).await
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Convert Python dict to HashMap<String, String>
|
|
||||||
#[allow(dead_code)]
|
|
||||||
fn dict_to_hashmap(dict: &Bound<'_, PyDict>) -> PyResult<HashMap<String, String>> {
|
|
||||||
let mut map = HashMap::new();
|
|
||||||
for (key, value) in dict.iter() {
|
|
||||||
let key_str: String = key.extract()?;
|
|
||||||
let value_str: String = value.extract()?;
|
|
||||||
map.insert(key_str, value_str);
|
|
||||||
}
|
|
||||||
Ok(map)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Extract an Arc<dyn LanceNamespace> from a Python namespace object.
|
|
||||||
///
|
|
||||||
/// This function wraps any Python namespace object with PyLanceNamespace.
|
|
||||||
/// The PyLanceNamespace wrapper uses DictWithModelDump to pass requests,
|
|
||||||
/// which works with both:
|
|
||||||
/// - Native namespaces (DirectoryNamespace, RestNamespace) that use depythonize (expects dict)
|
|
||||||
/// - Custom Python implementations that call .model_dump() on the request
|
|
||||||
pub fn extract_namespace_arc(
|
|
||||||
py: Python<'_>,
|
|
||||||
ns: Py<PyAny>,
|
|
||||||
) -> PyResult<Arc<dyn LanceNamespaceTrait>> {
|
|
||||||
let ns_ref = ns.bind(py);
|
|
||||||
PyLanceNamespace::create_arc(py, ns_ref)
|
|
||||||
}
|
|
||||||
@@ -316,19 +316,6 @@ impl<'py> IntoPyObject<'py> for PySelect {
|
|||||||
Select::All => Ok(py.None().into_bound(py).into_any()),
|
Select::All => Ok(py.None().into_bound(py).into_any()),
|
||||||
Select::Columns(columns) => Ok(columns.into_pyobject(py)?.into_any()),
|
Select::Columns(columns) => Ok(columns.into_pyobject(py)?.into_any()),
|
||||||
Select::Dynamic(columns) => Ok(columns.into_pyobject(py)?.into_any()),
|
Select::Dynamic(columns) => Ok(columns.into_pyobject(py)?.into_any()),
|
||||||
Select::Expr(pairs) => {
|
|
||||||
// Serialize DataFusion Expr -> SQL string so Python sees the same
|
|
||||||
// format as Select::Dynamic: a list of (name, sql_string) tuples.
|
|
||||||
let sql_pairs: PyResult<Vec<(String, String)>> = pairs
|
|
||||||
.into_iter()
|
|
||||||
.map(|(name, expr)| {
|
|
||||||
lancedb::expr::expr_to_sql_string(&expr)
|
|
||||||
.map(|sql| (name, sql))
|
|
||||||
.map_err(|e| PyRuntimeError::new_err(e.to_string()))
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
Ok(sql_pairs?.into_pyobject(py)?.into_any())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -426,17 +426,6 @@ impl Table {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn prewarm_data(
|
|
||||||
self_: PyRef<'_, Self>,
|
|
||||||
columns: Option<Vec<String>>,
|
|
||||||
) -> PyResult<Bound<'_, PyAny>> {
|
|
||||||
let inner = self_.inner_ref()?.clone();
|
|
||||||
future_into_py(self_.py(), async move {
|
|
||||||
inner.prewarm_data(columns).await.infer_error()?;
|
|
||||||
Ok(())
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn list_indices(self_: PyRef<'_, Self>) -> PyResult<Bound<'_, PyAny>> {
|
pub fn list_indices(self_: PyRef<'_, Self>) -> PyResult<Bound<'_, PyAny>> {
|
||||||
let inner = self_.inner_ref()?.clone();
|
let inner = self_.inner_ref()?.clone();
|
||||||
future_into_py(self_.py(), async move {
|
future_into_py(self_.py(), async move {
|
||||||
@@ -721,6 +710,9 @@ impl Table {
|
|||||||
if let Some(use_index) = parameters.use_index {
|
if let Some(use_index) = parameters.use_index {
|
||||||
builder.use_index(use_index);
|
builder.use_index(use_index);
|
||||||
}
|
}
|
||||||
|
if let Some(mem_wal) = parameters.mem_wal {
|
||||||
|
builder.mem_wal(mem_wal);
|
||||||
|
}
|
||||||
|
|
||||||
future_into_py(self_.py(), async move {
|
future_into_py(self_.py(), async move {
|
||||||
let res = builder.execute(Box::new(batches)).await.infer_error()?;
|
let res = builder.execute(Box::new(batches)).await.infer_error()?;
|
||||||
@@ -881,6 +873,7 @@ pub struct MergeInsertParams {
|
|||||||
when_not_matched_by_source_condition: Option<String>,
|
when_not_matched_by_source_condition: Option<String>,
|
||||||
timeout: Option<std::time::Duration>,
|
timeout: Option<std::time::Duration>,
|
||||||
use_index: Option<bool>,
|
use_index: Option<bool>,
|
||||||
|
mem_wal: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[pyclass]
|
#[pyclass]
|
||||||
|
|||||||
4
python/uv.lock
generated
4
python/uv.lock
generated
@@ -2006,7 +2006,7 @@ requires-dist = [
|
|||||||
{ name = "botocore", marker = "extra == 'embeddings'", specifier = ">=1.31.57" },
|
{ name = "botocore", marker = "extra == 'embeddings'", specifier = ">=1.31.57" },
|
||||||
{ name = "cohere", marker = "extra == 'embeddings'" },
|
{ name = "cohere", marker = "extra == 'embeddings'" },
|
||||||
{ name = "colpali-engine", marker = "extra == 'embeddings'", specifier = ">=0.3.10" },
|
{ name = "colpali-engine", marker = "extra == 'embeddings'", specifier = ">=0.3.10" },
|
||||||
{ name = "datafusion", marker = "extra == 'tests'", specifier = "<52" },
|
{ name = "datafusion", marker = "extra == 'tests'" },
|
||||||
{ name = "deprecation" },
|
{ name = "deprecation" },
|
||||||
{ name = "duckdb", marker = "extra == 'tests'" },
|
{ name = "duckdb", marker = "extra == 'tests'" },
|
||||||
{ name = "google-generativeai", marker = "extra == 'embeddings'" },
|
{ name = "google-generativeai", marker = "extra == 'embeddings'" },
|
||||||
@@ -2035,7 +2035,7 @@ requires-dist = [
|
|||||||
{ name = "pyarrow-stubs", marker = "extra == 'tests'" },
|
{ name = "pyarrow-stubs", marker = "extra == 'tests'" },
|
||||||
{ name = "pydantic", specifier = ">=1.10" },
|
{ name = "pydantic", specifier = ">=1.10" },
|
||||||
{ name = "pylance", marker = "extra == 'pylance'", specifier = ">=1.0.0b14" },
|
{ name = "pylance", marker = "extra == 'pylance'", specifier = ">=1.0.0b14" },
|
||||||
{ name = "pylance", marker = "extra == 'tests'", specifier = ">=1.0.0b14,<3.0.0" },
|
{ name = "pylance", marker = "extra == 'tests'", specifier = ">=1.0.0b14" },
|
||||||
{ name = "pyright", marker = "extra == 'dev'" },
|
{ name = "pyright", marker = "extra == 'dev'" },
|
||||||
{ name = "pytest", marker = "extra == 'tests'" },
|
{ name = "pytest", marker = "extra == 'tests'" },
|
||||||
{ name = "pytest-asyncio", marker = "extra == 'tests'" },
|
{ name = "pytest-asyncio", marker = "extra == 'tests'" },
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "lancedb"
|
name = "lancedb"
|
||||||
version = "0.27.0"
|
version = "0.27.0-beta.3"
|
||||||
edition.workspace = true
|
edition.workspace = true
|
||||||
description = "LanceDB: A serverless, low-latency vector database for AI applications"
|
description = "LanceDB: A serverless, low-latency vector database for AI applications"
|
||||||
license.workspace = true
|
license.workspace = true
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# LanceDB Rust SDK
|
# LanceDB Rust
|
||||||
|
|
||||||
<a href="https://crates.io/crates/vectordb"></a>
|
<a href="https://crates.io/crates/vectordb"></a>
|
||||||
<a href="https://docs.rs/vectordb/latest/vectordb/"></a>
|
<a href="https://docs.rs/vectordb/latest/vectordb/"></a>
|
||||||
|
|||||||
@@ -136,7 +136,6 @@ impl OpenTableBuilder {
|
|||||||
lance_read_params: None,
|
lance_read_params: None,
|
||||||
location: None,
|
location: None,
|
||||||
namespace_client: None,
|
namespace_client: None,
|
||||||
managed_versioning: None,
|
|
||||||
},
|
},
|
||||||
embedding_registry,
|
embedding_registry,
|
||||||
}
|
}
|
||||||
@@ -236,29 +235,6 @@ impl OpenTableBuilder {
|
|||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set a namespace client for managed versioning support.
|
|
||||||
///
|
|
||||||
/// When a namespace client is provided and the table has `managed_versioning` enabled,
|
|
||||||
/// the table will use the namespace's commit handler to notify the namespace of
|
|
||||||
/// version changes. This enables features like event emission for table modifications.
|
|
||||||
pub fn namespace_client(mut self, client: Arc<dyn lance_namespace::LanceNamespace>) -> Self {
|
|
||||||
self.request.namespace_client = Some(client);
|
|
||||||
self
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Set whether managed versioning is enabled for this table.
|
|
||||||
///
|
|
||||||
/// When set to `Some(true)`, the table will use namespace-managed commits.
|
|
||||||
/// When set to `Some(false)`, the table will use local commits even if namespace_client is set.
|
|
||||||
/// When set to `None` (default), the value will be fetched from the namespace if namespace_client is set.
|
|
||||||
///
|
|
||||||
/// This is typically set when the caller has already queried the namespace and knows the
|
|
||||||
/// managed_versioning status, avoiding a redundant describe_table call.
|
|
||||||
pub fn managed_versioning(mut self, enabled: bool) -> Self {
|
|
||||||
self.request.managed_versioning = Some(enabled);
|
|
||||||
self
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Open the table
|
/// Open the table
|
||||||
pub async fn execute(self) -> Result<Table> {
|
pub async fn execute(self) -> Result<Table> {
|
||||||
let table = self.parent.open_table(self.request).await?;
|
let table = self.parent.open_table(self.request).await?;
|
||||||
@@ -318,12 +294,6 @@ impl CloneTableBuilder {
|
|||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Set a namespace client for managed versioning support.
|
|
||||||
pub fn namespace_client(mut self, client: Arc<dyn lance_namespace::LanceNamespace>) -> Self {
|
|
||||||
self.request.namespace_client = Some(client);
|
|
||||||
self
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Execute the clone operation
|
/// Execute the clone operation
|
||||||
pub async fn execute(self) -> Result<Table> {
|
pub async fn execute(self) -> Result<Table> {
|
||||||
let parent = self.parent.clone();
|
let parent = self.parent.clone();
|
||||||
@@ -814,13 +784,19 @@ impl ConnectBuilder {
|
|||||||
message: "An api_key is required when connecting to LanceDb Cloud".to_string(),
|
message: "An api_key is required when connecting to LanceDb Cloud".to_string(),
|
||||||
})?;
|
})?;
|
||||||
|
|
||||||
|
// Propagate mem_wal_enabled from options to client_config
|
||||||
|
let mut client_config = self.request.client_config;
|
||||||
|
if options.mem_wal_enabled.is_some() {
|
||||||
|
client_config.mem_wal_enabled = options.mem_wal_enabled;
|
||||||
|
}
|
||||||
|
|
||||||
let storage_options = StorageOptions(options.storage_options.clone());
|
let storage_options = StorageOptions(options.storage_options.clone());
|
||||||
let internal = Arc::new(crate::remote::db::RemoteDatabase::try_new(
|
let internal = Arc::new(crate::remote::db::RemoteDatabase::try_new(
|
||||||
&self.request.uri,
|
&self.request.uri,
|
||||||
&api_key,
|
&api_key,
|
||||||
®ion,
|
®ion,
|
||||||
options.host_override,
|
options.host_override,
|
||||||
self.request.client_config,
|
client_config,
|
||||||
storage_options.into(),
|
storage_options.into(),
|
||||||
)?);
|
)?);
|
||||||
Ok(Connection {
|
Ok(Connection {
|
||||||
|
|||||||
@@ -66,10 +66,6 @@ pub struct OpenTableRequest {
|
|||||||
/// Optional namespace client for server-side query execution.
|
/// Optional namespace client for server-side query execution.
|
||||||
/// When set, queries will be executed on the namespace server instead of locally.
|
/// When set, queries will be executed on the namespace server instead of locally.
|
||||||
pub namespace_client: Option<Arc<dyn LanceNamespace>>,
|
pub namespace_client: Option<Arc<dyn LanceNamespace>>,
|
||||||
/// Whether managed versioning is enabled for this table.
|
|
||||||
/// When Some(true), the table will use namespace-managed commits instead of local commits.
|
|
||||||
/// When None and namespace_client is provided, the value will be fetched from the namespace.
|
|
||||||
pub managed_versioning: Option<bool>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl std::fmt::Debug for OpenTableRequest {
|
impl std::fmt::Debug for OpenTableRequest {
|
||||||
@@ -81,7 +77,6 @@ impl std::fmt::Debug for OpenTableRequest {
|
|||||||
.field("lance_read_params", &self.lance_read_params)
|
.field("lance_read_params", &self.lance_read_params)
|
||||||
.field("location", &self.location)
|
.field("location", &self.location)
|
||||||
.field("namespace_client", &self.namespace_client)
|
.field("namespace_client", &self.namespace_client)
|
||||||
.field("managed_versioning", &self.managed_versioning)
|
|
||||||
.finish()
|
.finish()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -166,9 +161,6 @@ pub struct CloneTableRequest {
|
|||||||
/// Whether to perform a shallow clone (true) or deep clone (false). Defaults to true.
|
/// Whether to perform a shallow clone (true) or deep clone (false). Defaults to true.
|
||||||
/// Currently only shallow clone is supported.
|
/// Currently only shallow clone is supported.
|
||||||
pub is_shallow: bool,
|
pub is_shallow: bool,
|
||||||
/// Optional namespace client for managed versioning support.
|
|
||||||
/// When set, enables the commit handler to track table versions through the namespace.
|
|
||||||
pub namespace_client: Option<Arc<dyn LanceNamespace>>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl CloneTableRequest {
|
impl CloneTableRequest {
|
||||||
@@ -180,7 +172,6 @@ impl CloneTableRequest {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -669,7 +669,6 @@ impl ListingDatabase {
|
|||||||
lance_read_params: None,
|
lance_read_params: None,
|
||||||
location: None,
|
location: None,
|
||||||
namespace_client: None,
|
namespace_client: None,
|
||||||
managed_versioning: None,
|
|
||||||
};
|
};
|
||||||
let req = (callback)(req);
|
let req = (callback)(req);
|
||||||
let table = self.open_table(req).await?;
|
let table = self.open_table(req).await?;
|
||||||
@@ -870,7 +869,6 @@ impl Database for ListingDatabase {
|
|||||||
Some(write_params),
|
Some(write_params),
|
||||||
self.read_consistency_interval,
|
self.read_consistency_interval,
|
||||||
request.namespace_client,
|
request.namespace_client,
|
||||||
false, // server_side_query_enabled - listing database doesn't support server-side queries
|
|
||||||
)
|
)
|
||||||
.await
|
.await
|
||||||
{
|
{
|
||||||
@@ -948,9 +946,7 @@ impl Database for ListingDatabase {
|
|||||||
self.store_wrapper.clone(),
|
self.store_wrapper.clone(),
|
||||||
None,
|
None,
|
||||||
self.read_consistency_interval,
|
self.read_consistency_interval,
|
||||||
request.namespace_client,
|
None,
|
||||||
false, // server_side_query_enabled - listing database doesn't support server-side queries
|
|
||||||
None, // managed_versioning - will be queried if namespace_client is provided
|
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
|
|
||||||
@@ -1026,8 +1022,6 @@ impl Database for ListingDatabase {
|
|||||||
Some(read_params),
|
Some(read_params),
|
||||||
self.read_consistency_interval,
|
self.read_consistency_interval,
|
||||||
request.namespace_client,
|
request.namespace_client,
|
||||||
false, // server_side_query_enabled - listing database doesn't support server-side queries
|
|
||||||
request.managed_versioning, // Pass through managed_versioning from request
|
|
||||||
)
|
)
|
||||||
.await?,
|
.await?,
|
||||||
);
|
);
|
||||||
@@ -1168,7 +1162,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@@ -1229,7 +1222,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@@ -1289,7 +1281,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1326,7 +1317,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: false, // Request deep clone
|
is_shallow: false, // Request deep clone
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1367,7 +1357,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1408,7 +1397,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1428,7 +1416,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1465,7 +1452,6 @@ mod tests {
|
|||||||
source_version: Some(1),
|
source_version: Some(1),
|
||||||
source_tag: Some("v1.0".to_string()),
|
source_tag: Some("v1.0".to_string()),
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -1539,7 +1525,6 @@ mod tests {
|
|||||||
source_version: Some(initial_version),
|
source_version: Some(initial_version),
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@@ -1618,7 +1603,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: Some("v1.0".to_string()),
|
source_tag: Some("v1.0".to_string()),
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@@ -1670,7 +1654,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@@ -1763,7 +1746,6 @@ mod tests {
|
|||||||
source_version: None,
|
source_version: None,
|
||||||
source_tag: None,
|
source_tag: None,
|
||||||
is_shallow: true,
|
is_shallow: true,
|
||||||
namespace_client: None,
|
|
||||||
})
|
})
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|||||||
@@ -7,7 +7,6 @@ use std::collections::HashMap;
|
|||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use async_trait::async_trait;
|
use async_trait::async_trait;
|
||||||
use lance::io::commit::namespace_manifest::LanceNamespaceExternalManifestStore;
|
|
||||||
use lance_io::object_store::{ObjectStoreParams, StorageOptionsAccessor};
|
use lance_io::object_store::{ObjectStoreParams, StorageOptionsAccessor};
|
||||||
use lance_namespace::{
|
use lance_namespace::{
|
||||||
LanceNamespace,
|
LanceNamespace,
|
||||||
@@ -19,8 +18,6 @@ use lance_namespace::{
|
|||||||
},
|
},
|
||||||
};
|
};
|
||||||
use lance_namespace_impls::ConnectBuilder;
|
use lance_namespace_impls::ConnectBuilder;
|
||||||
use lance_table::io::commit::CommitHandler;
|
|
||||||
use lance_table::io::commit::external_manifest::ExternalManifestCommitHandler;
|
|
||||||
|
|
||||||
use crate::database::ReadConsistency;
|
use crate::database::ReadConsistency;
|
||||||
use crate::error::{Error, Result};
|
use crate::error::{Error, Result};
|
||||||
@@ -208,12 +205,14 @@ impl Database for LanceNamespaceDatabase {
|
|||||||
let mut table_id = request.namespace.clone();
|
let mut table_id = request.namespace.clone();
|
||||||
table_id.push(request.name.clone());
|
table_id.push(request.name.clone());
|
||||||
|
|
||||||
|
// Try declare_table first, falling back to create_empty_table for backwards
|
||||||
|
// compatibility with older namespace clients that don't support declare_table
|
||||||
let declare_request = DeclareTableRequest {
|
let declare_request = DeclareTableRequest {
|
||||||
id: Some(table_id.clone()),
|
id: Some(table_id.clone()),
|
||||||
..Default::default()
|
..Default::default()
|
||||||
};
|
};
|
||||||
|
|
||||||
let (location, initial_storage_options, managed_versioning) = {
|
let (location, initial_storage_options) = {
|
||||||
let response = self.namespace.declare_table(declare_request).await?;
|
let response = self.namespace.declare_table(declare_request).await?;
|
||||||
let loc = response.location.ok_or_else(|| Error::Runtime {
|
let loc = response.location.ok_or_else(|| Error::Runtime {
|
||||||
message: "Table location is missing from declare_table response".to_string(),
|
message: "Table location is missing from declare_table response".to_string(),
|
||||||
@@ -223,33 +222,21 @@ impl Database for LanceNamespaceDatabase {
|
|||||||
.storage_options
|
.storage_options
|
||||||
.or_else(|| Some(self.storage_options.clone()))
|
.or_else(|| Some(self.storage_options.clone()))
|
||||||
.filter(|o| !o.is_empty());
|
.filter(|o| !o.is_empty());
|
||||||
(loc, opts, response.managed_versioning)
|
(loc, opts)
|
||||||
};
|
};
|
||||||
|
|
||||||
// Build write params with storage options and commit handler
|
let write_params = if let Some(storage_opts) = initial_storage_options {
|
||||||
let mut params = request.write_options.lance_write_params.unwrap_or_default();
|
let mut params = request.write_options.lance_write_params.unwrap_or_default();
|
||||||
|
|
||||||
// Set up storage options if provided
|
|
||||||
if let Some(storage_opts) = initial_storage_options {
|
|
||||||
let store_params = params
|
let store_params = params
|
||||||
.store_params
|
.store_params
|
||||||
.get_or_insert_with(ObjectStoreParams::default);
|
.get_or_insert_with(ObjectStoreParams::default);
|
||||||
store_params.storage_options_accessor = Some(Arc::new(
|
store_params.storage_options_accessor = Some(Arc::new(
|
||||||
StorageOptionsAccessor::with_static_options(storage_opts),
|
StorageOptionsAccessor::with_static_options(storage_opts),
|
||||||
));
|
));
|
||||||
}
|
Some(params)
|
||||||
|
} else {
|
||||||
// Set up commit handler when managed_versioning is enabled
|
request.write_options.lance_write_params
|
||||||
if managed_versioning == Some(true) {
|
};
|
||||||
let external_store =
|
|
||||||
LanceNamespaceExternalManifestStore::new(self.namespace.clone(), table_id.clone());
|
|
||||||
let commit_handler: Arc<dyn CommitHandler> = Arc::new(ExternalManifestCommitHandler {
|
|
||||||
external_manifest_store: Arc::new(external_store),
|
|
||||||
});
|
|
||||||
params.commit_handler = Some(commit_handler);
|
|
||||||
}
|
|
||||||
|
|
||||||
let write_params = Some(params);
|
|
||||||
|
|
||||||
let native_table = NativeTable::create_from_namespace(
|
let native_table = NativeTable::create_from_namespace(
|
||||||
self.namespace.clone(),
|
self.namespace.clone(),
|
||||||
|
|||||||
@@ -339,12 +339,6 @@ impl PermutationReader {
|
|||||||
}
|
}
|
||||||
Ok(false)
|
Ok(false)
|
||||||
}
|
}
|
||||||
Select::Expr(columns) => {
|
|
||||||
// For Expr projections, we check if any alias is _rowid.
|
|
||||||
// We can't validate the expression itself (it may differ from _rowid)
|
|
||||||
// but we allow it through; the column will be included.
|
|
||||||
Ok(columns.iter().any(|(alias, _)| alias == ROW_ID))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -47,25 +47,6 @@ pub enum Select {
|
|||||||
///
|
///
|
||||||
/// See [`Query::select`] for more details and examples
|
/// See [`Query::select`] for more details and examples
|
||||||
Dynamic(Vec<(String, String)>),
|
Dynamic(Vec<(String, String)>),
|
||||||
/// Advanced selection using type-safe DataFusion expressions
|
|
||||||
///
|
|
||||||
/// Similar to [`Select::Dynamic`] but uses [`datafusion_expr::Expr`] instead of
|
|
||||||
/// raw SQL strings. Use [`crate::expr`] helpers to build expressions:
|
|
||||||
///
|
|
||||||
/// ```
|
|
||||||
/// use lancedb::expr::{col, lit};
|
|
||||||
/// use lancedb::query::Select;
|
|
||||||
///
|
|
||||||
/// // SELECT id, id * 2 AS id2 FROM ...
|
|
||||||
/// let selection = Select::expr_projection(&[
|
|
||||||
/// ("id", col("id")),
|
|
||||||
/// ("id2", col("id") * lit(2)),
|
|
||||||
/// ]);
|
|
||||||
/// ```
|
|
||||||
///
|
|
||||||
/// Note: For remote/server-side queries the expressions are serialized to SQL strings
|
|
||||||
/// automatically (same as [`Select::Dynamic`]).
|
|
||||||
Expr(Vec<(String, datafusion_expr::Expr)>),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Select {
|
impl Select {
|
||||||
@@ -88,29 +69,6 @@ impl Select {
|
|||||||
.collect(),
|
.collect(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
/// Create a typed-expression projection.
|
|
||||||
///
|
|
||||||
/// This is a convenience method for creating a [`Select::Expr`] variant from
|
|
||||||
/// a slice of `(name, expr)` pairs where each `expr` is a [`datafusion_expr::Expr`].
|
|
||||||
///
|
|
||||||
/// # Example
|
|
||||||
/// ```
|
|
||||||
/// use lancedb::expr::{col, lit};
|
|
||||||
/// use lancedb::query::Select;
|
|
||||||
///
|
|
||||||
/// let selection = Select::expr_projection(&[
|
|
||||||
/// ("id", col("id")),
|
|
||||||
/// ("id2", col("id") * lit(2)),
|
|
||||||
/// ]);
|
|
||||||
/// ```
|
|
||||||
pub fn expr_projection(columns: &[(impl AsRef<str>, datafusion_expr::Expr)]) -> Self {
|
|
||||||
Self::Expr(
|
|
||||||
columns
|
|
||||||
.iter()
|
|
||||||
.map(|(name, expr)| (name.as_ref().to_string(), expr.clone()))
|
|
||||||
.collect(),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// A trait for converting a type to a query vector
|
/// A trait for converting a type to a query vector
|
||||||
@@ -1633,58 +1591,6 @@ mod tests {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_select_with_expr_projection() {
|
|
||||||
// Mirrors test_select_with_transform but uses Select::Expr instead of Select::Dynamic
|
|
||||||
let tmp_dir = tempdir().unwrap();
|
|
||||||
let dataset_path = tmp_dir.path().join("test_expr.lance");
|
|
||||||
let uri = dataset_path.to_str().unwrap();
|
|
||||||
|
|
||||||
let batches = make_non_empty_batches();
|
|
||||||
let conn = connect(uri).execute().await.unwrap();
|
|
||||||
let table = conn
|
|
||||||
.create_table("my_table", batches)
|
|
||||||
.execute()
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
use crate::expr::{col, lit};
|
|
||||||
let query = table.query().limit(10).select(Select::expr_projection(&[
|
|
||||||
("id2", col("id") * lit(2i32)),
|
|
||||||
("id", col("id")),
|
|
||||||
]));
|
|
||||||
|
|
||||||
let schema = query.output_schema().await.unwrap();
|
|
||||||
assert_eq!(
|
|
||||||
schema,
|
|
||||||
Arc::new(ArrowSchema::new(vec![
|
|
||||||
ArrowField::new("id2", DataType::Int32, true),
|
|
||||||
ArrowField::new("id", DataType::Int32, true),
|
|
||||||
]))
|
|
||||||
);
|
|
||||||
|
|
||||||
let result = query.execute().await;
|
|
||||||
let mut batches = result
|
|
||||||
.expect("should have result")
|
|
||||||
.try_collect::<Vec<_>>()
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
assert_eq!(batches.len(), 1);
|
|
||||||
let batch = batches.pop().unwrap();
|
|
||||||
|
|
||||||
// id and id2
|
|
||||||
assert_eq!(batch.num_columns(), 2);
|
|
||||||
|
|
||||||
let id: &Int32Array = batch.column_by_name("id").unwrap().as_primitive();
|
|
||||||
let id2: &Int32Array = batch.column_by_name("id2").unwrap().as_primitive();
|
|
||||||
|
|
||||||
id.iter().zip(id2.iter()).for_each(|(id, id2)| {
|
|
||||||
let id = id.unwrap();
|
|
||||||
let id2 = id2.unwrap();
|
|
||||||
assert_eq!(id * 2, id2);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn test_execute_no_vector() {
|
async fn test_execute_no_vector() {
|
||||||
// TODO: Switch back to memory://foo after https://github.com/lancedb/lancedb/issues/1051
|
// TODO: Switch back to memory://foo after https://github.com/lancedb/lancedb/issues/1051
|
||||||
|
|||||||
@@ -14,6 +14,7 @@ use crate::remote::db::RemoteOptions;
|
|||||||
use crate::remote::retry::{ResolvedRetryConfig, RetryCounter};
|
use crate::remote::retry::{ResolvedRetryConfig, RetryCounter};
|
||||||
|
|
||||||
const REQUEST_ID_HEADER: HeaderName = HeaderName::from_static("x-request-id");
|
const REQUEST_ID_HEADER: HeaderName = HeaderName::from_static("x-request-id");
|
||||||
|
const MEM_WAL_ENABLED_HEADER: HeaderName = HeaderName::from_static("x-lancedb-mem-wal-enabled");
|
||||||
|
|
||||||
/// Configuration for TLS/mTLS settings.
|
/// Configuration for TLS/mTLS settings.
|
||||||
#[derive(Clone, Debug, Default)]
|
#[derive(Clone, Debug, Default)]
|
||||||
@@ -52,6 +53,10 @@ pub struct ClientConfig {
|
|||||||
pub tls_config: Option<TlsConfig>,
|
pub tls_config: Option<TlsConfig>,
|
||||||
/// Provider for custom headers to be added to each request
|
/// Provider for custom headers to be added to each request
|
||||||
pub header_provider: Option<Arc<dyn HeaderProvider>>,
|
pub header_provider: Option<Arc<dyn HeaderProvider>>,
|
||||||
|
/// Enable MemWAL write path for streaming writes.
|
||||||
|
/// When true, write operations will use the MemWAL architecture
|
||||||
|
/// for high-performance streaming writes.
|
||||||
|
pub mem_wal_enabled: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl std::fmt::Debug for ClientConfig {
|
impl std::fmt::Debug for ClientConfig {
|
||||||
@@ -67,6 +72,7 @@ impl std::fmt::Debug for ClientConfig {
|
|||||||
"header_provider",
|
"header_provider",
|
||||||
&self.header_provider.as_ref().map(|_| "Some(...)"),
|
&self.header_provider.as_ref().map(|_| "Some(...)"),
|
||||||
)
|
)
|
||||||
|
.field("mem_wal_enabled", &self.mem_wal_enabled)
|
||||||
.finish()
|
.finish()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -81,6 +87,7 @@ impl Default for ClientConfig {
|
|||||||
id_delimiter: None,
|
id_delimiter: None,
|
||||||
tls_config: None,
|
tls_config: None,
|
||||||
header_provider: None,
|
header_provider: None,
|
||||||
|
mem_wal_enabled: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -426,11 +433,14 @@ impl<S: HttpSend> RestfulLanceDbClient<S> {
|
|||||||
})?,
|
})?,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
if let Some(prefix) = db_prefix {
|
if db_prefix.is_some() {
|
||||||
headers.insert(
|
headers.insert(
|
||||||
HeaderName::from_static("x-lancedb-database-prefix"),
|
HeaderName::from_static("x-lancedb-database-prefix"),
|
||||||
HeaderValue::from_str(prefix).map_err(|_| Error::InvalidInput {
|
HeaderValue::from_str(db_prefix.unwrap()).map_err(|_| Error::InvalidInput {
|
||||||
message: format!("non-ascii database prefix '{}' provided", prefix),
|
message: format!(
|
||||||
|
"non-ascii database prefix '{}' provided",
|
||||||
|
db_prefix.unwrap()
|
||||||
|
),
|
||||||
})?,
|
})?,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
@@ -474,6 +484,11 @@ impl<S: HttpSend> RestfulLanceDbClient<S> {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Add MemWAL header if enabled
|
||||||
|
if let Some(true) = config.mem_wal_enabled {
|
||||||
|
headers.insert(MEM_WAL_ENABLED_HEADER, HeaderValue::from_static("true"));
|
||||||
|
}
|
||||||
|
|
||||||
Ok(headers)
|
Ok(headers)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -78,6 +78,7 @@ pub const OPT_REMOTE_PREFIX: &str = "remote_database_";
|
|||||||
pub const OPT_REMOTE_API_KEY: &str = "remote_database_api_key";
|
pub const OPT_REMOTE_API_KEY: &str = "remote_database_api_key";
|
||||||
pub const OPT_REMOTE_REGION: &str = "remote_database_region";
|
pub const OPT_REMOTE_REGION: &str = "remote_database_region";
|
||||||
pub const OPT_REMOTE_HOST_OVERRIDE: &str = "remote_database_host_override";
|
pub const OPT_REMOTE_HOST_OVERRIDE: &str = "remote_database_host_override";
|
||||||
|
pub const OPT_REMOTE_MEM_WAL_ENABLED: &str = "remote_database_mem_wal_enabled";
|
||||||
// TODO: add support for configuring client config via key/value options
|
// TODO: add support for configuring client config via key/value options
|
||||||
|
|
||||||
#[derive(Clone, Debug, Default)]
|
#[derive(Clone, Debug, Default)]
|
||||||
@@ -98,6 +99,12 @@ pub struct RemoteDatabaseOptions {
|
|||||||
/// These options are only used for LanceDB Enterprise and only a subset of options
|
/// These options are only used for LanceDB Enterprise and only a subset of options
|
||||||
/// are supported.
|
/// are supported.
|
||||||
pub storage_options: HashMap<String, String>,
|
pub storage_options: HashMap<String, String>,
|
||||||
|
/// Enable MemWAL write path for high-performance streaming writes.
|
||||||
|
///
|
||||||
|
/// When enabled, write operations (insert, merge_insert, etc.) will use
|
||||||
|
/// the MemWAL architecture which buffers writes in memory and Write-Ahead Log
|
||||||
|
/// before asynchronously merging to the base table.
|
||||||
|
pub mem_wal_enabled: Option<bool>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl RemoteDatabaseOptions {
|
impl RemoteDatabaseOptions {
|
||||||
@@ -109,6 +116,9 @@ impl RemoteDatabaseOptions {
|
|||||||
let api_key = map.get(OPT_REMOTE_API_KEY).cloned();
|
let api_key = map.get(OPT_REMOTE_API_KEY).cloned();
|
||||||
let region = map.get(OPT_REMOTE_REGION).cloned();
|
let region = map.get(OPT_REMOTE_REGION).cloned();
|
||||||
let host_override = map.get(OPT_REMOTE_HOST_OVERRIDE).cloned();
|
let host_override = map.get(OPT_REMOTE_HOST_OVERRIDE).cloned();
|
||||||
|
let mem_wal_enabled = map
|
||||||
|
.get(OPT_REMOTE_MEM_WAL_ENABLED)
|
||||||
|
.map(|v| v.to_lowercase() == "true");
|
||||||
let storage_options = map
|
let storage_options = map
|
||||||
.iter()
|
.iter()
|
||||||
.filter(|(key, _)| !key.starts_with(OPT_REMOTE_PREFIX))
|
.filter(|(key, _)| !key.starts_with(OPT_REMOTE_PREFIX))
|
||||||
@@ -119,6 +129,7 @@ impl RemoteDatabaseOptions {
|
|||||||
region,
|
region,
|
||||||
host_override,
|
host_override,
|
||||||
storage_options,
|
storage_options,
|
||||||
|
mem_wal_enabled,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -137,6 +148,12 @@ impl DatabaseOptions for RemoteDatabaseOptions {
|
|||||||
if let Some(host_override) = &self.host_override {
|
if let Some(host_override) = &self.host_override {
|
||||||
map.insert(OPT_REMOTE_HOST_OVERRIDE.to_string(), host_override.clone());
|
map.insert(OPT_REMOTE_HOST_OVERRIDE.to_string(), host_override.clone());
|
||||||
}
|
}
|
||||||
|
if let Some(mem_wal_enabled) = &self.mem_wal_enabled {
|
||||||
|
map.insert(
|
||||||
|
OPT_REMOTE_MEM_WAL_ENABLED.to_string(),
|
||||||
|
mem_wal_enabled.to_string(),
|
||||||
|
);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -181,6 +198,20 @@ impl RemoteDatabaseOptionsBuilder {
|
|||||||
self.options.host_override = Some(host_override);
|
self.options.host_override = Some(host_override);
|
||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Enable MemWAL write path for high-performance streaming writes.
|
||||||
|
///
|
||||||
|
/// When enabled, write operations will use the MemWAL architecture
|
||||||
|
/// which buffers writes in memory and Write-Ahead Log before
|
||||||
|
/// asynchronously merging to the base table.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `enabled` - Whether to enable MemWAL writes
|
||||||
|
pub fn mem_wal_enabled(mut self, enabled: bool) -> Self {
|
||||||
|
self.options.mem_wal_enabled = Some(enabled);
|
||||||
|
self
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
@@ -464,7 +495,6 @@ impl<S: HttpSend> Database for RemoteDatabase<S> {
|
|||||||
lance_read_params: None,
|
lance_read_params: None,
|
||||||
location: None,
|
location: None,
|
||||||
namespace_client: None,
|
namespace_client: None,
|
||||||
managed_versioning: None,
|
|
||||||
};
|
};
|
||||||
let req = (callback)(req);
|
let req = (callback)(req);
|
||||||
self.open_table(req).await
|
self.open_table(req).await
|
||||||
|
|||||||
@@ -62,6 +62,7 @@ use std::time::Duration;
|
|||||||
use tokio::sync::RwLock;
|
use tokio::sync::RwLock;
|
||||||
|
|
||||||
const REQUEST_TIMEOUT_HEADER: HeaderName = HeaderName::from_static("x-request-timeout-ms");
|
const REQUEST_TIMEOUT_HEADER: HeaderName = HeaderName::from_static("x-request-timeout-ms");
|
||||||
|
const MEM_WAL_ENABLED_HEADER: HeaderName = HeaderName::from_static("x-lancedb-mem-wal-enabled");
|
||||||
const METRIC_TYPE_KEY: &str = "metric_type";
|
const METRIC_TYPE_KEY: &str = "metric_type";
|
||||||
const INDEX_TYPE_KEY: &str = "index_type";
|
const INDEX_TYPE_KEY: &str = "index_type";
|
||||||
const SCHEMA_CACHE_TTL: Duration = Duration::from_secs(30);
|
const SCHEMA_CACHE_TTL: Duration = Duration::from_secs(30);
|
||||||
@@ -477,16 +478,6 @@ impl<S: HttpSend> RemoteTable<S> {
|
|||||||
}));
|
}));
|
||||||
body["columns"] = alias_map.into();
|
body["columns"] = alias_map.into();
|
||||||
}
|
}
|
||||||
Select::Expr(pairs) => {
|
|
||||||
let alias_map: Result<serde_json::Map<String, serde_json::Value>> = pairs
|
|
||||||
.iter()
|
|
||||||
.map(|(name, expr)| {
|
|
||||||
expr_to_sql_string(expr)
|
|
||||||
.map(|sql| (name.clone(), serde_json::Value::String(sql)))
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
body["columns"] = alias_map?.into();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if params.fast_search {
|
if params.fast_search {
|
||||||
@@ -1369,6 +1360,7 @@ impl<S: HttpSend> BaseTable for RemoteTable<S> {
|
|||||||
self.check_mutable().await?;
|
self.check_mutable().await?;
|
||||||
|
|
||||||
let timeout = params.timeout;
|
let timeout = params.timeout;
|
||||||
|
let mem_wal = params.mem_wal;
|
||||||
|
|
||||||
let query = MergeInsertRequest::try_from(params)?;
|
let query = MergeInsertRequest::try_from(params)?;
|
||||||
let mut request = self
|
let mut request = self
|
||||||
@@ -1384,6 +1376,10 @@ impl<S: HttpSend> BaseTable for RemoteTable<S> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if mem_wal {
|
||||||
|
request = request.header(MEM_WAL_ENABLED_HEADER, "true");
|
||||||
|
}
|
||||||
|
|
||||||
let (request_id, response) = self.send_streaming(request, new_data, true).await?;
|
let (request_id, response) = self.send_streaming(request, new_data, true).await?;
|
||||||
|
|
||||||
let response = self.check_table_response(&request_id, response).await?;
|
let response = self.check_table_response(&request_id, response).await?;
|
||||||
@@ -1655,33 +1651,10 @@ impl<S: HttpSend> BaseTable for RemoteTable<S> {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn prewarm_index(&self, index_name: &str) -> Result<()> {
|
async fn prewarm_index(&self, _index_name: &str) -> Result<()> {
|
||||||
let request = self.client.post(&format!(
|
Err(Error::NotSupported {
|
||||||
"/v1/table/{}/index/{}/prewarm/",
|
message: "prewarm_index is not yet supported on LanceDB cloud.".into(),
|
||||||
self.identifier, index_name
|
})
|
||||||
));
|
|
||||||
let (request_id, response) = self.send(request, true).await?;
|
|
||||||
if response.status() == StatusCode::NOT_FOUND {
|
|
||||||
return Err(Error::IndexNotFound {
|
|
||||||
name: index_name.to_string(),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
self.check_table_response(&request_id, response).await?;
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn prewarm_data(&self, columns: Option<Vec<String>>) -> Result<()> {
|
|
||||||
let mut request = self.client.post(&format!(
|
|
||||||
"/v1/table/{}/page_cache/prewarm/",
|
|
||||||
self.identifier
|
|
||||||
));
|
|
||||||
let body = serde_json::json!({
|
|
||||||
"columns": columns.unwrap_or_default(),
|
|
||||||
});
|
|
||||||
request = request.json(&body);
|
|
||||||
let (request_id, response) = self.send(request, true).await?;
|
|
||||||
self.check_table_response(&request_id, response).await?;
|
|
||||||
Ok(())
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn table_definition(&self) -> Result<TableDefinition> {
|
async fn table_definition(&self) -> Result<TableDefinition> {
|
||||||
@@ -3562,64 +3535,6 @@ mod tests {
|
|||||||
assert_eq!(result.version, if old_server { 0 } else { 43 });
|
assert_eq!(result.version, if old_server { 0 } else { 43 });
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_prewarm_index() {
|
|
||||||
let table = Table::new_with_handler("my_table", |request| {
|
|
||||||
assert_eq!(request.method(), "POST");
|
|
||||||
assert_eq!(
|
|
||||||
request.url().path(),
|
|
||||||
"/v1/table/my_table/index/my_index/prewarm/"
|
|
||||||
);
|
|
||||||
http::Response::builder().status(200).body("{}").unwrap()
|
|
||||||
});
|
|
||||||
table.prewarm_index("my_index").await.unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_prewarm_index_not_found() {
|
|
||||||
let table = Table::new_with_handler("my_table", |request| {
|
|
||||||
assert_eq!(
|
|
||||||
request.url().path(),
|
|
||||||
"/v1/table/my_table/index/my_index/prewarm/"
|
|
||||||
);
|
|
||||||
http::Response::builder().status(404).body("{}").unwrap()
|
|
||||||
});
|
|
||||||
let e = table.prewarm_index("my_index").await.unwrap_err();
|
|
||||||
assert!(matches!(e, Error::IndexNotFound { .. }));
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_prewarm_data() {
|
|
||||||
let table = Table::new_with_handler("my_table", |request| {
|
|
||||||
assert_eq!(request.method(), "POST");
|
|
||||||
assert_eq!(
|
|
||||||
request.url().path(),
|
|
||||||
"/v1/table/my_table/page_cache/prewarm/"
|
|
||||||
);
|
|
||||||
http::Response::builder().status(200).body("{}").unwrap()
|
|
||||||
});
|
|
||||||
table.prewarm_data(None).await.unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_prewarm_data_with_columns() {
|
|
||||||
let table = Table::new_with_handler("my_table", |request| {
|
|
||||||
assert_eq!(request.method(), "POST");
|
|
||||||
assert_eq!(
|
|
||||||
request.url().path(),
|
|
||||||
"/v1/table/my_table/page_cache/prewarm/"
|
|
||||||
);
|
|
||||||
let body = request.body().unwrap().as_bytes().unwrap();
|
|
||||||
let body: serde_json::Value = serde_json::from_slice(body).unwrap();
|
|
||||||
assert_eq!(body["columns"], serde_json::json!(["col_a", "col_b"]));
|
|
||||||
http::Response::builder().status(200).body("{}").unwrap()
|
|
||||||
});
|
|
||||||
table
|
|
||||||
.prewarm_data(Some(vec!["col_a".into(), "col_b".into()]))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn test_drop_index() {
|
async fn test_drop_index() {
|
||||||
let table = Table::new_with_handler("my_table", |request| {
|
let table = Table::new_with_handler("my_table", |request| {
|
||||||
|
|||||||
@@ -34,13 +34,9 @@ use lance_index::vector::sq::builder::SQBuildParams;
|
|||||||
use lance_io::object_store::{LanceNamespaceStorageOptionsProvider, StorageOptionsAccessor};
|
use lance_io::object_store::{LanceNamespaceStorageOptionsProvider, StorageOptionsAccessor};
|
||||||
pub use query::AnyQuery;
|
pub use query::AnyQuery;
|
||||||
|
|
||||||
use lance::io::commit::namespace_manifest::LanceNamespaceExternalManifestStore;
|
|
||||||
use lance_namespace::LanceNamespace;
|
use lance_namespace::LanceNamespace;
|
||||||
use lance_namespace::models::DescribeTableRequest;
|
|
||||||
use lance_table::format::Manifest;
|
use lance_table::format::Manifest;
|
||||||
use lance_table::io::commit::CommitHandler;
|
|
||||||
use lance_table::io::commit::ManifestNamingScheme;
|
use lance_table::io::commit::ManifestNamingScheme;
|
||||||
use lance_table::io::commit::external_manifest::ExternalManifestCommitHandler;
|
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::format;
|
use std::format;
|
||||||
@@ -277,13 +273,8 @@ pub trait BaseTable: std::fmt::Display + std::fmt::Debug + Send + Sync {
|
|||||||
async fn list_indices(&self) -> Result<Vec<IndexConfig>>;
|
async fn list_indices(&self) -> Result<Vec<IndexConfig>>;
|
||||||
/// Drop an index from the table.
|
/// Drop an index from the table.
|
||||||
async fn drop_index(&self, name: &str) -> Result<()>;
|
async fn drop_index(&self, name: &str) -> Result<()>;
|
||||||
/// Prewarm an index in the table.
|
/// Prewarm an index in the table
|
||||||
async fn prewarm_index(&self, name: &str) -> Result<()>;
|
async fn prewarm_index(&self, name: &str) -> Result<()>;
|
||||||
/// Prewarm data for the table.
|
|
||||||
///
|
|
||||||
/// Currently only supported on remote tables.
|
|
||||||
/// If `columns` is `None`, all columns are prewarmed.
|
|
||||||
async fn prewarm_data(&self, columns: Option<Vec<String>>) -> Result<()>;
|
|
||||||
/// Get statistics about the index.
|
/// Get statistics about the index.
|
||||||
async fn index_stats(&self, index_name: &str) -> Result<Option<IndexStatistics>>;
|
async fn index_stats(&self, index_name: &str) -> Result<Option<IndexStatistics>>;
|
||||||
/// Merge insert new records into the table.
|
/// Merge insert new records into the table.
|
||||||
@@ -951,7 +942,17 @@ impl Table {
|
|||||||
/// * Prune: Removes old versions of the dataset
|
/// * Prune: Removes old versions of the dataset
|
||||||
/// * Index: Optimizes the indices, adding new data to existing indices
|
/// * Index: Optimizes the indices, adding new data to existing indices
|
||||||
///
|
///
|
||||||
/// The frequency an application should call optimize is based on the frequency of
|
/// <section class="warning">Experimental API</section>
|
||||||
|
///
|
||||||
|
/// The optimization process is undergoing active development and may change.
|
||||||
|
/// Our goal with these changes is to improve the performance of optimization and
|
||||||
|
/// reduce the complexity.
|
||||||
|
///
|
||||||
|
/// That being said, it is essential today to run optimize if you want the best
|
||||||
|
/// performance. It should be stable and safe to use in production, but it our
|
||||||
|
/// hope that the API may be simplified (or not even need to be called) in the future.
|
||||||
|
///
|
||||||
|
/// The frequency an application shoudl call optimize is based on the frequency of
|
||||||
/// data modifications. If data is frequently added, deleted, or updated then
|
/// data modifications. If data is frequently added, deleted, or updated then
|
||||||
/// optimize should be run frequently. A good rule of thumb is to run optimize if
|
/// optimize should be run frequently. A good rule of thumb is to run optimize if
|
||||||
/// you have added or modified 100,000 or more records or run more than 20 data
|
/// you have added or modified 100,000 or more records or run more than 20 data
|
||||||
@@ -1118,45 +1119,22 @@ impl Table {
|
|||||||
self.inner.drop_index(name).await
|
self.inner.drop_index(name).await
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Prewarm an index in the table.
|
/// Prewarm an index in the table
|
||||||
///
|
///
|
||||||
/// This is a hint to the database that the index will be accessed in the
|
/// This is a hint to fully load the index into memory. It can be used to
|
||||||
/// future and should be loaded into memory if possible. This can reduce
|
/// avoid cold starts
|
||||||
/// cold-start latency for subsequent queries.
|
|
||||||
///
|
|
||||||
/// This call initiates prewarming and returns once the request is accepted.
|
|
||||||
/// It is idempotent and safe to call from multiple clients concurrently.
|
|
||||||
///
|
///
|
||||||
/// It is generally wasteful to call this if the index does not fit into the
|
/// It is generally wasteful to call this if the index does not fit into the
|
||||||
/// available cache. Not all index types support prewarming; unsupported
|
/// available cache.
|
||||||
/// indices will silently ignore the request.
|
///
|
||||||
|
/// Note: This function is not yet supported on all indices, in which case it
|
||||||
|
/// may do nothing.
|
||||||
///
|
///
|
||||||
/// Use [`Self::list_indices()`] to find the names of the indices.
|
/// Use [`Self::list_indices()`] to find the names of the indices.
|
||||||
pub async fn prewarm_index(&self, name: &str) -> Result<()> {
|
pub async fn prewarm_index(&self, name: &str) -> Result<()> {
|
||||||
self.inner.prewarm_index(name).await
|
self.inner.prewarm_index(name).await
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Prewarm data for the table.
|
|
||||||
///
|
|
||||||
/// This is a hint to the database that the given columns will be accessed in
|
|
||||||
/// the future and the database should prefetch the data if possible. This
|
|
||||||
/// can reduce cold-start latency for subsequent queries. Currently only
|
|
||||||
/// supported on remote tables.
|
|
||||||
///
|
|
||||||
/// This call initiates prewarming and returns once the request is accepted.
|
|
||||||
/// It is idempotent and safe to call from multiple clients concurrently —
|
|
||||||
/// calling it on already-prewarmed columns is a no-op on the server.
|
|
||||||
///
|
|
||||||
/// This operation has a large upfront cost but can speed up future queries
|
|
||||||
/// that need to fetch the given columns. Large columns such as embeddings
|
|
||||||
/// or binary data may not be practical to prewarm. This feature is intended
|
|
||||||
/// for workloads that issue many queries against the same columns.
|
|
||||||
///
|
|
||||||
/// If `columns` is `None`, all columns are prewarmed.
|
|
||||||
pub async fn prewarm_data(&self, columns: Option<Vec<String>>) -> Result<()> {
|
|
||||||
self.inner.prewarm_data(columns).await
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Poll until the columns are fully indexed. Will return Error::Timeout if the columns
|
/// Poll until the columns are fully indexed. Will return Error::Timeout if the columns
|
||||||
/// are not fully indexed within the timeout.
|
/// are not fully indexed within the timeout.
|
||||||
pub async fn wait_for_index(
|
pub async fn wait_for_index(
|
||||||
@@ -1234,13 +1212,10 @@ pub struct NativeTable {
|
|||||||
// This comes from the connection options. We store here so we can pass down
|
// This comes from the connection options. We store here so we can pass down
|
||||||
// to the dataset when we recreate it (for example, in checkout_latest).
|
// to the dataset when we recreate it (for example, in checkout_latest).
|
||||||
read_consistency_interval: Option<std::time::Duration>,
|
read_consistency_interval: Option<std::time::Duration>,
|
||||||
// Optional namespace client for namespace operations (e.g., managed versioning).
|
// Optional namespace client for server-side query execution.
|
||||||
// pub(crate) so query.rs can access the field for server-side query execution.
|
// When set, queries will be executed on the namespace server instead of locally.
|
||||||
|
// pub (crate) namespace_client so query.rs can access the fields
|
||||||
pub(crate) namespace_client: Option<Arc<dyn LanceNamespace>>,
|
pub(crate) namespace_client: Option<Arc<dyn LanceNamespace>>,
|
||||||
// Whether to enable server-side query execution via the namespace client.
|
|
||||||
// When true and namespace_client is set, queries will be executed on the
|
|
||||||
// namespace server instead of locally.
|
|
||||||
pub(crate) server_side_query_enabled: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl std::fmt::Debug for NativeTable {
|
impl std::fmt::Debug for NativeTable {
|
||||||
@@ -1252,7 +1227,6 @@ impl std::fmt::Debug for NativeTable {
|
|||||||
.field("uri", &self.uri)
|
.field("uri", &self.uri)
|
||||||
.field("read_consistency_interval", &self.read_consistency_interval)
|
.field("read_consistency_interval", &self.read_consistency_interval)
|
||||||
.field("namespace_client", &self.namespace_client)
|
.field("namespace_client", &self.namespace_client)
|
||||||
.field("server_side_query_enabled", &self.server_side_query_enabled)
|
|
||||||
.finish()
|
.finish()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1289,7 +1263,7 @@ impl NativeTable {
|
|||||||
/// * A [NativeTable] object.
|
/// * A [NativeTable] object.
|
||||||
pub async fn open(uri: &str) -> Result<Self> {
|
pub async fn open(uri: &str) -> Result<Self> {
|
||||||
let name = Self::get_table_name(uri)?;
|
let name = Self::get_table_name(uri)?;
|
||||||
Self::open_with_params(uri, &name, vec![], None, None, None, None, false, None).await
|
Self::open_with_params(uri, &name, vec![], None, None, None, None).await
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Opens an existing Table
|
/// Opens an existing Table
|
||||||
@@ -1299,10 +1273,7 @@ impl NativeTable {
|
|||||||
/// * `base_path` - The base path where the table is located
|
/// * `base_path` - The base path where the table is located
|
||||||
/// * `name` The Table name
|
/// * `name` The Table name
|
||||||
/// * `params` The [ReadParams] to use when opening the table
|
/// * `params` The [ReadParams] to use when opening the table
|
||||||
/// * `namespace_client` - Optional namespace client for namespace operations
|
/// * `namespace_client` - Optional namespace client for server-side query execution
|
||||||
/// * `server_side_query_enabled` - Whether to enable server-side query execution
|
|
||||||
/// * `managed_versioning` - Whether managed versioning is enabled. If None and namespace_client
|
|
||||||
/// is provided, the value will be fetched via describe_table.
|
|
||||||
///
|
///
|
||||||
/// # Returns
|
/// # Returns
|
||||||
///
|
///
|
||||||
@@ -1316,8 +1287,6 @@ impl NativeTable {
|
|||||||
params: Option<ReadParams>,
|
params: Option<ReadParams>,
|
||||||
read_consistency_interval: Option<std::time::Duration>,
|
read_consistency_interval: Option<std::time::Duration>,
|
||||||
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
||||||
server_side_query_enabled: bool,
|
|
||||||
managed_versioning: Option<bool>,
|
|
||||||
) -> Result<Self> {
|
) -> Result<Self> {
|
||||||
let params = params.unwrap_or_default();
|
let params = params.unwrap_or_default();
|
||||||
// patch the params if we have a write store wrapper
|
// patch the params if we have a write store wrapper
|
||||||
@@ -1326,54 +1295,17 @@ impl NativeTable {
|
|||||||
None => params,
|
None => params,
|
||||||
};
|
};
|
||||||
|
|
||||||
// Build table_id from namespace + name
|
let dataset = DatasetBuilder::from_uri(uri)
|
||||||
let mut table_id = namespace.clone();
|
.with_read_params(params)
|
||||||
table_id.push(name.to_string());
|
.load()
|
||||||
|
.await
|
||||||
// Determine if managed_versioning is enabled
|
.map_err(|e| match e {
|
||||||
// Use the provided value if available, otherwise query the namespace
|
lance::Error::DatasetNotFound { .. } => Error::TableNotFound {
|
||||||
let managed_versioning = match managed_versioning {
|
name: name.to_string(),
|
||||||
Some(value) => value,
|
source: Box::new(e),
|
||||||
None if namespace_client.is_some() => {
|
},
|
||||||
let ns_client = namespace_client.as_ref().unwrap();
|
e => e.into(),
|
||||||
let describe_request = DescribeTableRequest {
|
})?;
|
||||||
id: Some(table_id.clone()),
|
|
||||||
..Default::default()
|
|
||||||
};
|
|
||||||
let response = ns_client
|
|
||||||
.describe_table(describe_request)
|
|
||||||
.await
|
|
||||||
.map_err(|e| Error::Runtime {
|
|
||||||
message: format!(
|
|
||||||
"Failed to describe table via namespace client: {}. \
|
|
||||||
If you don't need managed versioning, don't pass namespace_client.",
|
|
||||||
e
|
|
||||||
),
|
|
||||||
})?;
|
|
||||||
response.managed_versioning == Some(true)
|
|
||||||
}
|
|
||||||
None => false,
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut builder = DatasetBuilder::from_uri(uri).with_read_params(params);
|
|
||||||
|
|
||||||
// Set up commit handler when managed_versioning is enabled
|
|
||||||
if managed_versioning && let Some(ref ns_client) = namespace_client {
|
|
||||||
let external_store =
|
|
||||||
LanceNamespaceExternalManifestStore::new(ns_client.clone(), table_id.clone());
|
|
||||||
let commit_handler: Arc<dyn CommitHandler> = Arc::new(ExternalManifestCommitHandler {
|
|
||||||
external_manifest_store: Arc::new(external_store),
|
|
||||||
});
|
|
||||||
builder = builder.with_commit_handler(commit_handler);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dataset = builder.load().await.map_err(|e| match e {
|
|
||||||
lance::Error::DatasetNotFound { .. } => Error::TableNotFound {
|
|
||||||
name: name.to_string(),
|
|
||||||
source: Box::new(e),
|
|
||||||
},
|
|
||||||
e => e.into(),
|
|
||||||
})?;
|
|
||||||
|
|
||||||
let dataset = DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval);
|
let dataset = DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval);
|
||||||
let id = Self::build_id(&namespace, name);
|
let id = Self::build_id(&namespace, name);
|
||||||
@@ -1386,7 +1318,6 @@ impl NativeTable {
|
|||||||
dataset,
|
dataset,
|
||||||
read_consistency_interval,
|
read_consistency_interval,
|
||||||
namespace_client,
|
namespace_client,
|
||||||
server_side_query_enabled,
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1490,7 +1421,6 @@ impl NativeTable {
|
|||||||
dataset,
|
dataset,
|
||||||
read_consistency_interval,
|
read_consistency_interval,
|
||||||
namespace_client: stored_namespace_client,
|
namespace_client: stored_namespace_client,
|
||||||
server_side_query_enabled,
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1530,8 +1460,7 @@ impl NativeTable {
|
|||||||
/// * `namespace` - The namespace path. When non-empty, an explicit URI must be provided.
|
/// * `namespace` - The namespace path. When non-empty, an explicit URI must be provided.
|
||||||
/// * `batches` RecordBatch to be saved in the database.
|
/// * `batches` RecordBatch to be saved in the database.
|
||||||
/// * `params` - Write parameters.
|
/// * `params` - Write parameters.
|
||||||
/// * `namespace_client` - Optional namespace client for namespace operations
|
/// * `namespace_client` - Optional namespace client for server-side query execution
|
||||||
/// * `server_side_query_enabled` - Whether to enable server-side query execution
|
|
||||||
///
|
///
|
||||||
/// # Returns
|
/// # Returns
|
||||||
///
|
///
|
||||||
@@ -1546,7 +1475,6 @@ impl NativeTable {
|
|||||||
params: Option<WriteParams>,
|
params: Option<WriteParams>,
|
||||||
read_consistency_interval: Option<std::time::Duration>,
|
read_consistency_interval: Option<std::time::Duration>,
|
||||||
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
||||||
server_side_query_enabled: bool,
|
|
||||||
) -> Result<Self> {
|
) -> Result<Self> {
|
||||||
// Default params uses format v1.
|
// Default params uses format v1.
|
||||||
let params = params.unwrap_or(WriteParams {
|
let params = params.unwrap_or(WriteParams {
|
||||||
@@ -1579,7 +1507,6 @@ impl NativeTable {
|
|||||||
dataset: DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval),
|
dataset: DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval),
|
||||||
read_consistency_interval,
|
read_consistency_interval,
|
||||||
namespace_client,
|
namespace_client,
|
||||||
server_side_query_enabled,
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1593,7 +1520,6 @@ impl NativeTable {
|
|||||||
params: Option<WriteParams>,
|
params: Option<WriteParams>,
|
||||||
read_consistency_interval: Option<std::time::Duration>,
|
read_consistency_interval: Option<std::time::Duration>,
|
||||||
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
namespace_client: Option<Arc<dyn LanceNamespace>>,
|
||||||
server_side_query_enabled: bool,
|
|
||||||
) -> Result<Self> {
|
) -> Result<Self> {
|
||||||
let data: Box<dyn Scannable> = Box::new(RecordBatch::new_empty(schema));
|
let data: Box<dyn Scannable> = Box::new(RecordBatch::new_empty(schema));
|
||||||
Self::create(
|
Self::create(
|
||||||
@@ -1605,7 +1531,6 @@ impl NativeTable {
|
|||||||
params,
|
params,
|
||||||
read_consistency_interval,
|
read_consistency_interval,
|
||||||
namespace_client,
|
namespace_client,
|
||||||
server_side_query_enabled,
|
|
||||||
)
|
)
|
||||||
.await
|
.await
|
||||||
}
|
}
|
||||||
@@ -1709,7 +1634,6 @@ impl NativeTable {
|
|||||||
dataset: DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval),
|
dataset: DatasetConsistencyWrapper::new_latest(dataset, read_consistency_interval),
|
||||||
read_consistency_interval,
|
read_consistency_interval,
|
||||||
namespace_client: stored_namespace_client,
|
namespace_client: stored_namespace_client,
|
||||||
server_side_query_enabled,
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2308,12 +2232,6 @@ impl BaseTable for NativeTable {
|
|||||||
Ok(dataset.prewarm_index(index_name).await?)
|
Ok(dataset.prewarm_index(index_name).await?)
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn prewarm_data(&self, _columns: Option<Vec<String>>) -> Result<()> {
|
|
||||||
Err(Error::NotSupported {
|
|
||||||
message: "prewarm_data is currently only supported on remote tables.".into(),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
async fn update(&self, update: UpdateBuilder) -> Result<UpdateResult> {
|
async fn update(&self, update: UpdateBuilder) -> Result<UpdateResult> {
|
||||||
// Delegate to the submodule implementation
|
// Delegate to the submodule implementation
|
||||||
update::execute_update(self, update).await
|
update::execute_update(self, update).await
|
||||||
@@ -2707,7 +2625,7 @@ mod tests {
|
|||||||
vec![Ok(batch.clone())],
|
vec![Ok(batch.clone())],
|
||||||
batch.schema(),
|
batch.schema(),
|
||||||
));
|
));
|
||||||
let table = NativeTable::create(uri, "test", vec![], reader, None, None, None, None, false)
|
let table = NativeTable::create(uri, "test", vec![], reader, None, None, None, None)
|
||||||
.await
|
.await
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
|
|||||||
@@ -3,13 +3,12 @@
|
|||||||
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use arrow_cast::can_cast_types;
|
|
||||||
use arrow_schema::{DataType, Field, FieldRef, Fields, Schema};
|
use arrow_schema::{DataType, Field, FieldRef, Fields, Schema};
|
||||||
use datafusion::functions::core::{get_field, named_struct};
|
use datafusion::functions::core::{get_field, named_struct};
|
||||||
use datafusion_common::ScalarValue;
|
use datafusion_common::ScalarValue;
|
||||||
use datafusion_common::config::ConfigOptions;
|
use datafusion_common::config::ConfigOptions;
|
||||||
use datafusion_physical_expr::ScalarFunctionExpr;
|
use datafusion_physical_expr::ScalarFunctionExpr;
|
||||||
use datafusion_physical_expr::expressions::{CastExpr, Literal};
|
use datafusion_physical_expr::expressions::{Literal, cast};
|
||||||
use datafusion_physical_plan::expressions::Column;
|
use datafusion_physical_plan::expressions::Column;
|
||||||
use datafusion_physical_plan::projection::ProjectionExec;
|
use datafusion_physical_plan::projection::ProjectionExec;
|
||||||
use datafusion_physical_plan::{ExecutionPlan, PhysicalExpr};
|
use datafusion_physical_plan::{ExecutionPlan, PhysicalExpr};
|
||||||
@@ -26,9 +25,12 @@ pub fn cast_to_table_schema(
|
|||||||
return Ok(input);
|
return Ok(input);
|
||||||
}
|
}
|
||||||
|
|
||||||
let exprs = build_field_exprs(input_schema.fields(), table_schema.fields(), &|idx| {
|
let exprs = build_field_exprs(
|
||||||
Arc::new(Column::new(input_schema.field(idx).name(), idx)) as Arc<dyn PhysicalExpr>
|
input_schema.fields(),
|
||||||
})?;
|
table_schema.fields(),
|
||||||
|
&|idx| Arc::new(Column::new(input_schema.field(idx).name(), idx)) as Arc<dyn PhysicalExpr>,
|
||||||
|
&input_schema,
|
||||||
|
)?;
|
||||||
|
|
||||||
let exprs: Vec<(Arc<dyn PhysicalExpr>, String)> = exprs
|
let exprs: Vec<(Arc<dyn PhysicalExpr>, String)> = exprs
|
||||||
.into_iter()
|
.into_iter()
|
||||||
@@ -49,6 +51,7 @@ fn build_field_exprs(
|
|||||||
input_fields: &Fields,
|
input_fields: &Fields,
|
||||||
table_fields: &Fields,
|
table_fields: &Fields,
|
||||||
get_input_expr: &dyn Fn(usize) -> Arc<dyn PhysicalExpr>,
|
get_input_expr: &dyn Fn(usize) -> Arc<dyn PhysicalExpr>,
|
||||||
|
input_schema: &Schema,
|
||||||
) -> Result<Vec<(Arc<dyn PhysicalExpr>, FieldRef)>> {
|
) -> Result<Vec<(Arc<dyn PhysicalExpr>, FieldRef)>> {
|
||||||
let config = Arc::new(ConfigOptions::default());
|
let config = Arc::new(ConfigOptions::default());
|
||||||
let mut result = Vec::new();
|
let mut result = Vec::new();
|
||||||
@@ -69,19 +72,24 @@ fn build_field_exprs(
|
|||||||
(DataType::Struct(in_children), DataType::Struct(tbl_children))
|
(DataType::Struct(in_children), DataType::Struct(tbl_children))
|
||||||
if in_children != tbl_children =>
|
if in_children != tbl_children =>
|
||||||
{
|
{
|
||||||
let sub_exprs = build_field_exprs(in_children, tbl_children, &|child_idx| {
|
let sub_exprs = build_field_exprs(
|
||||||
let child_name = in_children[child_idx].name();
|
in_children,
|
||||||
Arc::new(ScalarFunctionExpr::new(
|
tbl_children,
|
||||||
&format!("get_field({child_name})"),
|
&|child_idx| {
|
||||||
get_field(),
|
let child_name = in_children[child_idx].name();
|
||||||
vec![
|
Arc::new(ScalarFunctionExpr::new(
|
||||||
input_expr.clone(),
|
&format!("get_field({child_name})"),
|
||||||
Arc::new(Literal::new(ScalarValue::from(child_name.as_str()))),
|
get_field(),
|
||||||
],
|
vec![
|
||||||
Arc::new(in_children[child_idx].as_ref().clone()),
|
input_expr.clone(),
|
||||||
config.clone(),
|
Arc::new(Literal::new(ScalarValue::from(child_name.as_str()))),
|
||||||
)) as Arc<dyn PhysicalExpr>
|
],
|
||||||
})?;
|
Arc::new(in_children[child_idx].as_ref().clone()),
|
||||||
|
config.clone(),
|
||||||
|
)) as Arc<dyn PhysicalExpr>
|
||||||
|
},
|
||||||
|
input_schema,
|
||||||
|
)?;
|
||||||
|
|
||||||
let output_struct_fields: Fields = sub_exprs
|
let output_struct_fields: Fields = sub_exprs
|
||||||
.iter()
|
.iter()
|
||||||
@@ -117,21 +125,17 @@ fn build_field_exprs(
|
|||||||
// Types match: pass through.
|
// Types match: pass through.
|
||||||
(inp, tbl) if inp == tbl => input_expr,
|
(inp, tbl) if inp == tbl => input_expr,
|
||||||
// Types differ: cast.
|
// Types differ: cast.
|
||||||
// safe: false (the default) means overflow/truncation errors surface at execution time.
|
_ => cast(input_expr, input_schema, table_field.data_type().clone()).map_err(|e| {
|
||||||
(_, _) if can_cast_types(input_field.data_type(), table_field.data_type()) => Arc::new(
|
Error::InvalidInput {
|
||||||
CastExpr::new(input_expr, table_field.data_type().clone(), None),
|
|
||||||
)
|
|
||||||
as Arc<dyn PhysicalExpr>,
|
|
||||||
(inp, tbl) => {
|
|
||||||
return Err(Error::InvalidInput {
|
|
||||||
message: format!(
|
message: format!(
|
||||||
"cannot cast field '{}' from {} to {}",
|
"cannot cast field '{}' from {} to {}: {}",
|
||||||
table_field.name(),
|
table_field.name(),
|
||||||
inp,
|
input_field.data_type(),
|
||||||
tbl,
|
table_field.data_type(),
|
||||||
|
e
|
||||||
),
|
),
|
||||||
});
|
}
|
||||||
}
|
})?,
|
||||||
};
|
};
|
||||||
|
|
||||||
let output_field = Arc::new(Field::new(
|
let output_field = Arc::new(Field::new(
|
||||||
@@ -149,12 +153,10 @@ fn build_field_exprs(
|
|||||||
mod tests {
|
mod tests {
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use arrow::buffer::OffsetBuffer;
|
|
||||||
use arrow_array::{
|
use arrow_array::{
|
||||||
Array, Float32Array, Float64Array, Int32Array, Int64Array, ListArray, RecordBatch,
|
Float32Array, Float64Array, Int32Array, Int64Array, RecordBatch, StringArray, StructArray,
|
||||||
StringArray, StructArray, UInt32Array, UInt64Array,
|
|
||||||
};
|
};
|
||||||
use arrow_schema::{DataType, Field, Fields, Schema};
|
use arrow_schema::{DataType, Field, Schema};
|
||||||
use datafusion::prelude::SessionContext;
|
use datafusion::prelude::SessionContext;
|
||||||
use datafusion_catalog::MemTable;
|
use datafusion_catalog::MemTable;
|
||||||
use futures::TryStreamExt;
|
use futures::TryStreamExt;
|
||||||
@@ -493,129 +495,4 @@ mod tests {
|
|||||||
assert_eq!(b.value(0), "hello");
|
assert_eq!(b.value(0), "hello");
|
||||||
assert_eq!(b.value(1), "world");
|
assert_eq!(b.value(1), "world");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_narrowing_numeric_cast_success() {
|
|
||||||
let input_batch = RecordBatch::try_new(
|
|
||||||
Arc::new(Schema::new(vec![Field::new("a", DataType::UInt64, false)])),
|
|
||||||
vec![Arc::new(UInt64Array::from(vec![1u64, 2, 3]))],
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
let table_schema = Schema::new(vec![Field::new("a", DataType::UInt32, false)]);
|
|
||||||
|
|
||||||
let plan = plan_from_batch(input_batch).await;
|
|
||||||
let casted = cast_to_table_schema(plan, &table_schema).unwrap();
|
|
||||||
let result = collect(casted).await;
|
|
||||||
|
|
||||||
assert_eq!(result.schema().field(0).data_type(), &DataType::UInt32);
|
|
||||||
let a: &UInt32Array = result.column(0).as_any().downcast_ref().unwrap();
|
|
||||||
assert_eq!(a.values(), &[1u32, 2, 3]);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_narrowing_numeric_cast_overflow_errors() {
|
|
||||||
let overflow_val = u32::MAX as u64 + 1;
|
|
||||||
let input_batch = RecordBatch::try_new(
|
|
||||||
Arc::new(Schema::new(vec![Field::new("a", DataType::UInt64, false)])),
|
|
||||||
vec![Arc::new(UInt64Array::from(vec![overflow_val]))],
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
let table_schema = Schema::new(vec![Field::new("a", DataType::UInt32, false)]);
|
|
||||||
|
|
||||||
let plan = plan_from_batch(input_batch).await;
|
|
||||||
// Planning succeeds — the overflow is only detected at execution time.
|
|
||||||
let casted = cast_to_table_schema(plan, &table_schema).unwrap();
|
|
||||||
|
|
||||||
let ctx = SessionContext::new();
|
|
||||||
let stream = casted.execute(0, ctx.task_ctx()).unwrap();
|
|
||||||
let result: Result<Vec<RecordBatch>, _> = stream.try_collect().await;
|
|
||||||
assert!(result.is_err(), "expected overflow error at execution time");
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::test]
|
|
||||||
async fn test_list_struct_field_reorder() {
|
|
||||||
// list<struct<a: Int32, b: Int32>> → list<struct<b: Int64, a: Int64>>
|
|
||||||
// Tests both reordering (a,b → b,a) and element-type widening (Int32 → Int64).
|
|
||||||
let inner_fields: Fields = vec![
|
|
||||||
Field::new("a", DataType::Int32, true),
|
|
||||||
Field::new("b", DataType::Int32, true),
|
|
||||||
]
|
|
||||||
.into();
|
|
||||||
let struct_array = StructArray::from(vec![
|
|
||||||
(
|
|
||||||
Arc::new(inner_fields[0].as_ref().clone()),
|
|
||||||
Arc::new(Int32Array::from(vec![1, 3])) as _,
|
|
||||||
),
|
|
||||||
(
|
|
||||||
Arc::new(inner_fields[1].as_ref().clone()),
|
|
||||||
Arc::new(Int32Array::from(vec![2, 4])) as _,
|
|
||||||
),
|
|
||||||
]);
|
|
||||||
// Offsets: one list element containing two struct rows (0..2).
|
|
||||||
let offsets = OffsetBuffer::from_lengths(vec![2]);
|
|
||||||
let list_array = ListArray::try_new(
|
|
||||||
Arc::new(Field::new("item", DataType::Struct(inner_fields), true)),
|
|
||||||
offsets,
|
|
||||||
Arc::new(struct_array),
|
|
||||||
None,
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
let input_batch = RecordBatch::try_new(
|
|
||||||
Arc::new(Schema::new(vec![Field::new(
|
|
||||||
"s_list",
|
|
||||||
list_array.data_type().clone(),
|
|
||||||
false,
|
|
||||||
)])),
|
|
||||||
vec![Arc::new(list_array)],
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
let table_inner: Fields = vec![
|
|
||||||
Field::new("b", DataType::Int64, true),
|
|
||||||
Field::new("a", DataType::Int64, true),
|
|
||||||
]
|
|
||||||
.into();
|
|
||||||
let table_schema = Schema::new(vec![Field::new(
|
|
||||||
"s_list",
|
|
||||||
DataType::List(Arc::new(Field::new(
|
|
||||||
"item",
|
|
||||||
DataType::Struct(table_inner),
|
|
||||||
true,
|
|
||||||
))),
|
|
||||||
false,
|
|
||||||
)]);
|
|
||||||
|
|
||||||
let plan = plan_from_batch(input_batch).await;
|
|
||||||
let casted = cast_to_table_schema(plan, &table_schema).unwrap();
|
|
||||||
let result = collect(casted).await;
|
|
||||||
|
|
||||||
let list_col = result
|
|
||||||
.column(0)
|
|
||||||
.as_any()
|
|
||||||
.downcast_ref::<ListArray>()
|
|
||||||
.unwrap();
|
|
||||||
let struct_col = list_col
|
|
||||||
.values()
|
|
||||||
.as_any()
|
|
||||||
.downcast_ref::<StructArray>()
|
|
||||||
.unwrap();
|
|
||||||
assert_eq!(struct_col.num_columns(), 2);
|
|
||||||
|
|
||||||
let b: &Int64Array = struct_col
|
|
||||||
.column_by_name("b")
|
|
||||||
.unwrap()
|
|
||||||
.as_any()
|
|
||||||
.downcast_ref()
|
|
||||||
.unwrap();
|
|
||||||
assert_eq!(b.values(), &[2, 4]);
|
|
||||||
let a: &Int64Array = struct_col
|
|
||||||
.column_by_name("a")
|
|
||||||
.unwrap()
|
|
||||||
.as_any()
|
|
||||||
.downcast_ref()
|
|
||||||
.unwrap();
|
|
||||||
assert_eq!(a.values(), &[1, 3]);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
use futures::FutureExt;
|
|
||||||
// SPDX-License-Identifier: Apache-2.0
|
// SPDX-License-Identifier: Apache-2.0
|
||||||
// SPDX-FileCopyrightText: Copyright The LanceDB Authors
|
// SPDX-FileCopyrightText: Copyright The LanceDB Authors
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
@@ -24,7 +23,7 @@ pub struct DeleteResult {
|
|||||||
pub(crate) async fn execute_delete(table: &NativeTable, predicate: &str) -> Result<DeleteResult> {
|
pub(crate) async fn execute_delete(table: &NativeTable, predicate: &str) -> Result<DeleteResult> {
|
||||||
table.dataset.ensure_mutable()?;
|
table.dataset.ensure_mutable()?;
|
||||||
let mut dataset = (*table.dataset.get().await?).clone();
|
let mut dataset = (*table.dataset.get().await?).clone();
|
||||||
let delete_result = dataset.delete(predicate).boxed().await?;
|
let delete_result = dataset.delete(predicate).await?;
|
||||||
let num_deleted_rows = delete_result.num_deleted_rows;
|
let num_deleted_rows = delete_result.num_deleted_rows;
|
||||||
let version = dataset.version().version;
|
let version = dataset.version().version;
|
||||||
table.dataset.update(dataset);
|
table.dataset.update(dataset);
|
||||||
|
|||||||
@@ -55,6 +55,7 @@ pub struct MergeInsertBuilder {
|
|||||||
pub(crate) when_not_matched_by_source_delete_filt: Option<String>,
|
pub(crate) when_not_matched_by_source_delete_filt: Option<String>,
|
||||||
pub(crate) timeout: Option<Duration>,
|
pub(crate) timeout: Option<Duration>,
|
||||||
pub(crate) use_index: bool,
|
pub(crate) use_index: bool,
|
||||||
|
pub(crate) mem_wal: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl MergeInsertBuilder {
|
impl MergeInsertBuilder {
|
||||||
@@ -69,6 +70,7 @@ impl MergeInsertBuilder {
|
|||||||
when_not_matched_by_source_delete_filt: None,
|
when_not_matched_by_source_delete_filt: None,
|
||||||
timeout: None,
|
timeout: None,
|
||||||
use_index: true,
|
use_index: true,
|
||||||
|
mem_wal: false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -148,13 +150,65 @@ impl MergeInsertBuilder {
|
|||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Enables MemWAL (Memory Write-Ahead Log) mode for this merge insert operation.
|
||||||
|
///
|
||||||
|
/// When enabled, the merge insert will route data through a memory node service
|
||||||
|
/// that buffers writes before flushing to storage. This is only supported for
|
||||||
|
/// remote (LanceDB Cloud) tables.
|
||||||
|
///
|
||||||
|
/// If not set, defaults to `false`.
|
||||||
|
pub fn mem_wal(&mut self, enabled: bool) -> &mut Self {
|
||||||
|
self.mem_wal = enabled;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
/// Executes the merge insert operation
|
/// Executes the merge insert operation
|
||||||
///
|
///
|
||||||
/// Returns version and statistics about the merge operation including the number of rows
|
/// Returns version and statistics about the merge operation including the number of rows
|
||||||
/// inserted, updated, and deleted.
|
/// inserted, updated, and deleted.
|
||||||
pub async fn execute(self, new_data: Box<dyn RecordBatchReader + Send>) -> Result<MergeResult> {
|
pub async fn execute(self, new_data: Box<dyn RecordBatchReader + Send>) -> Result<MergeResult> {
|
||||||
|
// Validate MemWAL constraints before execution
|
||||||
|
if self.mem_wal {
|
||||||
|
self.validate_mem_wal_pattern()?;
|
||||||
|
}
|
||||||
self.table.clone().merge_insert(self, new_data).await
|
self.table.clone().merge_insert(self, new_data).await
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Validate that the merge insert pattern is supported by MemWAL.
|
||||||
|
///
|
||||||
|
/// MemWAL only supports the upsert pattern:
|
||||||
|
/// - when_matched_update_all (without filter)
|
||||||
|
/// - when_not_matched_insert_all
|
||||||
|
/// - NO when_not_matched_by_source_delete
|
||||||
|
fn validate_mem_wal_pattern(&self) -> Result<()> {
|
||||||
|
// Must have when_matched_update_all without filter
|
||||||
|
if !self.when_matched_update_all {
|
||||||
|
return Err(Error::InvalidInput {
|
||||||
|
message: "MemWAL requires when_matched_update_all() to be set".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
if self.when_matched_update_all_filt.is_some() {
|
||||||
|
return Err(Error::InvalidInput {
|
||||||
|
message: "MemWAL does not support conditional when_matched_update_all (no filter allowed)".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// Must have when_not_matched_insert_all
|
||||||
|
if !self.when_not_matched_insert_all {
|
||||||
|
return Err(Error::InvalidInput {
|
||||||
|
message: "MemWAL requires when_not_matched_insert_all() to be set".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// Must NOT have when_not_matched_by_source_delete
|
||||||
|
if self.when_not_matched_by_source_delete {
|
||||||
|
return Err(Error::InvalidInput {
|
||||||
|
message: "MemWAL does not support when_not_matched_by_source_delete()".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Internal implementation of the merge insert logic
|
/// Internal implementation of the merge insert logic
|
||||||
@@ -165,6 +219,14 @@ pub(crate) async fn execute_merge_insert(
|
|||||||
params: MergeInsertBuilder,
|
params: MergeInsertBuilder,
|
||||||
new_data: Box<dyn RecordBatchReader + Send>,
|
new_data: Box<dyn RecordBatchReader + Send>,
|
||||||
) -> Result<MergeResult> {
|
) -> Result<MergeResult> {
|
||||||
|
if params.mem_wal {
|
||||||
|
return Err(Error::NotSupported {
|
||||||
|
message: "MemWAL is not supported for native (local) tables. \
|
||||||
|
MemWAL is only available for remote (LanceDB Cloud) tables."
|
||||||
|
.to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
let dataset = table.dataset.get().await?;
|
let dataset = table.dataset.get().await?;
|
||||||
let mut builder = LanceMergeInsertBuilder::try_new(dataset.clone(), params.on)?;
|
let mut builder = LanceMergeInsertBuilder::try_new(dataset.clone(), params.on)?;
|
||||||
match (
|
match (
|
||||||
@@ -324,4 +386,139 @@ mod tests {
|
|||||||
merge_insert_builder.execute(new_batches).await.unwrap();
|
merge_insert_builder.execute(new_batches).await.unwrap();
|
||||||
assert_eq!(table.count_rows(None).await.unwrap(), 25);
|
assert_eq!(table.count_rows(None).await.unwrap(), 25);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_mem_wal_validation_valid_pattern() {
|
||||||
|
let conn = connect("memory://").execute().await.unwrap();
|
||||||
|
let batches = merge_insert_test_batches(0, 0);
|
||||||
|
let table = conn
|
||||||
|
.create_table("mem_wal_test", batches)
|
||||||
|
.execute()
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// Valid MemWAL pattern: when_matched_update_all + when_not_matched_insert_all
|
||||||
|
let new_batches = merge_insert_test_batches(5, 1);
|
||||||
|
let mut builder = table.merge_insert(&["i"]);
|
||||||
|
builder.when_matched_update_all(None);
|
||||||
|
builder.when_not_matched_insert_all();
|
||||||
|
builder.mem_wal(true);
|
||||||
|
|
||||||
|
// Should fail because native tables don't support MemWAL, but validation passes
|
||||||
|
let result = builder.execute(new_batches).await;
|
||||||
|
assert!(result.is_err());
|
||||||
|
let err = result.unwrap_err().to_string();
|
||||||
|
assert!(
|
||||||
|
err.contains("MemWAL is not supported for native"),
|
||||||
|
"Expected native table error, got: {}",
|
||||||
|
err
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_mem_wal_validation_missing_when_matched() {
|
||||||
|
let conn = connect("memory://").execute().await.unwrap();
|
||||||
|
let batches = merge_insert_test_batches(0, 0);
|
||||||
|
let table = conn
|
||||||
|
.create_table("mem_wal_test2", batches)
|
||||||
|
.execute()
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// Missing when_matched_update_all
|
||||||
|
let new_batches = merge_insert_test_batches(5, 1);
|
||||||
|
let mut builder = table.merge_insert(&["i"]);
|
||||||
|
builder.when_not_matched_insert_all();
|
||||||
|
builder.mem_wal(true);
|
||||||
|
|
||||||
|
let result = builder.execute(new_batches).await;
|
||||||
|
assert!(result.is_err());
|
||||||
|
let err = result.unwrap_err().to_string();
|
||||||
|
assert!(
|
||||||
|
err.contains("requires when_matched_update_all"),
|
||||||
|
"Expected validation error, got: {}",
|
||||||
|
err
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_mem_wal_validation_missing_when_not_matched() {
|
||||||
|
let conn = connect("memory://").execute().await.unwrap();
|
||||||
|
let batches = merge_insert_test_batches(0, 0);
|
||||||
|
let table = conn
|
||||||
|
.create_table("mem_wal_test3", batches)
|
||||||
|
.execute()
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// Missing when_not_matched_insert_all
|
||||||
|
let new_batches = merge_insert_test_batches(5, 1);
|
||||||
|
let mut builder = table.merge_insert(&["i"]);
|
||||||
|
builder.when_matched_update_all(None);
|
||||||
|
builder.mem_wal(true);
|
||||||
|
|
||||||
|
let result = builder.execute(new_batches).await;
|
||||||
|
assert!(result.is_err());
|
||||||
|
let err = result.unwrap_err().to_string();
|
||||||
|
assert!(
|
||||||
|
err.contains("requires when_not_matched_insert_all"),
|
||||||
|
"Expected validation error, got: {}",
|
||||||
|
err
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_mem_wal_validation_with_filter() {
|
||||||
|
let conn = connect("memory://").execute().await.unwrap();
|
||||||
|
let batches = merge_insert_test_batches(0, 0);
|
||||||
|
let table = conn
|
||||||
|
.create_table("mem_wal_test4", batches)
|
||||||
|
.execute()
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// With conditional filter - not allowed
|
||||||
|
let new_batches = merge_insert_test_batches(5, 1);
|
||||||
|
let mut builder = table.merge_insert(&["i"]);
|
||||||
|
builder.when_matched_update_all(Some("target.age > 0".to_string()));
|
||||||
|
builder.when_not_matched_insert_all();
|
||||||
|
builder.mem_wal(true);
|
||||||
|
|
||||||
|
let result = builder.execute(new_batches).await;
|
||||||
|
assert!(result.is_err());
|
||||||
|
let err = result.unwrap_err().to_string();
|
||||||
|
assert!(
|
||||||
|
err.contains("does not support conditional"),
|
||||||
|
"Expected filter validation error, got: {}",
|
||||||
|
err
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_mem_wal_validation_with_delete() {
|
||||||
|
let conn = connect("memory://").execute().await.unwrap();
|
||||||
|
let batches = merge_insert_test_batches(0, 0);
|
||||||
|
let table = conn
|
||||||
|
.create_table("mem_wal_test5", batches)
|
||||||
|
.execute()
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// With when_not_matched_by_source_delete - not allowed
|
||||||
|
let new_batches = merge_insert_test_batches(5, 1);
|
||||||
|
let mut builder = table.merge_insert(&["i"]);
|
||||||
|
builder.when_matched_update_all(None);
|
||||||
|
builder.when_not_matched_insert_all();
|
||||||
|
builder.when_not_matched_by_source_delete(None);
|
||||||
|
builder.mem_wal(true);
|
||||||
|
|
||||||
|
let result = builder.execute(new_batches).await;
|
||||||
|
assert!(result.is_err());
|
||||||
|
let err = result.unwrap_err().to_string();
|
||||||
|
assert!(
|
||||||
|
err.contains("does not support when_not_matched_by_source_delete"),
|
||||||
|
"Expected delete validation error, got: {}",
|
||||||
|
err
|
||||||
|
);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -64,9 +64,6 @@ pub enum OptimizeAction {
|
|||||||
older_than: Option<Duration>,
|
older_than: Option<Duration>,
|
||||||
/// Because they may be part of an in-progress transaction, files newer than 7 days old are not deleted by default.
|
/// Because they may be part of an in-progress transaction, files newer than 7 days old are not deleted by default.
|
||||||
/// If you are sure that there are no in-progress transactions, then you can set this to True to delete all files older than `older_than`.
|
/// If you are sure that there are no in-progress transactions, then you can set this to True to delete all files older than `older_than`.
|
||||||
///
|
|
||||||
/// **WARNING**: This should only be set to true if you can guarantee that no other process is
|
|
||||||
/// currently working on this dataset. Otherwise the dataset could be put into a corrupted state.
|
|
||||||
delete_unverified: Option<bool>,
|
delete_unverified: Option<bool>,
|
||||||
/// If true, an error will be returned if there are any old versions that are still tagged.
|
/// If true, an error will be returned if there are any old versions that are still tagged.
|
||||||
error_if_tagged_old_versions: Option<bool>,
|
error_if_tagged_old_versions: Option<bool>,
|
||||||
@@ -120,10 +117,6 @@ pub(crate) async fn optimize_indices(table: &NativeTable, options: &OptimizeOpti
|
|||||||
/// If you are sure that there are no in-progress transactions, then you
|
/// If you are sure that there are no in-progress transactions, then you
|
||||||
/// can set this to True to delete all files older than `older_than`.
|
/// can set this to True to delete all files older than `older_than`.
|
||||||
///
|
///
|
||||||
/// **WARNING**: This should only be set to true if you can guarantee that
|
|
||||||
/// no other process is currently working on this dataset. Otherwise the
|
|
||||||
/// dataset could be put into a corrupted state.
|
|
||||||
///
|
|
||||||
/// This calls into [lance::dataset::Dataset::cleanup_old_versions] and
|
/// This calls into [lance::dataset::Dataset::cleanup_old_versions] and
|
||||||
/// returns the result.
|
/// returns the result.
|
||||||
pub(crate) async fn cleanup_old_versions(
|
pub(crate) async fn cleanup_old_versions(
|
||||||
|
|||||||
@@ -40,10 +40,8 @@ pub async fn execute_query(
|
|||||||
query: &AnyQuery,
|
query: &AnyQuery,
|
||||||
options: QueryExecutionOptions,
|
options: QueryExecutionOptions,
|
||||||
) -> Result<DatasetRecordBatchStream> {
|
) -> Result<DatasetRecordBatchStream> {
|
||||||
// If server-side query is enabled and namespace client is configured, use server-side query execution
|
// If namespace client is configured, use server-side query execution
|
||||||
if table.server_side_query_enabled
|
if let Some(ref namespace_client) = table.namespace_client {
|
||||||
&& let Some(ref namespace_client) = table.namespace_client
|
|
||||||
{
|
|
||||||
return execute_namespace_query(table, namespace_client.clone(), query, options).await;
|
return execute_namespace_query(table, namespace_client.clone(), query, options).await;
|
||||||
}
|
}
|
||||||
execute_generic_query(table, query, options).await
|
execute_generic_query(table, query, options).await
|
||||||
@@ -186,13 +184,6 @@ pub async fn create_plan(
|
|||||||
Select::Dynamic(ref select_with_transform) => {
|
Select::Dynamic(ref select_with_transform) => {
|
||||||
scanner.project_with_transform(select_with_transform.as_slice())?;
|
scanner.project_with_transform(select_with_transform.as_slice())?;
|
||||||
}
|
}
|
||||||
Select::Expr(ref expr_pairs) => {
|
|
||||||
let sql_pairs: crate::Result<Vec<(String, String)>> = expr_pairs
|
|
||||||
.iter()
|
|
||||||
.map(|(name, expr)| expr_to_sql_string(expr).map(|sql| (name.clone(), sql)))
|
|
||||||
.collect();
|
|
||||||
scanner.project_with_transform(sql_pairs?.as_slice())?;
|
|
||||||
}
|
|
||||||
Select::All => {}
|
Select::All => {}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -347,17 +338,6 @@ fn convert_to_namespace_query(query: &AnyQuery) -> Result<NsQueryTableRequest> {
|
|||||||
.to_string(),
|
.to_string(),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
Select::Expr(pairs) => {
|
|
||||||
let sql_pairs: crate::Result<Vec<(String, String)>> = pairs
|
|
||||||
.iter()
|
|
||||||
.map(|(name, expr)| expr_to_sql_string(expr).map(|sql| (name.clone(), sql)))
|
|
||||||
.collect();
|
|
||||||
let sql_pairs = sql_pairs?;
|
|
||||||
Some(Box::new(QueryTableRequestColumns {
|
|
||||||
column_names: None,
|
|
||||||
column_aliases: Some(sql_pairs.into_iter().collect()),
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// Check for unsupported features
|
// Check for unsupported features
|
||||||
@@ -429,17 +409,6 @@ fn convert_to_namespace_query(query: &AnyQuery) -> Result<NsQueryTableRequest> {
|
|||||||
.to_string(),
|
.to_string(),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
Select::Expr(pairs) => {
|
|
||||||
let sql_pairs: crate::Result<Vec<(String, String)>> = pairs
|
|
||||||
.iter()
|
|
||||||
.map(|(name, expr)| expr_to_sql_string(expr).map(|sql| (name.clone(), sql)))
|
|
||||||
.collect();
|
|
||||||
let sql_pairs = sql_pairs?;
|
|
||||||
Some(Box::new(QueryTableRequestColumns {
|
|
||||||
column_names: None,
|
|
||||||
column_aliases: Some(sql_pairs.into_iter().collect()),
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// Handle full text search if present
|
// Handle full text search if present
|
||||||
|
|||||||
Reference in New Issue
Block a user