mirror of
https://github.com/dlt-hub/dlt.git
synced 2025-12-17 19:31:30 +00:00
Fixes historic builds (#3412)
* fixes historic builds * fix broken link * constrain docs build env to python 3.10 * switch snippets testing to python 3.10 * allows python up to py3.12 in docs project --------- Co-authored-by: Marcin Rudolf <rudolfix@rudolfix.org>
This commit is contained in:
2
.github/workflows/build_docs.yml
vendored
2
.github/workflows/build_docs.yml
vendored
@@ -24,7 +24,7 @@ jobs:
|
||||
- name: Install uv
|
||||
uses: astral-sh/setup-uv@v6
|
||||
with:
|
||||
python-version: "3.11"
|
||||
python-version: "3.10"
|
||||
activate-environment: true
|
||||
|
||||
- name: Install node dependencies
|
||||
|
||||
4
.github/workflows/test_docs.yml
vendored
4
.github/workflows/test_docs.yml
vendored
@@ -66,12 +66,12 @@ jobs:
|
||||
- name: Setup Python
|
||||
uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: "3.11"
|
||||
python-version: "3.10"
|
||||
|
||||
- name: Install uv
|
||||
uses: astral-sh/setup-uv@v6
|
||||
with:
|
||||
python-version: "3.11"
|
||||
python-version: "3.10"
|
||||
activate-environment: true
|
||||
enable-cache: true
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@ name = "dlt-docs-tools"
|
||||
version = "0.1.0"
|
||||
description = "Add your description here"
|
||||
readme = "README.md"
|
||||
requires-python = ">=3.9.2, <3.14"
|
||||
requires-python = ">=3.10, <3.13" # databind not available over 3.10, we need to update the build
|
||||
|
||||
# NOTE: working here is always dev enviroment, so we don't need a dev group
|
||||
dependencies = [
|
||||
@@ -42,6 +42,7 @@ dependencies = [
|
||||
"weaviate-client>=3.26.7,<4.0.0",
|
||||
"regex>=2025.10.23",
|
||||
"pytest-forked>=1.6.0",
|
||||
"databind>=4.5.2",
|
||||
]
|
||||
|
||||
|
||||
|
||||
2421
docs/uv.lock
generated
2421
docs/uv.lock
generated
File diff suppressed because it is too large
Load Diff
@@ -19,7 +19,7 @@ You can run dlt within Snowflake AI Data Cloud in a few different ways.
|
||||
Package dlt into a container and run it on SPCS for in-Snowflake execution (available starting with Snowpark 1.43.0).
|
||||
|
||||
4. **External runner with Snowflake as destination**
|
||||
[Run dlt outside](../../walkthroughs/deploy-a-pipeline) Snowflake (your infra or a third party) and load into Snowflake as the destination.
|
||||
[Run dlt outside](../walkthroughs/deploy-a-pipeline) Snowflake (your infra or a third party) and load into Snowflake as the destination.
|
||||
|
||||
## How to choose
|
||||
|
||||
|
||||
@@ -97,10 +97,10 @@ for (const version of selectedVersions) {
|
||||
|
||||
// build doc version, we also run preprocessing and markdown gen for each doc version
|
||||
console.log(`Building docs...`)
|
||||
proc.execSync(`cd ${REPO_DOCS_DIR} && npm run preprocess-docs:cloudflare && PYTHONPATH=. pydoc-markdown && python clean_pydoc_sidebar.py`)
|
||||
proc.execSync(`cd ${REPO_DOCS_DIR} && cd .. && make preprocess-docs`)
|
||||
|
||||
console.log(`Snapshotting version...`)
|
||||
proc.execSync(`cd ${REPO_DOCS_DIR} && npm run docusaurus docs:version ${version}`)
|
||||
proc.execSync(`cd ${REPO_DOCS_DIR} && npx docusaurus docs:version ${version}`)
|
||||
|
||||
console.log(`Moving snapshot`)
|
||||
fs.cpSync(REPO_DOCS_DIR+"/"+VERSIONED_DOCS_FOLDER, VERSIONED_DOCS_FOLDER, {recursive: true})
|
||||
|
||||
Reference in New Issue
Block a user