Fix studio setup installing deps into system Python instead of venv#4400
Fix studio setup installing deps into system Python instead of venv#4400danielhanchen wants to merge 7 commits intomainfrom
Conversation
|
Warning You have reached your daily quota limit. Please wait up to 24 hours and I will start processing your requests again! |
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: d905e3e690
ℹ️ About Codex in GitHub
Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
Codex can also answer questions or update the PR. Try commenting "@codex address that feedback".
studio/setup.ps1
Outdated
| Write-Host " Installing uv package manager..." -ForegroundColor Cyan | ||
| powershell -ExecutionPolicy ByPass -c "irm https://astral.sh/uv/install.ps1 | iex" 2>&1 | Out-Null | ||
| # Refresh PATH to pick up uv | ||
| Refresh-Environment |
There was a problem hiding this comment.
Re-activate venv after refreshing PATH
When uv is not already installed, this call resets PATH to machine+user values (see Refresh-Environment at the top of this script), which drops the activated venv’s Scripts directory from the process path. As a result, subsequent (Get-Command python).Source and python "$PSScriptRoot\install_python_stack.py" can resolve to the system interpreter, so dependencies are installed outside ~/.unsloth/studio/.venv on fresh Windows setups—the same class of regression this change is trying to prevent.
Useful? React with 👍 / 👎.
Skip the frontend npm install + build if frontend/dist already exists. Previously setup.ps1 nuked node_modules and package-lock.json on every run, and both scripts always rebuilt even when dist/ was already present. On a git clone editable install, the first setup run still builds the frontend as before. Subsequent runs skip it, saving several minutes. To force a rebuild, delete frontend/dist and re-run setup.
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: 657cd0bd3b
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
studio/setup.sh
Outdated
| # Create venv under ~/.unsloth/studio/ (shared location, not in repo) | ||
| # All platforms (including Colab) use the same isolated venv so that | ||
| # studio dependencies are never installed into the system Python. | ||
| STUDIO_HOME="$HOME/.unsloth/studio" | ||
| VENV_DIR="$STUDIO_HOME/.venv" |
There was a problem hiding this comment.
Keep Colab installs in the notebook interpreter
When setup.sh is run via !./studio/setup.sh in the official studio/Unsloth_Studio_Colab.ipynb, the source "$VENV_DIR/bin/activate" later in this block only affects that subshell. The next notebook cell imports from colab import start in Colab's system Python, and that path immediately needs installed packages such as structlog (studio/backend/loggers/config.py). By moving Colab onto the shared ~/.unsloth/studio/.venv path, the notebook kernel no longer sees the dependencies that install_python_stack.py just installed, so the documented Colab startup flow breaks after setup completes.
Useful? React with 👍 / 👎.
Instead of only checking if dist/ exists, compare source file timestamps against the dist/ directory. If any file in frontend/src/ is newer than dist/, trigger a rebuild. This handles the case where a developer pulls new frontend changes and re-runs setup -- stale assets get rebuilt automatically.
Address review feedback: - Check entire frontend/ directory for changes, not just src/. The build also depends on package.json, vite.config.ts, tailwind.config.ts, public/, and other config files. A change to any of these now triggers a rebuild. - Move oxc-validator npm install outside the frontend build gate in setup.sh so it always runs on setup, matching setup.ps1 which already had it outside the gate.
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: 41ac58186f
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
| for pkg in ( | ||
| f"transformers=={TRANSFORMERS_5_VERSION}", | ||
| "huggingface_hub==1.7.1", | ||
| "hf_xet==1.4.2", |
There was a problem hiding this comment.
Rebuild .venv_t5 when its package set changes
Now that .venv_t5 is expected to contain huggingface_hub==1.7.1 and hf_xet==1.4.2, _ensure_venv_t5_exists() still accepts any non-empty ~/.unsloth/studio/.venv_t5 as valid. Upgraders who already have the old two-package directory from a previous Studio release will skip this install block entirely and keep running with the stale 1.3.0 layout, so the first 5.x training/inference/export path still reproduces the missing-dependency failure until they manually delete the directory.
Useful? React with 👍 / 👎.
| run_quiet "install transformers 5.x" fast_install --target "$VENV_T5_DIR" --no-deps "transformers==5.3.0" | ||
| run_quiet "install huggingface_hub for t5" fast_install --target "$VENV_T5_DIR" --no-deps "huggingface_hub==1.7.1" | ||
| run_quiet "install hf_xet for t5" fast_install --target "$VENV_T5_DIR" --no-deps "hf_xet==1.4.2" |
There was a problem hiding this comment.
Build Colab's .venv_t5 with the notebook interpreter
On Colab, the notebook still launches Studio from the system kernel (Unsloth_Studio_Colab.ipynb imports from colab import start after setup), but these installs run after re-activating $VENV_DIR, so .venv_t5 is populated for whatever BEST_PY the venv used. If the image exposes a newer python3.x than the notebook kernel, the new hf_xet native module in .venv_t5 is built for the wrong ABI and the later sys.path prepend in core/*/worker.py can fail once a 5.x model download needs the xet backend.
Useful? React with 👍 / 👎.
Addresses reviewer feedback on the frontend caching logic: 1. setup.sh: Fix broken find command that caused exit under pipefail. The piped `find | xargs find -newer` had paths after the expression which GNU find rejects. Replaced with a simpler `find -maxdepth 1 -type f -newer dist/` that checks ALL top-level files (catches index.html, bun.lock, etc. that the extension allowlist missed). 2. setup.sh: Guard oxc-validator npm install behind `command -v npm` check. When the frontend build is skipped (dist/ is cached), Node bootstrap is also skipped, so npm may not be available. 3. setup.ps1: Replace Get-ChildItem -Include with explicit path probing for src/ and public/. PowerShell's -Include without a trailing wildcard silently returns nothing, so src/public changes were never detected. Also check ALL top-level files instead of just .json/.ts/.js/.mjs extensions.
41ac581 to
2b1e53d
Compare
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: 2b1e53d1bd
ℹ️ About Codex in GitHub
Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
Codex can also answer questions or update the PR. Try commenting "@codex address that feedback".
| _changed=$(find "$SCRIPT_DIR/frontend/src" "$SCRIPT_DIR/frontend/public" \ | ||
| -type f -newer "$SCRIPT_DIR/frontend/dist" 2>/dev/null | head -1) |
There was a problem hiding this comment.
Avoid pipefail exit when probing changed frontend files
Because the script runs with set -euo pipefail, the find ... | head -1 probe can terminate setup with exit 141 when more than one file is newer than frontend/dist (head exits early, find gets SIGPIPE, pipeline becomes non-zero). In that case setup aborts before it reaches the frontend rebuild path, so upgrades with multiple changed frontend files can fail unexpectedly instead of rebuilding.
Useful? React with 👍 / 👎.
- All platforms (including Colab) now create ~/.unsloth/studio/.venv with --without-pip fallback for broken ensurepip environments - Add --python sys.executable to uv pip install in install_python_stack.py so uv targets the correct venv instead of system Python - Centralize .venv_t5 bootstrap in transformers_version.py with proper validation (checks required packages exist, not just non-empty dir) - Replace ~150 lines of duplicated install code across 3 worker files with calls to the shared _ensure_venv_t5_exists() helper - Use uv-if-present with pip fallback; do not install uv at runtime - Add site.addsitedir() shim in colab.py so notebook cells can import studio packages from the venv without system-Python double-install - Update .venv_t5 packages: huggingface_hub 1.3.0->1.7.1, add hf_xet - Bump transformers pin 4.57.1->4.57.6 in requirements + constraints - Add Fast-Install helper to setup.ps1 with uv+pip fallback - Keep Colab-specific completion banner in setup.sh
for more information, see https://pre-commit.ci
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: 446f829da4
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
| venv_lib = Path.home() / ".unsloth" / "studio" / ".venv" / "lib" | ||
| for sp in venv_lib.glob("python*/site-packages"): | ||
| if str(sp) not in sys.path: | ||
| site.addsitedir(str(sp)) |
There was a problem hiding this comment.
Prepend the Studio venv before system site-packages
On Colab, site.addsitedir() appends the venv's site-packages after the notebook kernel's existing entries. If the kernel already has packages like fastapi, transformers, or huggingface_hub installed, from run import run_server / main.py will keep importing those system copies instead of the versions just installed into ~/.unsloth/studio/.venv, so the new Colab path can still run against the wrong dependency set.
Useful? React with 👍 / 👎.
| _VENV_T5_DIR, | ||
| "--no-deps", |
There was a problem hiding this comment.
Upgrade existing .venv_t5 contents during runtime repair
When a user already has an older .venv_t5 and the worker repairs it on a machine without uv on PATH, this fallback uses pip install --target without --upgrade. pip install --help explicitly says --target installs do not replace existing files/folders unless --upgrade is given, so stale packages such as the previous huggingface_hub can survive and the new self-heal path still leaves .venv_t5 inconsistent.
Useful? React with 👍 / 👎.
|
/gemini review |
There was a problem hiding this comment.
Code Review
This pull request effectively addresses a critical issue with dependency installation on Colab by ensuring the studio's virtual environment is consistently used. The changes are well-implemented, particularly the unified venv creation in setup.sh, the robust fallback for ensurepip, and the clever _bootstrap_studio_venv for the Colab notebook environment. The refactoring to prefer uv for installations is a great performance enhancement. I have one suggestion to further improve the robustness of the setup script.
| cd "$SCRIPT_DIR/backend/core/data_recipe/oxc-validator" | ||
| run_quiet "npm install (oxc validator runtime)" npm install | ||
| cd "$SCRIPT_DIR" |
There was a problem hiding this comment.
Using cd can be risky in scripts with set -e, as an interruption or failure in the middle of the block could leave the script in an unexpected directory. Using pushd and popd is a more robust way to handle temporary directory changes, as popd will always return to the previous directory from the stack.
| cd "$SCRIPT_DIR/backend/core/data_recipe/oxc-validator" | |
| run_quiet "npm install (oxc validator runtime)" npm install | |
| cd "$SCRIPT_DIR" | |
| pushd "$SCRIPT_DIR/backend/core/data_recipe/oxc-validator" >/dev/null | |
| run_quiet "npm install (oxc validator runtime)" npm install | |
| popd >/dev/null |
… CPU support (#4413) * Allow Windows setup to complete without NVIDIA GPU setup.ps1 previously hard-exited if nvidia-smi was not found, blocking setup entirely on CPU-only or non-NVIDIA machines. The backend already supports CPU and MLX (Apple Silicon) in chat-only GGUF mode, and the Linux/Mac setup.sh handles missing GPUs gracefully. Changes: - Convert the GPU check from a hard exit to a warning - Guard CUDA toolkit installation behind $HasNvidiaSmi - Install CPU-only PyTorch when no GPU is detected - Build llama.cpp without CUDA flags when no GPU is present - Update doc comment to reflect CPU support * Cache frontend build across setup runs Skip the frontend npm install + build if frontend/dist already exists. Previously setup.ps1 nuked node_modules and package-lock.json on every run, and both scripts always rebuilt even when dist/ was already present. On a git clone editable install, the first setup run still builds the frontend as before. Subsequent runs skip it, saving several minutes. To force a rebuild, delete frontend/dist and re-run setup. * Show pip progress for PyTorch download on Windows The torch CUDA wheel is ~2.8 GB and the CPU wheel is ~300 MB. With | Out-Null suppressing all output, the install appeared completely frozen with no feedback. Remove | Out-Null for the torch install lines so pip's download progress bar is visible. Add a size hint so users know the download is expected to take a while. Also moves the Triton success message inside the GPU branch so it only prints when Triton was actually installed. * Guard CUDA env re-sanitization behind GPU check in llama.cpp build The CUDA_PATH re-sanitization block (lines 1020-1033) references $CudaToolkitRoot which is only set when $HasNvidiaSmi is true and the CUDA Toolkit section runs. On CPU-only machines, $CudaToolkitRoot is null, causing Split-Path to throw: Split-Path : Cannot bind argument to parameter 'Path' because it is null. Wrap the entire block in `if ($HasNvidiaSmi -and $CudaToolkitRoot)`. * Rebuild frontend when source files are newer than dist/ Instead of only checking if dist/ exists, compare source file timestamps against the dist/ directory. If any file in frontend/src/ is newer than dist/, trigger a rebuild. This handles the case where a developer pulls new frontend changes and re-runs setup -- stale assets get rebuilt automatically. * Fix cmake not found on Windows after winget install Two issues fixed: 1. After winget installs cmake, Refresh-Environment may not pick up the new PATH entry (MSI PATH changes sometimes need a new shell). Added a fallback that probes cmake's default install locations (Program Files, LocalAppData) and adds the directory to PATH explicitly if found. 2. If cmake is still unavailable when the llama.cpp build starts (e.g. winget failed silently or PATH was not updated), the build now skips gracefully with a [SKIP] warning instead of crashing with "cmake : The term 'cmake' is not recognized". * Fix frontend rebuild detection and decouple oxc-validator install Address review feedback: - Check entire frontend/ directory for changes, not just src/. The build also depends on package.json, vite.config.ts, tailwind.config.ts, public/, and other config files. A change to any of these now triggers a rebuild. - Move oxc-validator npm install outside the frontend build gate in setup.sh so it always runs on setup, matching setup.ps1 which already had it outside the gate. * Show cmake errors on failure and retry CUDA VS integration with elevation Two fixes for issue #4405 (Windows setup fails at cmake configure): 1. cmake configure: capture output and display it on failure instead of piping to Out-Null. When the error mentions "No CUDA toolset found", print a hint about the CUDA VS integration files. 2. CUDA VS integration copy: when the direct Copy-Item fails (needs admin access to write to Program Files), retry with Start-Process -Verb RunAs to prompt for elevation. This is the root cause of the "No CUDA toolset found" cmake failure -- the .targets files that let MSBuild compile .cu files are missing from the VS BuildCustomizations directory. * Address reviewer feedback: cmake PATH persistence, stale cache, torch error check 1. Persist cmake PATH to user registry so Refresh-Environment cannot drop it later in the same setup run. Previously the process-only PATH addition at phase 1 could vanish when Refresh-Environment rebuilt PATH from registry during phase 2/3 installs. 2. Clean stale CMake cache before configure. If a previous run built with CUDA and the user reruns without a GPU (or vice versa), the cached GGML_CUDA value would persist. Now the build dir is removed before configure. 3. Explicitly set -DGGML_CUDA=OFF for CPU-only builds instead of just omitting CUDA flags. This prevents cmake from auto-detecting a partial CUDA installation. 4. Fix CUDA cmake flag indentation -- was misaligned from the original PR, now consistently indented inside the if/else block. 5. Fail hard if pip install torch returns a non-zero exit code instead of silently continuing with a broken environment. * Remove extra CUDA cmake flags to align Windows with Linux build Drop GGML_CUDA_FA_ALL_QUANTS, GGML_CUDA_F16, GGML_CUDA_GRAPHS, GGML_CUDA_FORCE_CUBLAS, and GGML_CUDA_PEER_MAX_BATCH_SIZE flags. The Linux build in setup.sh only sets GGML_CUDA=ON and lets llama.cpp use its defaults for everything else. Keep Windows consistent. * Address reviewer round 2: GPU probe fallback, Triton check, stale binary rebuild 1. GPU detection: fallback to default nvidia-smi install locations (Program Files\NVIDIA Corporation\NVSMI, System32) when nvidia-smi is not on PATH. Prevents silent CPU-only provisioning on machines that have a GPU but a broken PATH. 2. Triton: check $LASTEXITCODE after pip install and print [WARN] on failure instead of unconditional [OK]. 3. Stale llama-server: check CMakeCache.txt for GGML_CUDA setting and rebuild if the existing binary does not match the current GPU mode (e.g. CUDA binary on a now-CPU-only rerun, or vice versa). * Fix frontend rebuild detection and npm dependency issues Addresses reviewer feedback on the frontend caching logic: 1. setup.sh: Fix broken find command that caused exit under pipefail. The piped `find | xargs find -newer` had paths after the expression which GNU find rejects. Replaced with a simpler `find -maxdepth 1 -type f -newer dist/` that checks ALL top-level files (catches index.html, bun.lock, etc. that the extension allowlist missed). 2. setup.sh: Guard oxc-validator npm install behind `command -v npm` check. When the frontend build is skipped (dist/ is cached), Node bootstrap is also skipped, so npm may not be available. 3. setup.ps1: Replace Get-ChildItem -Include with explicit path probing for src/ and public/. PowerShell's -Include without a trailing wildcard silently returns nothing, so src/public changes were never detected. Also check ALL top-level files instead of just .json/.ts/.js/.mjs extensions. * Fix studio setup: venv isolation, centralized .venv_t5, uv targeting - All platforms (including Colab) now create ~/.unsloth/studio/.venv with --without-pip fallback for broken ensurepip environments - Add --python sys.executable to uv pip install in install_python_stack.py so uv targets the correct venv instead of system Python - Centralize .venv_t5 bootstrap in transformers_version.py with proper validation (checks required packages exist, not just non-empty dir) - Replace ~150 lines of duplicated install code across 3 worker files with calls to the shared _ensure_venv_t5_exists() helper - Use uv-if-present with pip fallback; do not install uv at runtime - Add site.addsitedir() shim in colab.py so notebook cells can import studio packages from the venv without system-Python double-install - Update .venv_t5 packages: huggingface_hub 1.3.0->1.7.1, add hf_xet - Bump transformers pin 4.57.1->4.57.6 in requirements + constraints - Add Fast-Install helper to setup.ps1 with uv+pip fallback - Keep Colab-specific completion banner in setup.sh * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Fix nvidia-smi PATH persistence and cmake requirement for CPU-only 1. Store nvidia-smi as an absolute path ($NvidiaSmiExe) on first detection. All later calls (Get-CudaComputeCapability, Get-PytorchCudaTag, CUDA toolkit detection) use this absolute path instead of relying on PATH. This survives Refresh-Environment which rebuilds PATH from the registry and drops process-only additions. 2. Make cmake fatal for CPU-only installs. CPU-only machines depend entirely on llama-server for GGUF chat mode, so reporting "Setup Complete!" without it is misleading. GPU machines can still skip the llama-server build since they have other inference paths. * Fix broken frontend freshness detection in setup scripts - setup.sh: Replace broken `find | xargs find -newer` pipeline with single `find ... -newer` call. The old pipeline produced "paths must precede expression" errors (silently suppressed by 2>/dev/null), causing top-level config changes to never trigger a rebuild. - setup.sh: Add `command -v npm` guard to oxc-validator block so it does not fail when Node was not installed (build-skip path). - setup.ps1: Replace `Get-ChildItem -Include` (unreliable without -Recurse on PS 5.1) with explicit directory paths for src/ and public/ scanning. - Both: Add *.html to tracked file patterns so index.html (Vite entry point) changes trigger a rebuild. - Both: Use -print -quit instead of piping to head -1 for efficiency. * Fix bugs found during review of PRs #4404, #4400, #4399 - setup.sh: Add || true guard to find command that checks frontend/src and frontend/public dirs, preventing script abort under set -euo pipefail when either directory is missing - colab.py: Use sys.path.insert(0, ...) instead of site.addsitedir() so Studio venv packages take priority over system copies. Add warning when venv is missing instead of silently failing. - transformers_version.py: _venv_t5_is_valid() now checks installed package versions via .dist-info metadata, not just directory presence. Prevents false positives from stale or wrong-version packages. - transformers_version.py: _install_to_venv_t5() now passes --upgrade so pip replaces existing stale packages in the target directory. - setup.ps1: CPU-only PyTorch install uses --index-url for cpu wheel and all install commands use Fast-Install (uv with pip fallback). * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Fix _venv_t5_is_valid dist-info loop exiting after first directory Remove premature break that caused the loop over .dist-info directories to exit after the first match even if it had no METADATA file. Now continues iterating until a valid METADATA is found or all dirs are exhausted. * Capture error output on failure instead of discarding with Out-Null setup.ps1: 6 locations changed from `| Out-Null` to `| Out-String` with output shown on failure -- PyTorch GPU/CPU install, Triton install, venv_t5 package loop, cmake llama-server and llama-quantize builds. transformers_version.py: clean stale .venv_t5 directory before reinstall when validation detects missing or version-mismatched packages. * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Fix ModuleNotFoundError when CLI imports studio.backend.core The backend uses bare "from utils.*" imports everywhere, relying on backend/ being on sys.path. Workers and routes add it at startup, but the CLI imports studio.backend.core as a package -- backend/ was never added. Add sys.path setup at the top of core/__init__.py so lazy imports resolve correctly regardless of entry point. Fixes: unsloth inference unsloth/Qwen3-8B "who are you" crashing with "No module named 'utils'" * Fix frontend freshness check to detect all top-level file changes The extension allowlist (*.json, *.ts, *.js, *.mjs, *.html) missed files like bun.lock, so lockfile-only dependency changes could skip the frontend rebuild. Check all top-level files instead. * Add tiktoken to .venv_t5 for Qwen-family tokenizers Qwen models use tiktoken-based tokenizers which fail when routed through the transformers 5.x overlay without tiktoken installed. Add it to the setup scripts (with deps for Windows) and runtime fallback list. Integrates PR #4418. * Fix tiktoken crash in _venv_t5_is_valid and stray brace in setup.ps1 _venv_t5_is_valid() crashed with ValueError on unpinned packages like "tiktoken" (no ==version). Handle by splitting safely and skipping version check for unpinned packages (existence check only). Also remove stray closing brace in setup.ps1 tiktoken install block. --------- Co-authored-by: Daniel Han <danielhanchen@users.noreply.github.com> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Summary
unsloth studio setupon Colab skipped venv creation entirely and installed all studio dependencies (structlog, fastapi, data-designer, etc.) into the system Python. Users with their own venv gotModuleNotFoundErroron launch because the deps were not visible in their environment.uv pip installsilently ignored the activated venv and targeted the system Python unless explicitly told which Python to use via--python.Root cause
Two issues in two files:
studio/setup.shhad a Colab-specific branch (IS_COLAB=true) that skipped creating~/.unsloth/studio/.venvand raninstall_python_stackagainst the bare system Python. Every other platform (Windows, Linux, Mac) already created the venv correctly.studio/install_python_stack.pybuiltuv pip installcommands without specifying--python. On environments like Colab,uvresolved to the system Python even when a venv was activated (uvprintedUsing Python 3.12.12 environment at: /usrdespiteVIRTUAL_ENVbeing set).Changes
setup.sh:IS_COLABspecial case that skipped venv creation~/.unsloth/studio/.venvconsistently--without-pipfallback withget-pip.pybootstrap for environments whereensurepipis broken (Colab)install_python_stack.py:--python sys.executableto alluv pip installcommands_bootstrap_uv()probe to also use--pythonfor consistent detectionTest plan
Tested on Colab with a clean venv (only pip installed):
pip install unslothinstalls cleanlyunsloth studio setupcreates~/.unsloth/studio/.venvwith all deps (structlog, fastapi, data-designer, torch, etc.)unsloth studio -H 0.0.0.0 -p 8888launches successfully ("Launching with studio venv...", frontend loaded, server running)uvinstalls target the studio venv, not the system Python