4203 Commits

Author SHA1 Message Date
John Alva
ef3c64449a Remove planning and PR helper docs from PR; keep copies on branch preinstall-enhancements-docs 2025-11-11 16:09:59 -06:00
John Alva
5b4c2ff924 UX: preflight banner, fast torch detection, weekly full-check, non-blocking prompts, optional updates, auto port selection and browser open 2025-11-11 11:18:02 -06:00
John-Caldwell
0443944dbb Merge branch 'preinstall-enhancements' of https://github.com/John-Caldwell/ComfyUI into preinstall-enhancements 2025-11-11 08:41:15 -06:00
John-Caldwell
d90159f58b Update ASCII art banner to display COMFY 2025-11-11 08:40:53 -06:00
John-Caldwell
5818270ea3 Add screenshots directory and finalize PR for review 2025-11-11 01:35:31 -06:00
John-Caldwell
cfee48d3b7 Update checklist: Feature Request issue #10705 created and referenced 2025-11-10 18:09:47 -06:00
John-Caldwell
a6769cb56e Add issue number #10705 to PR description
Feature Request issue created successfully. PR now references the issue as required by contribution guidelines.
2025-11-10 18:08:34 -06:00
John-Caldwell
04139fe528 Update submission checklist and add final summary
- Updated PR_SUBMISSION_CHECKLIST.md with Feature Request issue creation step
- Added FINAL_SUMMARY.md with complete status and next steps
- All documentation and preparation complete
2025-11-10 18:00:29 -06:00
John-Caldwell
55b7ea2bbd Add Feature Request issue content and update PR description for compliance
- Created FEATURE_REQUEST_ISSUE.md with complete issue content
- Created CREATE_ISSUE_INSTRUCTIONS.md with step-by-step instructions
- Created PR_COMPLIANCE_ANALYSIS.md analyzing compliance with wiki requirements
- Created SEARCH_RESULTS_SUMMARY.md documenting search for existing issues/PRs
- Updated PR_DESCRIPTION.md with:
  - Issue Addressed section (explicitly states problems solved)
  - Potential Concerns and Side Effects section (required by wiki)
  - PR Size Note (acknowledges large PR size)
  - Placeholder for Related Issue number
2025-11-10 17:59:35 -06:00
John-Caldwell
2bfd78c46a Add PR submission checklist and screenshot recommendations 2025-11-10 17:21:41 -06:00
John-Caldwell
1a56b1dcea Add comprehensive PR description for preinstall enhancements 2025-11-10 17:20:54 -06:00
John-Caldwell
52d13ef3a8 Add plan document for preinstall enhancements PR 2025-11-10 16:30:23 -06:00
John-Caldwell
f65290f9a5 Add create_shortcut.ps1 for desktop shortcut creation 2025-11-10 11:54:36 -06:00
John-Caldwell
1365bbf859 Enhanced run_comfyui.bat with UTF-8 encoding, progress bars, and CUDA PyTorch auto-installation
- Added UTF-8 encoding (chcp 65001) to fix Unicode character display in ASCII art header

- Enabled progress bars for all pip installations (--progress-bar on)

- Fixed CUDA PyTorch auto-installation logic to properly continue to ComfyUI launch

- Updated CUDA availability variables after successful installation

- Fixed misleading Restart message to accurately reflect Continue to launch

- Improved error handling and user feedback throughout the installation process
2025-11-10 11:53:36 -06:00
John-Caldwell
0f93e63be4 Add enhanced batch file with optional dependency checking
Enhanced run_comfyui.bat with:
- Automatic detection of missing critical dependencies
- Virtual environment detection and warnings
- Optional user-prompted installation with clear warnings
- Comprehensive dependency checking for all essential packages

NOTE: Author is not a professional coder and relied heavily on Cursor AI for implementation. Please review code thoroughly before merging.
2025-11-10 10:27:52 -06:00
comfyanonymous
2abd2b5c20
Make ScaleROPE node work on Flux. (#10686) 2025-11-08 15:52:02 -05:00
comfyanonymous
a1a70362ca
Only unpin tensor if it was pinned by ComfyUI (#10677) 2025-11-07 11:15:05 -05:00
rattus
cf97b033ee
mm: guard against double pin and unpin explicitly (#10672)
As commented, if you let cuda be the one to detect double pin/unpinning
it actually creates an asyc GPU error.
2025-11-06 21:20:48 -05:00
comfyanonymous
eb1c42f649
Tell users they need to upload their logs in bug reports. (#10671) 2025-11-06 20:24:28 -05:00
comfyanonymous
e05c907126
Clarify release cycle. (#10667) 2025-11-06 04:11:30 -05:00
comfyanonymous
09dc24c8a9
Pinned mem also seems to work on AMD. (#10658) 2025-11-05 19:11:15 -05:00
comfyanonymous
1d69245981
Enable pinned memory by default on Nvidia. (#10656)
Removed the --fast pinned_memory flag.

You can use --disable-pinned-memory to disable it. Please report if it
causes any issues.
2025-11-05 18:08:13 -05:00
comfyanonymous
97f198e421
Fix qwen controlnet regression. (#10657) 2025-11-05 18:07:35 -05:00
Alexander Piskun
bda0eb2448
feat(API-nodes): move Rodin3D nodes to new client; removed old api client.py (#10645) 2025-11-05 02:16:00 -08:00
comfyanonymous
c4a6b389de
Lower ltxv mem usage to what it was before previous pr. (#10643)
Bring back qwen behavior to what it was before previous pr.
2025-11-04 22:47:35 -05:00
contentis
4cd881866b
Use single apply_rope function across models (#10547) 2025-11-04 20:10:11 -05:00
comfyanonymous
265adad858 ComfyUI version v0.3.68 v0.3.68 2025-11-04 19:42:23 -05:00
comfyanonymous
7f3e4d486c
Limit amount of pinned memory on windows to prevent issues. (#10638) 2025-11-04 17:37:50 -05:00
rattus
a389ee01bb
caching: Handle None outputs tuple case (#10637) 2025-11-04 14:14:10 -08:00
ComfyUI Wiki
9c71a66790
chore: update workflow templates to v0.2.11 (#10634) 2025-11-04 10:51:53 -08:00
comfyanonymous
af4b7b5edb
More fp8 torch.compile regressions fixed. (#10625) 2025-11-03 22:14:20 -05:00
comfyanonymous
0f4ef3afa0
This seems to slow things down slightly on Linux. (#10624) 2025-11-03 21:47:14 -05:00
comfyanonymous
6b88478f9f
Bring back fp8 torch compile performance to what it should be. (#10622) 2025-11-03 19:22:10 -05:00
comfyanonymous
e199c8cc67
Fixes (#10621) 2025-11-03 17:58:24 -05:00
comfyanonymous
0652cb8e2d
Speed up torch.compile (#10620) 2025-11-03 17:37:12 -05:00
comfyanonymous
958a17199a
People should update their pytorch versions. (#10618) 2025-11-03 17:08:30 -05:00
ComfyUI Wiki
e974e554ca
chore: update embedded docs to v0.3.1 (#10614) 2025-11-03 10:59:44 -08:00
Alexander Piskun
4e2110c794
feat(Pika-API-nodes): use new API client (#10608) 2025-11-03 00:29:08 -08:00
Alexander Piskun
e617cddf24
convert nodes_openai.py to V3 schema (#10604) 2025-11-03 00:28:13 -08:00
Alexander Piskun
1f3f7a2823
convert nodes_hypernetwork.py to V3 schema (#10583) 2025-11-03 00:21:47 -08:00
EverNebula
88df172790
fix(caching): treat bytes as hashable (#10567) 2025-11-03 00:16:40 -08:00
Alexander Piskun
6d6a18b0b7
fix(api-nodes-cloud): stop using sub-folder and absolute path for output of Rodin3D nodes (#10556) 2025-11-03 00:04:56 -08:00
comfyanonymous
97ff9fae7e
Clarify help text for --fast argument (#10609)
Updated help text for the --fast argument to clarify potential risks.
2025-11-02 13:14:04 -05:00
rattus
135fa49ec2
Small speed improvements to --async-offload (#10593)
* ops: dont take an offload stream if you dont need one

* ops: prioritize mem transfer

The async offload streams reason for existence is to transfer from
RAM to GPU. The post processing compute steps are a bonus on the side
stream, but if the compute stream is running a long kernel, it can
stall the side stream, as it wait to type-cast the bias before
transferring the weight. So do a pure xfer of the weight straight up,
then do everything bias, then go back to fix the weight type and do
weight patches.
2025-11-01 18:48:53 -04:00
comfyanonymous
44869ff786
Fix issue with pinned memory. (#10597) 2025-11-01 17:25:59 -04:00
Alexander Piskun
20182a393f
convert StabilityAI to use new API client (#10582) 2025-11-01 12:14:06 -07:00
Alexander Piskun
5f109fe6a0
added 12s-20s as available output durations for the LTXV API nodes (#10570) 2025-11-01 12:13:39 -07:00
comfyanonymous
c58c13b2ba
Fix torch compile regression on fp8 ops. (#10580) 2025-11-01 00:25:17 -04:00
comfyanonymous
7f374e42c8
ScaleROPE now works on Lumina models. (#10578) 2025-10-31 15:41:40 -04:00
comfyanonymous
27d1bd8829
Fix rope scaling. (#10560) 2025-10-30 22:51:58 -04:00