Files
ComfyUI_frontend/browser_tests
snomiao 2240645a7d [feat] Add Cloudflare Pages deployment for Playwright test reports (#5045)
* [feat] Add Cloudflare Pages deployment for Playwright test reports

- Deploy test reports to separate Cloudflare projects per browser
- Add real-time PR comments with progressive test status updates
- Use wrangler-action for unified Cloudflare tooling
- Support cross-browser testing with individual report links
- Document CI/CD integration in browser_tests/README.md

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* Fix Cloudflare project name for chromium-0.5x browser

* Extract project name transformation to variable for consistent URL formatting

* chore(ci): update branch filters for push and pull_request events in test-ui workflow to refine CI triggers

* [feat] Improve test-ui deployment with branch isolation and building page

- Use Cloudflare Pages --branch flag for proper branch isolation instead of modifying project names
- Add auto-refresh building page that shows test progress in real-time
- Deploy building page immediately when tests start for instant feedback
- Update URL generation to use branch-based Cloudflare Pages URLs format
- Maintain clean project names while isolating branches properly

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore(test-ui.yaml): increase sleep duration from 5 to 10 seconds for cache propagation and restore cached setup steps for improved workflow efficiency

* [refactor] Remove building-page to reduce complexity

- Remove auto-refresh building page and related deployment steps
- Simplify PR comments to show basic test status without progress page
- Keep branch-based deployment for proper isolation while reducing complexity
- Maintain clean workflow focused on core functionality

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore(test-ui.yaml): add a separator in the workflow file for better readability and organization of the test status section

* [feat] Add Cloudflare Pages deployment for Playwright test reports

- Deploy test reports to separate Cloudflare projects per browser
- Add real-time PR comments with progressive test status updates
- Use wrangler-action for unified Cloudflare tooling
- Support cross-browser testing with individual report links
- Document CI/CD integration in browser_tests/README.md

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* Fix Cloudflare project name for chromium-0.5x browser

* Extract project name transformation to variable for consistent URL formatting

* chore(ci): update branch filters for push and pull_request events in test-ui workflow to refine CI triggers

* [feat] Improve test-ui deployment with branch isolation and building page

- Use Cloudflare Pages --branch flag for proper branch isolation instead of modifying project names
- Add auto-refresh building page that shows test progress in real-time
- Deploy building page immediately when tests start for instant feedback
- Update URL generation to use branch-based Cloudflare Pages URLs format
- Maintain clean project names while isolating branches properly

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore(test-ui.yaml): increase sleep duration from 5 to 10 seconds for cache propagation and restore cached setup steps for improved workflow efficiency

* [refactor] Remove building-page to reduce complexity

- Remove auto-refresh building page and related deployment steps
- Simplify PR comments to show basic test status without progress page
- Keep branch-based deployment for proper isolation while reducing complexity
- Maintain clean workflow focused on core functionality

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore(test-ui.yaml): add a separator in the workflow file for better readability and organization of the test status section

* [fix] Address PR review feedback - improve workflow architecture and security

- [HIGH] Fix continue-on-error masking by adding final test result check that fails CI on test failures
- [MEDIUM] Move branch sanitization to setup job to reduce performance overhead
- [MEDIUM] Add compatibility-date to Cloudflare deployment for stability
- [LOW] Extract date format to environment variable to follow DRY principle
- [LOW] Quote shell variables properly to prevent word splitting
- [LOW] Update documentation to use dynamic branch-specific URLs

Addresses all security, performance, and code quality issues raised in automated PR review.
Maintains test report deployment while ensuring CI integrity.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore(test-ui.yaml): replace loading emoji with an image for better visual consistency in test logs
style(test-ui.yaml): clean up whitespace in the workflow file for improved readability

* style(test-ui.yaml): format message to combine two lines into one for better readability

* chore(test-ui.yaml): add a blank line for better readability in the workflow file

* style(test-ui.yaml): update loading image alt text and format messages for better readability in GitHub Actions workflow

* [architecture] Separate test execution from deployment - clean CI design

BREAKING: Remove continue-on-error from test execution for proper CI integrity

**Clean Architecture Changes:**
- Remove `continue-on-error: true` from Playwright test execution
- Create separate `deploy-reports` job that always runs for debugging
- Test jobs now properly fail when tests fail (maintains CI integrity)
- Reports still deploy for debugging via dedicated deployment job
- Capture and pass actual exit codes between jobs via artifacts

**Benefits:**
-  CI fails when tests fail (no longer masked)
-  Reports still deploy for debugging regardless of test outcome
-  Clean separation of concerns (test vs deploy responsibilities)
-  Proper job dependencies and error handling
-  Individual browser test results preserved

**Job Flow:**
1. `setup` - Cache and prepare environment
2. `playwright-tests` - Run tests, fail if tests fail, upload artifacts
3. `deploy-reports` - Always deploy reports using artifacts (parallel)
4. `comment-summary` - Generate summary and fail workflow if needed

This addresses the high-priority architecture concern about continue-on-error
masking test failures while maintaining report deployment functionality.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* [refactor] Simplify deployment architecture - remove over-engineering

**Reverted to clean, simple approach based on feedback:**

1.  **Faster deployment** - Deploy immediately after each test (no waiting for matrix completion)
2.  **Remove unnecessary GITHUB_OUTPUT** - Don't save exit codes, use step.conclusion instead
3.  **Single job approach** - Use `if: always()` instead of separate deploy-reports job

**Key Changes:**
- Removed separate `deploy-reports` job (86 lines deleted!)
- Deploy in same job with `if: always()` - much faster
- Use `steps.playwright.conclusion` instead of captured exit codes
- Cleaner, simpler architecture with same functionality

**Benefits:**
- 🚀 **Much faster** - Reports deploy immediately per browser, not waiting for all tests
- 🧹 **Simpler** - One job handles test + deploy, easier to understand
-  **Still maintains CI integrity** - Tests fail properly when they should
- 📊 **Reports always deploy** - Available for debugging regardless of test outcome

The previous approach was over-engineered. This is much cleaner and faster.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* fix(workflow): reorder condition in PR comment step for clarity and consistency

* chore(test-ui.yaml): update deployment command to remove compatibility date for better maintainability
docs(test-ui.yaml): add note to always() condition for clarity on artifact upload behavior

* [performance] Remove redundant branch sanitization - 75% processing reduction

**Issue**: Complex bash string operations running 4 times per build in matrix jobs
**Solution**: Remove duplicate branch sanitization, use pre-computed value from setup job

**Before**: Branch sanitization ran in both setup job AND each matrix job (5 total times)
**After**: Branch sanitization runs only once in setup job, reused via outputs

**Performance Impact**:
- 4 redundant tr/sed operations eliminated (matrix chromium, chromium-2x, chromium-0.5x, mobile-chrome)
- 75% reduction in branch name processing overhead
- Cleaner, more maintainable code

**Implementation**:
- Setup job: Computes `sanitized-branch` output once
- Matrix jobs: Use `${{ needs.setup.outputs.sanitized-branch }}` directly
- No duplicate string processing logic

Addresses PR review comment: [performance] medium Priority - Complex bash string operations in GitHub Actions matrix

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

---------

Co-authored-by: Claude <noreply@anthropic.com>
2025-08-19 10:22:45 -07:00
..
2025-08-14 14:53:46 -07:00
2024-12-30 17:26:37 -05:00
2024-12-30 17:26:37 -05:00
2024-12-30 17:26:37 -05:00

Playwright Testing for ComfyUI_frontend

This document outlines the setup, usage, and common patterns for Playwright browser tests in the ComfyUI_frontend project.

Prerequisites

CRITICAL: Start ComfyUI backend with --multi-user flag:

python main.py --multi-user

Without this flag, parallel tests will conflict and fail randomly.

Setup

ComfyUI devtools

Clone https://github.com/Comfy-Org/ComfyUI_devtools to your custom_nodes directory.
ComfyUI_devtools adds additional API endpoints and nodes to ComfyUI for browser testing.

Node.js & Playwright Prerequisites

Ensure you have Node.js v20 or v22 installed. Then, set up the Chromium test driver:

npx playwright install chromium --with-deps

Environment Configuration

Create .env from the template:

cp .env_example .env

Key settings for debugging:

# Remove Vue dev overlay that blocks UI elements
DISABLE_VUE_PLUGINS=true

# Test against dev server (recommended) or backend directly
PLAYWRIGHT_TEST_URL=http://localhost:5173  # Dev server
# PLAYWRIGHT_TEST_URL=http://localhost:8188  # Direct backend

# Path to ComfyUI for backing up user data/settings before tests
TEST_COMFYUI_DIR=/path/to/your/ComfyUI

Common Setup Issues

Most tests require the new menu system - Add to your test:

test.beforeEach(async ({ comfyPage }) => {
  await comfyPage.setSetting('Comfy.UseNewMenu', 'Top')
})

Release API Mocking

By default, all tests mock the release API (api.comfy.org/releases) to prevent release notification popups from interfering with test execution. This is necessary because the release notifications can appear over UI elements and block test interactions.

To test with real release data, you can disable mocking:

await comfyPage.setup({ mockReleases: false })

For tests that specifically need to test release functionality, see the example in tests/releaseNotifications.spec.ts.

Running Tests

Always use UI mode for development:

npx playwright test --ui

UI mode features:

  • Locator picker: Click the target icon, then click any element to get the exact locator code to use in your test. The code appears in the Locator tab.
  • Step debugging: Step through your test line-by-line by clicking Source tab
  • Time travel: In the Actions tab/panel, click any step to see the browser state at that moment
  • Console/Network Tabs: View logs and API calls at each step
  • Attachments Tab: View all snapshots with expected and actual images

Playwright UI Mode

For CI or headless testing:

npx playwright test                    # Run all tests
npx playwright test widget.spec.ts     # Run specific test file

Local Development Config

For debugging, you can try adjusting these settings in playwright.config.ts:

export default defineConfig({
  // VERY HELPFUL: Skip screenshot tests locally
  grep: process.env.CI ? undefined : /^(?!.*screenshot).*$/

  retries: 0, // No retries while debugging. Increase if writing new tests. that may be flaky.
  workers: 1, // Single worker for easier debugging. Increase to match CPU cores if you want to run a lot of tests in parallel.
  timeout: 30000, // Longer timeout for breakpoints

  use: {
    trace: 'on', // Always capture traces (CI uses 'on-first-retry')
    video: 'on' // Always record video (CI uses 'retain-on-failure')
  },

})

Test Structure

Browser tests in this project follow a specific organization pattern:

  • Fixtures: Located in fixtures/ - These provide test setup and utilities

    • ComfyPage.ts - The main fixture for interacting with ComfyUI
    • ComfyMouse.ts - Utility for mouse interactions with the canvas
    • Components fixtures in fixtures/components/ - Page object models for UI components
  • Tests: Located in tests/ - The actual test specifications

    • Organized by functionality (e.g., widget.spec.ts, interaction.spec.ts)
    • Snapshot directories (e.g., widget.spec.ts-snapshots/) contain reference screenshots
  • Utilities: Located in utils/ - Common utility functions

    • litegraphUtils.ts - Utilities for working with LiteGraph nodes

Writing Effective Tests

When writing new tests, follow these patterns:

Test Structure

// Import the test fixture
import { comfyPageFixture as test } from '../fixtures/ComfyPage'

test.describe('Feature Name', () => {
  // Set up test environment if needed
  test.beforeEach(async ({ comfyPage }) => {
    // Common setup
  })

  test('should do something specific', async ({ comfyPage }) => {
    // Test implementation
  })
})

Leverage Existing Fixtures and Helpers

Always check for existing helpers and fixtures before implementing new ones:

  • ComfyPage: Main fixture with methods for canvas interaction and node management
  • ComfyMouse: Helper for precise mouse operations on the canvas
  • Helpers: Check browser_tests/helpers/ for specialized helpers like:
    • actionbar.ts: Interact with the action bar
    • manageGroupNode.ts: Group node management operations
    • templates.ts: Template workflows operations
  • Component Fixtures: Check browser_tests/fixtures/components/ for UI component helpers
  • Utility Functions: Check browser_tests/utils/ and browser_tests/fixtures/utils/ for shared utilities

Most common testing needs are already addressed by these helpers, which will make your tests more consistent and reliable.

Key Testing Patterns

  1. Focus elements explicitly: Canvas-based elements often need explicit focus before interaction:

    // Click the canvas first to focus it before pressing keys
    await comfyPage.canvas.click()
    await comfyPage.page.keyboard.press('a')
    
  2. Mark canvas as dirty if needed: Some interactions need explicit canvas updates:

    // After programmatically changing node state, mark canvas dirty
    await comfyPage.page.evaluate(() => {
      window['app'].graph.setDirtyCanvas(true, true)
    })
    
  3. Use node references over coordinates: Node references from fixtures/utils/litegraphUtils.ts provide stable ways to interact with nodes:

    // Prefer this:
    const node = await comfyPage.getNodeRefsByType('LoadImage')[0]
    await node.click('title')
    
    // Over this:
    await comfyPage.canvas.click({ position: { x: 100, y: 100 } })
    
  4. Wait for canvas to render after UI interactions:

    await comfyPage.nextFrame()
    
  5. Clean up persistent server state: While most state is reset between tests, anything stored on the server persists:

    // Reset settings that affect other tests (these are stored on server)
    await comfyPage.setSetting('Comfy.ColorPalette', 'dark')
    await comfyPage.setSetting('Comfy.NodeBadge.NodeIdBadgeMode', 'None')
    
    // Clean up uploaded files if needed
    await comfyPage.request.delete(`${comfyPage.url}/api/delete/image.png`)
    
  6. Prefer functional assertions over screenshots: Use screenshots only when visual verification is necessary:

    // Prefer this:
    expect(await node.isPinned()).toBe(true)
    expect(await node.getProperty('title')).toBe('Expected Title')
    
    // Over this - only use when needed:
    await expect(comfyPage.canvas).toHaveScreenshot('state.png')
    
  7. Use minimal test workflows: When creating test workflows, keep them as minimal as possible:

    // Include only the components needed for the test
    await comfyPage.loadWorkflow('single_ksampler')
    
  8. Debug helpers for visual debugging (remove before committing):

    ComfyPage includes temporary debug methods for troubleshooting:

    test('debug failing interaction', async ({ comfyPage }, testInfo) => {
      // Add visual markers to see click positions
      await comfyPage.debugAddMarker({ x: 100, y: 200 })
    
      // Attach screenshot with markers to test report
      await comfyPage.debugAttachScreenshot(testInfo, 'node-positions', {
        element: 'canvas',
        markers: [{ position: { x: 100, y: 200 } }]
      })
    
      // Show canvas overlay for easier debugging
      await comfyPage.debugShowCanvasOverlay()
    
      // Remember to remove debug code before committing!
    })
    

    Available debug methods:

    • debugAddMarker(position) - Red circle at position
    • debugAttachScreenshot(testInfo, name) - Attach to test report
    • debugShowCanvasOverlay() - Show canvas as overlay
    • debugGetCanvasDataURL() - Get canvas as base64

Common Patterns and Utilities

Page Object Pattern

Tests use the Page Object pattern to create abstractions over the UI:

// Using the ComfyPage fixture
test('Can toggle boolean widget', async ({ comfyPage }) => {
  await comfyPage.loadWorkflow('widgets/boolean_widget')
  const node = (await comfyPage.getFirstNodeRef())!
  const widget = await node.getWidget(0)
  await widget.click()
})

Node References

The NodeReference class provides helpers for interacting with LiteGraph nodes:

// Getting node by type and interacting with it
const nodes = await comfyPage.getNodeRefsByType('LoadImage')
const loadImageNode = nodes[0]
const widget = await loadImageNode.getWidget(0)
await widget.click()

Visual Regression Testing

Tests use screenshot comparisons to verify UI state:

// Take a screenshot and compare to reference
await expect(comfyPage.canvas).toHaveScreenshot('boolean_widget_toggled.png')

Waiting for Animations

Always call nextFrame() after actions that trigger animations:

await comfyPage.canvas.click({ position: { x: 100, y: 100 } })
await comfyPage.nextFrame() // Wait for canvas to redraw

Mouse Interactions

Canvas operations use special helpers to ensure proper timing:

// Using ComfyMouse for drag and drop
await comfyMouse.dragAndDrop(
  { x: 100, y: 100 }, // From
  { x: 200, y: 200 } // To
)

// Standard ComfyPage helpers
await comfyPage.drag({ x: 100, y: 100 }, { x: 200, y: 200 })
await comfyPage.pan({ x: 200, y: 200 })
await comfyPage.zoom(-100) // Zoom in

Workflow Management

Tests use workflows stored in assets/ for consistent starting points:

// Load a test workflow
await comfyPage.loadWorkflow('single_ksampler')

// Wait for workflow to load and stabilize
await comfyPage.nextFrame()

Custom Assertions

The project includes custom Playwright assertions through comfyExpect:

// Check if a node is in a specific state
await expect(node).toBePinned()
await expect(node).toBeBypassed()
await expect(node).toBeCollapsed()

Troubleshooting Common Issues

Flaky Tests

  • Timing Issues: Always wait for animations to complete with nextFrame()
  • Coordinate Sensitivity: Canvas coordinates are viewport-relative; use node references when possible
  • Test Isolation: Tests run in parallel; avoid dependencies between tests
  • Screenshots vary: Ensure your OS and browser match the reference environment (Linux)
  • Async / await: Race conditions are a very common cause of test flakiness

Screenshot Testing

Due to variations in system font rendering, screenshot expectations are platform-specific. Please note:

  • Do not commit local screenshot expectations to the repository
  • We maintain Linux screenshot expectations as our GitHub Action runner operates in a Linux environment
  • While developing, you can generate local screenshots for your tests, but these will differ from CI-generated ones

Working with Screenshots Locally

Option 1 - Skip screenshot tests (add to playwright.config.ts):

export default defineConfig({
  grep: process.env.CI ? undefined : /^(?!.*screenshot).*$/
})

Option 2 - Generate local baselines for comparison:

npx playwright test --update-snapshots

Getting Test Artifacts from GitHub Actions

When tests fail in CI, you can download screenshots and traces:

  1. Go to the failed workflow run in GitHub Actions
  2. Scroll to "Artifacts" section at the bottom
  3. Download playwright-report or test-results
  4. Extract and open the HTML report locally
  5. View actual vs expected screenshots and execution traces

Creating New Screenshot Baselines

For PRs from Comfy-Org/ComfyUI_frontend branches:

  1. Write test with toHaveScreenshot('filename.png')
  2. Create PR and add New Browser Test Expectation label
  3. CI will generate and commit the Linux baseline screenshots

Note: Fork PRs cannot auto-commit screenshots. A maintainer will need to commit the screenshots manually for you (don't worry, they'll do it).

CI/CD Integration

Automated Test Deployment

The project automatically deploys Playwright test reports to Cloudflare Pages for every PR and push to main branches. This provides:

  • Live test reports with interactive HTML views
  • Cross-browser testing across chromium, mobile-chrome, and different viewport sizes
  • Real-time PR comments with test status and links to detailed reports

How it works:

  1. Test execution: All browser tests run in parallel across multiple browsers

  2. Report generation: HTML reports are generated for each browser configuration

  3. Cloudflare deployment: Each browser's report deploys to its own Cloudflare Pages project with branch isolation:

    • comfyui-playwright-chromium (with branch-specific URLs)
    • comfyui-playwright-mobile-chrome (with branch-specific URLs)
    • comfyui-playwright-chromium-2x (2x scale, with branch-specific URLs)
    • comfyui-playwright-chromium-0-5x (0.5x scale, with branch-specific URLs)
  4. PR comments: GitHub automatically updates PR comments with:

    • / Test status for each browser
    • Direct links to interactive test reports
    • Real-time progress updates as tests complete

Accessing test reports:

  • From PR comments: Click the "View Report" links for each browser
  • From GitHub Actions: Download artifacts from failed runs
  • Direct URLs: Reports are available at https://[branch].comfyui-playwright-[browser].pages.dev (branch-specific deployments)

Report features:

  • Interactive HTML reports with test results, screenshots, and traces
  • Detailed failure analysis with before/after screenshots
  • Test execution videos for failed tests
  • Network logs and console output for debugging

This integration ensures that test results are easily accessible to reviewers and maintainers, making it simple to verify that changes don't break existing functionality across different browsers and viewport sizes.

Resources