Run FLUX, Stable Diffusion, and Train LoRAs — One-Click GPU Setup
Setting up an image generation environment with FLUX, Stable Diffusion, and LoRA training involves installing ComfyUI, downloading multiple model checkpoints (each several gigabytes), configuring Kohya for training, and ensuring all dependencies resolve cleanly. Depending on your starting point, this takes one to several hours.
We built a VM image with everything pre-installed. FLUX and Stable Diffusion models are downloaded. ComfyUI is configured. Kohya training tools are ready. You deploy the VM, open port 9093 in your browser, and start generating or training.
What You Get
The Barrack ComfyUI SD-FLUX-KOHYA image ships with:
- FLUX.1 Schnell — Black Forest Labs' fast image generation model (
flux1-schnell-fp8.safetensors), already downloaded - Stable Diffusion — SDXL and SD model support, pre-configured
- Kohya — LoRA training toolkit for fine-tuning models on your own images
- ComfyUI — node-based visual interface, serving on port 9093
- Pre-built workflow templates — including the FLUX Schnell workflow, ready to load
- Ubuntu 22.04 with NVIDIA drivers and CUDA pre-configured
All services start automatically on boot.
Compatible GPUs
This image is available on the following GPUs:
| GPU | VRAM |
|---|---|
| RTX A6000 | 48 GB |
| L40 | 48 GB |
| A100 PCIe | 80 GB |
| H100 PCIe | 80 GB |
| H100 PCIe NVLink | 80 GB |
FLUX Schnell in fp8 quantization runs comfortably on all of the above. LoRA training benefits from higher VRAM — the A100 and H100 allow larger batch sizes and faster training. On the RTX A6000 (48GB), FLUX Schnell generates a 1024×1024 image in approximately 10–15 seconds at 4 inference steps.
Step 1 — Create an Account
Go to barrack.ai/signup. Register with email or Google OAuth.
Complete your billing profile at My Account: full name, billing address, postal code, and country.
Purchase credits. Minimum deposit: $5.00 (USD), €5.00 (EUR), or ₹100.00 (INR).
Step 2 — Deploy the VM
- Go to barrack.ai/dashboard
- Select your GPU — this image is compatible with RTX A6000, L40, A100 PCIe, H100 PCIe, and H100 PCIe NVLink
- Set GPU count to 1
- In the OS Image dropdown, select Barrack ComfyUI SD-FLUX-KOHYA
- Create or select an SSH key
- Click Deploy
The VM enters provisioning within seconds. Wait until the status shows Active.
Prefer API deployment? See the API deployment documentation.
Step 3 — Find Your IP Address
- Go to barrack.ai/dashboard
- Click the dropdown at the top of the page
- Select your instance
- Click Details
- Your public IP address is displayed there
Public IP is automatically enabled for this image.
Step 4 — Open ComfyUI
Open your browser and navigate to:
http://YOUR_PUBLIC_IP:9093
This loads the ComfyUI interface with an empty canvas.
Step 5 — Load a Workflow and Generate
Generate with FLUX
- Click Workflow in the top menu
- Click Load
- Select the
flux_schnell_workflowtemplate - The workflow loads with all nodes connected: Load Checkpoint → CLIP Text Encode → KSampler → VAE Decode → Save Image
- The FLUX Schnell model (
flux1-schnell-fp8.safetensors) is pre-selected in the checkpoint node - Enter your prompt in the CLIP Text Encode (Positive Prompt) node
- Click Queue Prompt or the Run button
- Image generates in approximately 10–15 seconds
Example Prompt
a bottle with a beautiful rainbow galaxy inside it on top of a wooden table in the middle of a modern kitchen, photorealistic, 8K
Default settings: 1024×1024 resolution, 4 steps, Euler sampler.
Generate with Stable Diffusion
ComfyUI supports loading Stable Diffusion checkpoints (SDXL, SD 1.5, SD 3.5) in the same interface. Swap the model in the Load Checkpoint node to switch between FLUX and SD models. Additional SD checkpoints can be downloaded to ~/ComfyUI/models/checkpoints/ via SSH.
Train LoRAs with Kohya
LoRA (Low-Rank Adaptation) lets you fine-tune a model on a small set of reference images — typically 20–50 images — to learn a specific style, character, or object. The resulting LoRA file is small (typically 100–300 MB) and can be loaded into ComfyUI alongside the base model.
Kohya training tools are pre-installed on this image. To use them:
- SSH into your VM:
ssh ubuntu@YOUR_PUBLIC_IP
- Prepare your training images in a directory on the VM. Upload them via
scp:
scp -r ./my_training_images/ ubuntu@YOUR_PUBLIC_IP:~/training_data/
-
Run Kohya training with your images. Training a LoRA on 50 images takes approximately 30 minutes on the RTX A6000.
-
Once training completes, move the output LoRA file to the ComfyUI models directory:
cp output.safetensors ~/ComfyUI/models/loras/
- In ComfyUI, add a Load LoRA node to your workflow and select your trained file. Generate images using your custom-trained style.
SSH Access
Connect to the VM for file management or training:
ssh ubuntu@YOUR_PUBLIC_IP
Generated images are saved in:
ls ~/ComfyUI/output/
Download generated images:
scp ubuntu@YOUR_PUBLIC_IP:~/ComfyUI/output/your_image.png ./
When to Use This
- Image generation — produce images with FLUX or Stable Diffusion for design, marketing, or creative projects
- Custom model training — train LoRAs on your own images to generate consistent characters, styles, or branded visuals
- Private generation — no content moderation, no watermarks, no data sent to external services
- Batch generation — ComfyUI supports queuing multiple prompts for overnight batch runs
- Experimentation — swap models, adjust samplers, modify workflows, and iterate without per-image fees
FLUX vs Stable Diffusion — When to Use Which
| FLUX Schnell | Stable Diffusion (SDXL) | |
|---|---|---|
| Speed | ~10–15 seconds per image | ~20–30 seconds per image |
| Quality at default settings | High detail, strong prompt adherence | High detail, large community fine-tune ecosystem |
| LoRA ecosystem | Growing | Massive — thousands of community LoRAs available |
| Best for | Fast iteration, photorealistic output | Style variety, community models, established workflows |
Both are available on this image. Use whichever fits your project.
Resources
- FLUX.1 — Black Forest Labs
- Stable Diffusion — Stability AI
- Kohya — GitHub
- ComfyUI — GitHub
- Barrack AI Deployment Guide
- Barrack AI API Documentation
Frequently Asked Questions
What is the Barrack ComfyUI SD-FLUX-KOHYA image?
A pre-configured virtual machine image that includes FLUX.1 Schnell (already downloaded), Stable Diffusion support, Kohya LoRA training tools, and ComfyUI as the visual interface. All dependencies are installed and services start automatically on boot.
Which GPUs can run this image?
The image is available on RTX A6000 (48GB), L40 (48GB), A100 PCIe (80GB), H100 PCIe (80GB), and H100 PCIe NVLink (80GB). FLUX Schnell in fp8 runs comfortably on all of them. LoRA training benefits from higher VRAM.
Do I need to install anything after deploying the VM?
No. FLUX Schnell is pre-downloaded, ComfyUI is configured, Kohya is installed, and workflow templates are ready to load. Everything starts automatically on boot.
How fast is image generation?
FLUX Schnell generates a 1024×1024 image in approximately 10–15 seconds at 4 inference steps on an RTX A6000. Stable Diffusion SDXL takes approximately 20–30 seconds per image.
What is LoRA training?
LoRA (Low-Rank Adaptation) is a technique for fine-tuning a base model on a small set of reference images. You provide 20–50 images of a specific style, character, or object, and the training process produces a lightweight file (100–300 MB) that biases the model toward generating content matching your references.
Can I use community LoRAs and checkpoints?
Yes. You can download any compatible checkpoint or LoRA file to the appropriate directory on the VM via SSH (~/ComfyUI/models/checkpoints/ for checkpoints, ~/ComfyUI/models/loras/ for LoRAs) and use them in ComfyUI immediately.
Does my data leave the VM?
No. All generation and training runs locally on your GPU. No images, prompts, or training data are sent to any external service.
What does it cost?
Barrack AI uses per-minute billing with no contracts. You pay only for the time your VM is running. There are no per-image fees and no monthly subscriptions. H100 PCIe starts at $1.99/hr.
How do I deploy via API instead of the dashboard?
Barrack AI provides a full deployment API. See the API documentation for programmatic instance creation, management, and termination.
Last updated: February 23, 2026
Barrack AI provides GPU cloud instances for AI workloads — per-minute billing, no contracts. Learn more →
