Nodes Browser
ComfyDeploy: How comfy_clip_blip_node works in ComfyUI?
What is comfy_clip_blip_node?
CLIPTextEncodeBLIP: This custom node provides a CLIP Encoder that is capable of receiving images as input.
How to install it in ComfyDeploy?
Head over to the machine page
- Click on the "Create a new machine" button
- Select the
Edit
build steps - Add a new step -> Custom Node
- Search for
comfy_clip_blip_node
and select it - Close the build step dialig and then click on the "Save" button to rebuild the machine
A ComfyUI Node for adding BLIP in CLIPTextEncode
Announcement: BLIP is now officially integrated into CLIPTextEncode
Dependencies
- ✅ Fairscale>=0.4.4 (NOT in ComfyUI)
- ✅ Transformers==4.26.1 (already in ComfyUI)
- ✅ Timm>=0.4.12 (already in ComfyUI)
- ✅ Gitpython (already in ComfyUI)
Local Installation
Inside ComfyUI_windows_portable\python_embeded, run:
<pre>python.exe -m pip install fairscale</pre>And, inside ComfyUI_windows_portable\ComfyUI\custom_nodes, run:
<pre>git clone https://github.com/paulo-coronado/comfy_clip_blip_node</pre>Google Colab Installation
Add a cell with the following code:
<pre> !pip install fairscale !cd custom_nodes && git clone https://github.com/paulo-coronado/comfy_clip_blip_node </pre>How to use
- Add the CLIPTextEncodeBLIP node;
- Connect the node with an image and select a value for min_length and max_length;
- Optional: if you want to embed the BLIP text in a prompt, use the keyword BLIP_TEXT (e.g. "a photo of BLIP_TEXT", medium shot, intricate details, highly detailed).
Acknowledgement
The implementation of CLIPTextEncodeBLIP relies on resources from <a href="https://github.com/salesforce/BLIP">BLIP</a>, <a href="https://github.com/salesforce/ALBEF">ALBEF</a>, <a href="https://github.com/huggingface/transformers">Huggingface Transformers</a>, and <a href="https://github.com/rwightman/pytorch-image-models/tree/master/timm">timm</a>. We thank the original authors for their open-sourcing.