glTF   "  �!  JSON{"asset": {"version": "2.0", "extras": {"prompt": "{\"3\": {\"inputs\": {\"seed\": 960690829976426, \"steps\": 16, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"70\", 0], \"positive\": [\"80\", 0], \"negative\": [\"80\", 1], \"latent_image\": [\"66\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"51\": {\"inputs\": {\"crop\": \"none\", \"clip_vision\": [\"54\", 1], \"image\": [\"56\", 0]}, \"class_type\": \"CLIPVisionEncode\", \"_meta\": {\"title\": \"CLIP Vision Encode\"}}, \"54\": {\"inputs\": {\"ckpt_name\": \"model.fp16.safetensors\"}, \"class_type\": \"ImageOnlyCheckpointLoader\", \"_meta\": {\"title\": \"Image Only Checkpoint Loader (img2vid model)\"}}, \"56\": {\"inputs\": {\"image\": \"00000000000000000000000Aexample.png\"}, \"class_type\": \"LoadImage\", \"_meta\": {\"title\": \"Load Image\"}, \"is_changed\": [\"a8e215ad32a0052fc4190e9c5863428d5ee35cd98b246243842ba8360511b7c4\"]}, \"61\": {\"inputs\": {\"num_chunks\": 2000, \"octree_resolution\": 128, \"samples\": [\"3\", 0], \"vae\": [\"54\", 2]}, \"class_type\": \"VAEDecodeHunyuan3D\", \"_meta\": {\"title\": \"VAEDecodeHunyuan3D\"}}, \"62\": {\"inputs\": {\"threshold\": 0.6000000000000001, \"voxel\": [\"61\", 0]}, \"class_type\": \"VoxelToMeshBasic\", \"_meta\": {\"title\": \"VoxelToMeshBasic\"}}, \"66\": {\"inputs\": {\"resolution\": 3072, \"batch_size\": 1}, \"class_type\": \"EmptyLatentHunyuan3Dv2\", \"_meta\": {\"title\": \"EmptyLatentHunyuan3Dv2\"}}, \"67\": {\"inputs\": {\"filename_prefix\": \"mesh/ComfyUI\", \"image\": \"\", \"mesh\": [\"62\", 0]}, \"class_type\": \"SaveGLB\", \"_meta\": {\"title\": \"SaveGLB\"}}, \"70\": {\"inputs\": {\"shift\": 1.0000000000000002, \"model\": [\"54\", 0]}, \"class_type\": \"ModelSamplingAuraFlow\", \"_meta\": {\"title\": \"ModelSamplingAuraFlow\"}}, \"80\": {\"inputs\": {\"clip_vision_output\": [\"51\", 0]}, \"class_type\": \"Hunyuan3Dv2Conditioning\", \"_meta\": {\"title\": \"Hunyuan3Dv2Conditioning\"}}}", "workflow": "{\"id\": \"81ccfe3e-3540-49e3-88ae-4fde173b8c87\", \"revision\": 0, \"last_node_id\": 80, \"last_link_id\": 166, \"nodes\": [{\"id\": 3, \"type\": \"KSampler\", \"pos\": [350, 310], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 156}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 165}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 166}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 143}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [130]}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [960690829976426, \"randomize\", 16, 1, \"euler\", \"normal\", 1]}, {\"id\": 51, \"type\": \"CLIPVisionEncode\", \"pos\": [28.491897583007812, 203.88284301757812], \"size\": [260, 80], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip_vision\", \"name\": \"clip_vision\", \"type\": \"CLIP_VISION\", \"link\": 111}, {\"label\": \"image\", \"name\": \"image\", \"type\": \"IMAGE\", \"link\": 145}], \"outputs\": [{\"label\": \"CLIP_VISION_OUTPUT\", \"name\": \"CLIP_VISION_OUTPUT\", \"type\": \"CLIP_VISION_OUTPUT\", \"slot_index\": 0, \"links\": [164]}], \"properties\": {\"Node name for S&R\": \"CLIPVisionEncode\"}, \"widgets_values\": [\"none\"]}, {\"id\": 54, \"type\": \"ImageOnlyCheckpointLoader\", \"pos\": [-391.26416015625, 66.70899200439453], \"size\": [369.6000061035156, 98], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [155]}, {\"label\": \"CLIP_VISION\", \"name\": \"CLIP_VISION\", \"type\": \"CLIP_VISION\", \"slot_index\": 1, \"links\": [111]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [158]}], \"properties\": {\"Node name for S&R\": \"ImageOnlyCheckpointLoader\"}, \"widgets_values\": [\"model.fp16.safetensors\"]}, {\"id\": 56, \"type\": \"LoadImage\", \"pos\": [-390, 210], \"size\": [370, 340], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [145]}, {\"label\": \"MASK\", \"name\": \"MASK\", \"type\": \"MASK\", \"slot_index\": 1, \"links\": []}], \"properties\": {\"Node name for S&R\": \"LoadImage\"}, \"widgets_values\": [\"00000000000000000000000Aexample.png\", \"image\", \"\"]}, {\"id\": 61, \"type\": \"VAEDecodeHunyuan3D\", \"pos\": [700, 50], \"size\": [315, 102], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 130}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 158}], \"outputs\": [{\"label\": \"VOXEL\", \"name\": \"VOXEL\", \"type\": \"VOXEL\", \"slot_index\": 0, \"links\": [132]}], \"properties\": {\"Node name for S&R\": \"VAEDecodeHunyuan3D\"}, \"widgets_values\": [2000, 128]}, {\"id\": 62, \"type\": \"VoxelToMeshBasic\", \"pos\": [700, 200], \"size\": [210, 58], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"voxel\", \"name\": \"voxel\", \"type\": \"VOXEL\", \"link\": 132}], \"outputs\": [{\"label\": \"MESH\", \"name\": \"MESH\", \"type\": \"MESH\", \"slot_index\": 0, \"links\": [146]}], \"properties\": {\"Node name for S&R\": \"VoxelToMeshBasic\"}, \"widgets_values\": [0.6000000000000001]}, {\"id\": 66, \"type\": \"EmptyLatentHunyuan3Dv2\", \"pos\": [30, 330], \"size\": [260.2345886230469, 82], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [143]}], \"properties\": {\"Node name for S&R\": \"EmptyLatentHunyuan3Dv2\"}, \"widgets_values\": [3072, 1]}, {\"id\": 67, \"type\": \"SaveGLB\", \"pos\": [700, 310], \"size\": [532.7781982421875, 481.68194580078125], \"flags\": {}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"label\": \"mesh\", \"name\": \"mesh\", \"type\": \"MESH\", \"link\": 146}], \"outputs\": [], \"properties\": {\"Node name for S&R\": \"SaveGLB\", \"Camera Info\": {\"position\": {\"x\": -5.450869037402971, \"y\": 2.221460971446785, \"z\": 5.041442306321533}, \"target\": {\"x\": 0, \"y\": 1.011904747431601, \"z\": 0}, \"zoom\": 1, \"cameraType\": \"perspective\"}, \"Background Color\": \"#141414\"}, \"widgets_values\": [\"mesh/ComfyUI\", \"\"]}, {\"id\": 70, \"type\": \"ModelSamplingAuraFlow\", \"pos\": [30, 70], \"size\": [260, 60], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 155}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [156]}], \"properties\": {\"Node name for S&R\": \"ModelSamplingAuraFlow\"}, \"widgets_values\": [1.0000000000000002]}, {\"id\": 77, \"type\": \"MarkdownNote\", \"pos\": [-380, -90], \"size\": [348.69091796875, 109.14118194580078], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"Download [hunyuan3d-dit-v2/model.fp16.safetensors](https://huggingface.co/tencent/Hunyuan3D-2/blob/main/hunyuan3d-dit-v2-0/model.fp16.safetensors) and rename to **hunyuan3d-dit-v2.safetensorss**\\n\\nPut it in the **ComfyUI/models/checkpoints** directory\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 80, \"type\": \"Hunyuan3Dv2Conditioning\", \"pos\": [350, 210], \"size\": [310, 50], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"clip_vision_output\", \"type\": \"CLIP_VISION_OUTPUT\", \"link\": 164}], \"outputs\": [{\"name\": \"positive\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [165]}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"slot_index\": 1, \"links\": [166]}], \"properties\": {\"Node name for S&R\": \"Hunyuan3Dv2Conditioning\"}, \"widgets_values\": []}], \"links\": [[111, 54, 1, 51, 0, \"CLIP_VISION\"], [130, 3, 0, 61, 0, \"LATENT\"], [132, 61, 0, 62, 0, \"VOXEL\"], [143, 66, 0, 3, 3, \"LATENT\"], [145, 56, 0, 51, 1, \"IMAGE\"], [146, 62, 0, 67, 0, \"MESH\"], [155, 54, 0, 70, 0, \"MODEL\"], [156, 70, 0, 3, 0, \"MODEL\"], [158, 54, 2, 61, 1, \"VAE\"], [164, 51, 0, 80, 0, \"CLIP_VISION_OUTPUT\"], [165, 80, 0, 3, 1, \"CONDITIONING\"], [166, 80, 1, 3, 2, \"CONDITIONING\"]], \"groups\": [], \"config\": {}, \"extra\": {}, \"version\": 0.4}"}}, "scenes": [{"nodes": []}], "nodes": [], "scene": 0}