Skip to content

Binding Touch to Everything: Learning Unified Multimodal Tactile Representations

Notifications You must be signed in to change notification settings

cfeng16/UniTouch

Repository files navigation

Binding Touch to Everything: Learning Unified Multimodal Tactile Representations (CVPR 2024)

arXiv HF Models Static Badge


This repo contains code of Touch-LLM for UniTouch. Our code is built on top of the ImageBind and LLaMA-Adapter codebases.

UniTouch model

Inference with Pretrained Models

  1. Download the pretrained touch encoder (last_new.ckpt) from the HuggingFace model hub and put it in the ./UniTouch folder, same level as touch_qa.py.
  2. Download the folder ckpts from the HuggingFace model hub and put it in the ./UniTouch folder, same level as touch_qa.py.
  3. Download the folder llama_ori from the HuggingFace model hub and put it in the ./UniTouch folder, same level as touch_qa.py.

For Touch-LLM:

CUDA_VISIBLE_DEVICES=0 python touch_qa.py

Citation

@inproceedings{yang2024binding,
  title={Binding touch to everything: Learning unified multimodal tactile representations},
  author={Yang, Fengyu and Feng, Chao and Chen, Ziyang and Park, Hyoungseob and Wang, Daniel and Dou, Yiming and Zeng, Ziyao and Chen, Xien and Gangopadhyay, Rit and Owens, Andrew and others},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={26340--26353},
  year={2024}
}

About

Binding Touch to Everything: Learning Unified Multimodal Tactile Representations

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages