Skip to content
/ DeCap Public
forked from dhg-wei/DeCap

ICLR 2023 DeCap: Decoding CLIP Latents for Zero-shot Captioning

Notifications You must be signed in to change notification settings

YqGao716/DeCap

 
 

Repository files navigation

Official implementation for DeCap

DeCap: Decoding CLIP Latents for Zero-Shot Captioning via Text-Only Training

Published at ICLR 2023

Paper link: DeCap

Data

Download coco_train to data. Download cc3m_train to data.

Training

./train_coco.sh

or

./train_cc3m.sh

Inferece

See inference_decap.ipynb.

Citation

@inproceedings{
li2023decap,
title={DeCap: Decoding {CLIP} Latents for Zero-Shot Captioning via Text-Only Training},
author={Wei Li and Linchao Zhu and Longyin Wen and Yi Yang},
booktitle={The Eleventh International Conference on Learning Representations },
year={2023},
url={https://openreview.net/forum?id=Lt8bMlhiwx2}
}

Acknowledgments

This repository is heavily based on ClipCap. For training we used the data of COCO dataset and Conceptual Captions.

About

ICLR 2023 DeCap: Decoding CLIP Latents for Zero-shot Captioning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 54.0%
  • Jupyter Notebook 42.0%
  • Shell 4.0%