-
Adobe Research
- San Jose, CA, US
-
06:34
(UTC -07:00) - http://david-yoon.github.io
- @david_s_yoon
Highlights
- Pro
Stars
Image to prompt with BLIP and CLIP
An open source implementation of CLIP.
Robust Speech Recognition via Large-Scale Weak Supervision
LAVIS - A One-stop Library for Language-Vision Intelligence
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Training and serving large-scale neural networks with auto parallelization.
Must-read papers on prompt-based tuning for pre-trained language models.
Factual consistency checking model for abstractive summaries (NAACL-22 Findings)
QACE is an image captioning metric using QA (EMNLP-21 Findings)
Code and dataset for AAAI 2022 paper "CAISE: Conversational Agent for Image Search and Editing" Hyounghun Kim, Doo Soon Kim, Seunghyun Yoon, Franck Dernoncourt, Trung Bui, and Mohit Bansal
PyTorch code for "Fine-grained Image Captioning with CLIP Reward" (Findings of NAACL 2022)
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
Perturbation CheckLists for Evaluating NLG Evaluation Metrics, EMNLP 2021
PyTorch implementation of a 1.3B text-to-image generation model trained on 14 million image-text pairs
☁️ Build multimodal AI applications with cloud-native stack
Simple Questions Generate Named Entity Recognition Datasets (EMNLP 2022)
A simple semantic search engine for scientific papers.
Explicit Alignment Objectives for Multilingual Bidirectional Encoders
An awesome README template to jumpstart your projects!
A collection of links and notes on forced alignment tools
A list of contrastive Learning papers
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
An unreferenced image captioning metric (ACL-21)