Article 69XV1 Vid2Seq: A pretrained visual language model for describing multi-event videos

Vid2Seq: A pretrained visual language model for describing multi-event videos

by
from Hacker News on (#69XV1)
Comments
External Content
Source RSS or Atom Feed
Feed Location http://news.ycombinator.com/rss
Feed Title Hacker News
Feed Link https://news.ycombinator.com/
Reply 0 comments