ronghanghu
Research Scientist, Meta AI
Repositories
Select a repository to view its commits, contributors, and more.seg_every_thing
Code release for Hu et al., Learning to Segment Every Thing. in CVPR, 2018.
n2nmn
Code release for Hu et al. Learning to Reason: End-to-End Module Networks for Visual Question Answering. in ICCV, 2017
tensorflow_compact_bilinear_pooling
Compact Bilinear Pooling in TensorFlow
speaker_follower
Code release for Fried et al., Speaker-Follower Models for Vision-and-Language Navigation. in NeurIPS, 2018.
natural-language-object-retrieval
Code release for Hu et al. Natural Language Object Retrieval, in CVPR, 2016
lcgn
Code release for Hu et al., Language-Conditioned Graph Networks for Relational Reasoning. in ICCV, 2019
text_objseg
Code release for Hu et al. Segmentation from Natural Language Expressions. in ECCV, 2016
snmn
Code release for Hu et al., Explainable Neural Computation via Stack Neural Module Networks. in ECCV, 2018
cmn
Code release for Hu et al. Modeling Relationships in Referential Expressions with Compositional Modular Networks. in CVPR, 2017
vit_10b_fsdp_example
See details in https://github.com/pytorch/xla/blob/r1.12/torch_xla/distributed/fsdp/README.md
gqa_single_hop_baseline
A simple but well-performing "single-hop" visual attention model for the GQA dataset