Search

Your search keyword '"Li, Hongyang"' showing total 107 results

Search Constraints

Start Over You searched for: Author "Li, Hongyang" Remove constraint Author: "Li, Hongyang" Publication Type Electronic Resources Remove constraint Publication Type: Electronic Resources
107 results on '"Li, Hongyang"'

Search Results

1. TAPTR: Tracking Any Point with Transformers as Detection

2. Generalized Predictive Model for Autonomous Driving

3. SparseFusion: Efficient Sparse Multi-Modal Fusion Framework for Long-Range 3D Perception

4. FastMAC: Stochastic Spectral Sampling of Correspondence Graph

5. Enhancing Generalization in Medical Visual Question Answering Tasks via Gradient-Guided Model Perturbation

6. Embodied Understanding of Driving Scenarios

7. Translating Images to Road Network:A Non-Autoregressive Sequence-to-Sequence Approach

8. Grounded SAM: Assembling Open-World Models for Diverse Visual Tasks

9. Learning Manipulation by Predicting Interaction

10. Vista: A Generalizable Driving World Model with High Fidelity and Versatile Controllability

11. Characterisation of novel mobile genetic elements and their association with antibiotic resistance genes in gram negative bacteria

12. Lite DETR : An Interleaved Multi-Scale Encoder for Efficient DETR

13. Mimic before Reconstruct: Enhancing Masked Autoencoders with Feature Mimicking

14. Introducing Depth into Transformer-based 3D Object Detection

15. Policy Pre-training for Autonomous Driving via Self-supervised Geometric Modeling

16. Scene as Occupancy

17. Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation

18. Think Twice before Driving: Towards Scalable Decoders for End-to-End Autonomous Driving

19. A Strong and Reproducible Object Detector with Only Public Datasets

20. OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping

21. Graph-based Topology Reasoning for Driving Scenes

22. Sparse Dense Fusion for 3D Object Detection

23. Detection Transformer with Stable Matching

24. Geometric-aware Pretraining for Vision-centric 3D Object Detection

25. 3D Data Augmentation for Driving Scenes on Camera

26. Grounded-SAM: Detect, Segment and Generate Anything

27. Visual Point Cloud Forecasting enables Scalable Autonomous Driving

28. Fully Sparse 3D Occupancy Prediction

29. LaneSegNet: Map Learning with Lane Segment Perception for Autonomous Driving

30. DriveLM: Driving with Graph Visual Question Answering

31. A Survey of Reasoning with Foundation Models

32. LLaVA-Grounding: Grounded Visual Chat with Large Multimodal Models

33. Open-sourced Data Ecosystem in Autonomous Driving: the Present and Future

34. Visual In-Context Prompting

35. LLM4Drive: A Survey of Large Language Models for Autonomous Driving

36. Leveraging Vision-Centric Multi-Modal Expertise for 3D Object Detection

37. DriveAdapter: Breaking the Coupling Barrier of Perception and Planning in End-to-End Autonomous Driving

38. DFA3D: 3D Deformable Attention For 2D-to-3D Feature Lifting

39. Density-invariant Features for Distant Point Cloud Registration

40. End-to-end Autonomous Driving: Challenges and Frontiers

41. detrex: Benchmarking Detection Transformers

42. Planning-oriented Autonomous Driving

43. BEVFormer v2: Adapting Modern Image Backbones to Bird's-Eye-View Recognition via Perspective Supervision

44. Stare at What You See: Masked Image Modeling without Reconstruction

45. DCL-Net: Deep Correspondence Learning Network for 6D Pose Estimation

46. Delving into the Devils of Bird's-eye-view Perception: A Review, Evaluation and Recipe

47. ST-P3: End-to-end Vision-based Autonomous Driving via Spatial-Temporal Feature Learning

48. HDGT: Heterogeneous Driving Graph Transformer for Multi-Agent Trajectory Prediction via Scene Encoding

49. BEVFormer: Learning Bird's-Eye-View Representation from Multi-Camera Images via Spatiotemporal Transformers

50. PersFormer: 3D Lane Detection via Perspective Transformer and the OpenLane Benchmark

Catalog

Books, media, physical & digital resources