oguzhanercan 's Collections Robotics
updated
Astra: Toward General-Purpose Mobile Robots via Hierarchical Multimodal
Learning
Paper
• 2506.06205
• Published
• 30
BitVLA: 1-bit Vision-Language-Action Models for Robotics Manipulation
Paper
• 2506.07530
• Published
• 20
Ark: An Open-source Python-based Framework for Robot Learning
Paper
• 2506.21628
• Published
• 16
RoboBrain 2.0 Technical Report
Paper
• 2507.02029
• Published
• 35
ThinkAct: Vision-Language-Action Reasoning via Reinforced Visual Latent
Planning
Paper
• 2507.16815
• Published
• 42
PRIX: Learning to Plan from Raw Pixels for End-to-End Autonomous Driving
Paper
• 2507.17596
• Published
• 7
MolmoAct: Action Reasoning Models that can Reason in Space
Paper
• 2508.07917
• Published
• 44
Spatial Forcing: Implicit Spatial Representation Alignment for
Vision-language-action Model
Paper
• 2510.12276
• Published
• 147
RLinf-VLA: A Unified and Efficient Framework for VLA+RL Training
Paper
• 2510.06710
• Published
• 42
Don't Blind Your VLA: Aligning Visual Representations for OOD
Generalization
Paper
• 2510.25616
• Published
• 105
Dream-VL & Dream-VLA: Open Vision-Language and Vision-Language-Action Models with Diffusion Language Model Backbone
Paper
• 2512.22615
• Published
• 49
Rethinking Video Generation Model for the Embodied World
Paper
• 2601.15282
• Published
• 43
GigaBrain-0.5M*: a VLA That Learns From World Model-Based Reinforcement Learning
Paper
• 2602.12099
• Published
• 56