Robot Learning With Implicit Representations
Robot Learning With Implicit Representations
Animesh Garg
RSS 2022 Workshop
What is Implicit Neural Representation?
✗ Memory
✗ Arbitrary Topologies
✗ Connectivity Structures
ü Continuous Representations
ü “Infinite” Spatial Resolution
ü Memory depends on signal complexity
Implicit
Voxels Point Clouds Mesh
Representation
𝑓: ℝ! → ℝ"
𝑉 = 𝑓(𝑟)
Images:
ü𝑟:Continuous
(x, y), 𝑉: (r, Representations
g, b)
ü “Infinite” Spatial Resolution
ü3DMemory
Scenesdepends
and Shapes
on signal
(as incomplexity
NeRFs)
𝑟: (𝑥, 𝑦, 𝑧, 𝜃, 𝜙), 𝑉: (𝑟, 𝑔, 𝑏, 𝜎)
✗ Not Analytically Tractable Implicit
Trajectories Representation
𝑟: (𝑞)"! generalized coordinates
𝑉: utility function
Implicit Representations in Visual Computing
Synergies Between Affordance and Geometry: 6-DOF Grasp Detection via Implicit Representations. In RSS, 2021.
3D Neural Scene Representations for Visuomotor Control. In CoRL, 2021.
Neural Descriptor Fields: SE(3)-Equivariant Object Representations for Manipulation. In ICRA, 2022.
Robot Learning with Implicit Representations
Algorithmic Development (perception and control)
+ Improved Simulation for Contact-rich Manipulation
Work in progress
Robot Learning with Implicit Representations
Algorithmic Development (perception and control)
+ Improved Simulation for Contact-rich Manipulation
9
Motivation 2 NeRF-2-NeRF
Registration Estimated Object
Poses
Collected NeRFs
In Memory
3
Combination
&
Rendering &
…
Interactive &
Interactable
Simulation
1 NeRF
Acquisition
What are Neural Radiance Fields (NeRFs)?
Training an MLP Composition & Rendering
Registration Problem in NeRFs
Unsupervised Training to Find T - Objective Function?
View/RGB Difference Correspondence Difference
Challenges:
Even if learned T is optimal: Error between rendered images We need a robust function applied to
is NOT zero! => MSE
The scenes are only partially overlapping. To make it more robust
Random
view 1
Random
view 2
NeRF 1 without transform NeRF 1 with transform Overlap of fixed NeRF2 Fixed NeRF 2 (target)
with sample points with sample points and moving NeRF1
Different Lightings (Failure Case)
If we use only radiance for registration,
then different lighting models on the object fail!
• Fix: Use Geometry features rather than radiance
Random
view 1
Random
view 2
(moving) NeRF 1 - initial NeRF 1 - registration Overlay of fixed NeRF 2 (fixed) NeRF 2 - target
pose iterations and moving NeRF 1
Robot Learning with Implicit Representations
Algorithmic Development (perception and control)
+ Improved Simulation for Contact-rich Manipulation
21
Existing Grasping Methods
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
+ Table-top object grasping
+ Grasping in clutter
+ Bin-picking
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
+ Table-top object grasping
+ Grasping in clutter
+ Bin-picking
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Existing Grasping Methods
+ Table-top object grasping
+ Grasping in clutter
+ Bin-picking
6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In ICCV, 2019.
Neural Motion Fields
Goal:
Learn a value function that can be used to plan a trajectory for grasping
P ). We propose Neural Motion Fields, which consists of cross-entropy loss function, which is defined as
More data helps with fine-grained rotation error with non-stationary objects
Ablation Study on Number of Anchor Grasps
Anchor grasps
37
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Only a fraction of possible contacts are active (in collision) at a given time.
Inactive contacts have no gradient.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Only a fraction of possible contacts are active (in collision) at a given time.
Inactive contacts have no gradient.
Can’t follow gradient to create new contacts.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Can’t follow gradient to create new contacts.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Can’t follow gradient to create new contacts.
2. Local flatness
Often compute ground-truth SDF from mesh.
If closest point is on triangle face, surface normal gradient is 0.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Can’t follow gradient to create new contacts.
2. Local flatness
Often compute ground-truth SDF from mesh.
If closest point is on triangle face, surface normal gradient is 0.
Can’t follow gradient to improve contact normals.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Can’t follow gradient to create new contacts.
2. Local flatness
Can’t follow gradient to improve contact normals.
Motivation
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges
1. Contact sparsity
Can’t follow gradient to create new contacts.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
2. Local flatness
Can’t follow gradient to improve contact normals.
Proper gradient
Figure from Werling, K., Omens, D., Lee, J., Exarchos, I., & Liu, C. K. Fast and Feature-Complete
Differentiable Physics for Articulated Rigid Bodies with Contact.
Challenge #2: Local flatness
SDF ground truth is often computed from a mesh.
But surface normal is constant on faces,
so contact normal (computed as positional derivative of SDF) has 0 gradient.
Figure from Phong Tessellation T Boubekeur, M Alexa ACM Transactions on Graphics 27 (5)
Challenge #3: Non-smooth geometry
Easy to optimize over surface of a spherical cow (⚽),
but most aren’t so smooth (🐄).
Challenge #3: Non-smooth geometry
Easy to optimize over surface of a spherical cow (⚽),
but most aren’t so smooth (🐄).
[*] Hasson, Y., Varol, G., Tzionas, D., Kalevatykh, I., Black, M. J., Laptev, I., & Schmid, C. (2019). Learning joint reconstruction of hands
and manipulated objects. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11807-11816).
Grasps from the ObMan dataset [*]
[*] Hasson, Y., Varol, G., Tzionas, D., Kalevatykh, I., Black, M. J., Laptev, I., & Schmid, C. (2019). Learning joint reconstruction of hands
and manipulated objects. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11807-11816).
Grasps from the ObMan dataset [*]
less contact
[*] Hasson, Y., Varol, G., Tzionas, D., Kalevatykh, I., Black, M. J., Laptev, I., & Schmid, C. (2019). Learning joint reconstruction of hands
and manipulated objects. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11807-11816).
Grasps from the ObMan dataset [*]
less contact
less stable
Less contact = less friction.
[*] Hasson, Y., Varol, G., Tzionas, D., Kalevatykh, I., Black, M. J., Laptev, I., & Schmid, C. (2019). Learning joint reconstruction of hands
and manipulated objects. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11807-11816).
Grasps from the ObMan dataset [*]
less contact
less stable less plausible
Less contact = less friction. Human grasping is contact-rich.
[*] Hasson, Y., Varol, G., Tzionas, D., Kalevatykh, I., Black, M. J., Laptev, I., & Schmid, C. (2019). Learning joint reconstruction of hands
and manipulated objects. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 11807-11816).
ObMan
Ours
~30x
ObMan
Ours
~30x
ObMan
Ours
Ours
Ours
~30x
~30x
Grasp’D: Take away
Goal: Make SDF-based contact forces friendly to gradient-based optimization.
Why? Planning in high-dimensional contact-rich scenarios,
e.g., robotic grasping and manipulation with multi-finger hands.
Challenges & Proposed Solutions
1. Contact sparsity à Leaky gradient
2. Local flatness à Phong SDF
3. Non-smooth object geometry à SDF Dilation
An example application: Generating contact-rich human & robotic grasps.
Robot Learning with Implicit Representations
Algorithmic Development (perception and control)
+ Improved Simulation for Contact-rich Manipulation
Animesh Garg
[email protected] | @animesh_garg