3D semantic attributes
(such as segmentation masks, geometric features, keypoints, and materials)
can be encoded as per-point probe functions
on 3D geometries.
β
probe functions: per-point functions defined on the shape surface
Given a collection of related 3D shapes,
this paper considers
"how to jointly analyze such probe functions over different shapes", and
"how to discover common latent structures using a neural network"
even absence of any correspondence information
Instead of a two-stage procedure
to first build independent functional spaces
and then relate them through correspondences(functional or traditional),
π½
This paper proposes a correspondence free framework
that directly learns consistent bases across a shape collection
that reflect the shared structure of the set of probe functions.
Paper produces a compact encoding
for meaningful functions over a collection of related 3D shapes
by learning a small functional basis for each shape
using neural netwroks.
Set of functional bases of each shape,
a.k.a a shape-dependent dictionary,
is computed as a set of functions on a point cloud
representing the underlying geometry
(a functional set whose span will include probe functions on that space.)
Training is accomplished in a very simple manner
by giving the network sequences of pairs
consisting of a shape geometry(as point clouds)
and a semantic probe function on that geometry
(that should be in the associated basis span).
The NN will maximize its representational capacity
by learning consistent bases
that reflect this shared functional structure,
leading in turn to consistent sparse function encodings.
Thus, consistent functional bases emerge from the network without explicit supervision.
Given a collection of shapes {Xi}, each of which has
a sample function of specific semantic meaning {fi}
(e.g. indicator of a subset of semantic parts or keypoints),
π½
we consider the problem of
1) sharing the semantic information across the shapes, and
2) predicting a functional dictionary A(X;Ξ) for each shape
that linearly spans all plausible semantic functions on the shape (Ξ: NN weights)
Assume that shape is given as n points sampled on its surface,
a function f is represented with a vector in β^n (a scalar per point),
and atoms of the dictionary are represented as columns of a matrix A(X;Ξ) β β^(n x k),
(k: sufficently large number for the size of dictionary)
Note that the column space of A(X;Ξ)
can include any function f
if it has all Dirac delta functions of all points as columns.
Paper aims at finding a much lower-dimensional vector space
that also contains all plausible semantic functions.
Also, force the columns of A(X;Ξ)
to encode atomic semantics in applications
such as atomic instances in segmentation,
by adding appropriate constraints.
Paper's network is trained on
point cloud representations of shape geometry
and associated semantic functions on that point cloud.
These functions express a shared semantic understanding of the shapes
but are not coordinated in any way.
Network is able to produce a small dictionary of basis functions for each shape,
a dictionary whose span includes
the semantic functions provided for that shape.
Even though our shapes have
independent discretizations and no functional correspondences are provided,
network can generate latent bases,
in a consistent order,
that reflect the shared semantic structure among the shapes.
NN input: pairs of a shape X
including n points and a function f β β^n
NN output: matrix A(X;Ξ) β β^(n x k)
as a dictionary of functions on the shape.
Proposed Loss function:
loss function is designed for minimizing both
1) projection error from the input function f
to the vector space A(X;Ξ)
2) number of atoms in the dictionary matrix.
where x β β^k : linear combination weight vector,
Ξ³: weight for a regularization
F(A(X;Ξ)): function that measures the projection error,
l2,1-norm : regularizer inducing structured sparsity,
encouraging more columns to be zero vectors.
C(A(X;Ξ), x): set of constraints on both A(X;Ξ) and x
depending on the applications.
For example, when the input function in an indicator(binary function,
we constrain all elements in both A(X;Ξ) and x
to be in [0, 1] range
Note paper's loss minimization is a min-min optimzation problem
π inner minimization (embedded in loss function in Eq.1)
optimizes the reconstruction coefficients
based on the shape dependent dictionary predicted by the network.
π outer minimization,
minimizes our loss function,
updates the NN weights to predict a best shape dependent dictionary.
Nested minimization generally does not have an analytic solution due to the constraint on x.
Thus, impossible to directly compute the gradient of L(A(X;Ξ); f) without x.
π½
solve this by an alternating minimization scheme as describe in Alogrithm.1
In a single gradient descent step,
first minimize F(A(X;Ξ); f) over x with the current A(X;Ξ),
and then compute the gradient of L(A(X;Ξ); f) while
fixing x.
Some constraints for both A(X;Ξ) and x
can be induced from the assumptions of
input function f and the properties of the dictionary atoms.
In the segmentation problem, we take an indicator function of a set of segments as an input,
and we desire that each atom in the output dictionary indicates an atomic part Figure 1 (a).
This, we restrict both A(X;Ξ) and x to have values in the [0 ,1] range.
Atomic parts in the dictionary must partition the shape,
meaning that each point must be assigned to one and only one atom.
π½ Thus, we add sum-to-one constraint for every row of A(X;Ξ)
The set of constraints for the segmentation problem is defined as follows:
where A(X;Ξ)i,j : (i, j)-th element of matrix A(X;Ξ)
0 and 1 : vectors/matrices with an appropriate size
β‘οΈ
The first constraint on x is incorporated in solving the inner minimization problem,
second and third constraints on A(X;Ξ) can simply be implemented
by using softmax activation at the last layer of the network.
Similarly with the segmentation problem,
the input function in the keypoint correspondence problem is also
an indicator finction of a set of points (Fig 1(b))
Thus, use the same [0,1] range constraint for both A(X;Ξ) and x.
Each atom needs to represent a single point,
and thus we add sum-to-one constraint
for every column of A(X;Ξ):
For robustness, a distance function from the keypoints can be used a input
instead of the binary indicator function.
Paper uses a normalized Gaussian-weighed distance function g in experiment:
red line: best one-to-one correspondences between GT and predicted keypoints for each shape
green line: correspondence between GT labels and atom indices for all shapes