SlideShare a Scribd company logo
1 of 36
Subspace Indexing on Grassmannian Manifold for Large
                    Scale Visual Identification


                                   Zhu Li
                           Media Networking Lab
                     FutureWei (Huawei) Technology, USA
                              Bridgewater, NJ




IBM Research, 2012                1                       Z. Li
Outline

• Short Self-Intro

• Large Scale Visual Analytics
    –   Applications
    –   Key Technical Challenges
    –   Query-Driven Local Subspaces
    –   Indexed Subspaces on Grassmannian Manifold
    –   Simulation
    –   Conclusion & Future Work




IBM Research, 2012                2                  Z. Li
About Me: http://users.eecs.northwestern.edu/~zli

    • Bio:
         – Sr. Staff Researcher, Core Networks R&D, Huawei Tech USA, 2010.10~
           to date
         – Asst Prof, HK Polytechnic Univ, CTO, Mudi Tech, 2008.04~2010.09
         – Senior, Senior Staff, and then Principal Staff Researcher, Multimedia
           Research Lab, Motorola Labs, USA, 2000-08.
    • Research Interests:
         – Biometrics, large scale fingerprint identification, face identification,
           behaviour biometrics
         – Visual search, Compact Visual Descriptor for Mobile Search. Check out
           Mudi's QbC visual search solution for dangdang.com:
            http://i.joyton.com/download/DangDang2.0.apk
         – Large scale audio/visual data analysis, storage and indexing, search and
           mining, Hadoop based map-reduce scheme for large scale visual
           analytics.




IBM Research, 2012                         3                                   Z. Li
The Large Scale Visual Analytics Problems

• Face Recognition
    – Identify face from 7 million HK ID face data set

• Image Search
    – Find out the category of given images




IBM Research, 2012                4                      Z. Li
The Problem

• Identification
    – Given a set of training image data and label{fk, lk}, and a probe p,
      identify the unique label associated with p.

• Why is it difficult ?
    – When the number of unique labels, m, and training data n are large....



                          X = f(I)




IBM Research, 2012                   5                                Z. Li
Appearance Modeling

• Find a “good” f()
    – Such that after projecting the appearance onto the subspace, the
      data points belong to different classes are easily separable




                              X = f(I)




IBM Research, 2012                6                             Z. Li
Global Linear LPP Models: f(X) = AX
• LPP (Xiaofei He, et.al):
   - Minimizing weighted distance (a graph) after projection
                         X
                 min           wj;k jjAxj ¡ Axk jj2
                     A
                         j;k



   -Solve by:

                                                                   X
                         T              T
                XLX A = ¸XDX A; s:t:L = D ¡ W; Dk;k =                  wj;k
                                                                   j



   - Embed a graph with pruned edges
                (
                 wj;k = e¡®jjxj ¡xk jj ;     if jjxj ¡ xk jj · ²
                 0;                         else


IBM Research, 2012                            7                               Z. Li
Global Linear LDA Models: f(X)=AX
• LDA:
    - Maximizing inter-class scatter over intra
           A = arg max | AT S B A |, s.t. | AT SW A |= 1
                       A

                 n                                            n
           s B = ∑ nk ( X k − X )( X k − X )   T        sW = ∑       ∑ ( X j − X k )( X j − X k )T
                k =1                                         k =1 P ( X j )=k




     -Solve by:

                S B A = λSW A


     - Embedding a graph with no edges among inter-class points
                           (
                                       1
                            wj;k =     mi ;         if xj ; xk 2 class i
                            0;                     else


IBM Research, 2012                                       8                                           Z. Li
Graph Embedding Interpretation

• Find the best embedding
    – LDA:
          » preserve the affinity matrix that has zero affinity for data points pairs
            that are not belonging to the same class
    – LPP:
          » Have more flexibility in modeling affinity w jk.




                      LPP Affinity                         LDA Affinity

IBM Research, 2012                        9                                 Z. Li
Non-Linear Models
• Appearance manifolds are non-linear in nature
    – Global linear models will suffer

• Non-Linear Solutions:
    – Kernel method: e.g K-PCA, K-LDA, K-LPP, SVM
        » Evaluate inner product <xj,xk> with a kernel function k(xj, xk), which if satisfy the conditions
          in Mercer’s Theorem, implicitly maps data via a non-linear function.
        » Typically involves a QP problem with a Hessian of size n x n, when n is large, not solvable.
    – LLE /Graph Laplacian:
        » An algorithm that maps input data {xk} to {yk} that tries to preserve an embedded graph
          structure among data points.
        » The mapping is data dependent and has difficulty handling new data outside the training set,
          e.g., a new query point

• How to compromise ?
    – Piece-wise Linear Approximation




IBM Research, 2012                              10                                            Z. Li
Piece-wise Linear : Query Driven
• Query-Driven Piece-wise Linear Model
    – No pre-determined structure on the training data
    – Local neighborhood data patch identified from query point q,
    – Local model built with local data, A(X, q)

                                    X                           Local Graph
                                                             Embedding Projection
                                                                   A(X,q)

q
                                                                         Y=A(X,q)X
                      +                     Local data:
                                              N(X, q)




                                                               +




IBM Research, 2012                        11                                   Z. Li
Local Model Discriminating Power Criteria
• What is a good N(X, q) ?

• Model power of a linear model:
    – A: Dxd, D=wxh

• Data Complexity: Graph Embedding Interpretation:
    –PCA: a fully connected graph
    –LDA: a graph with edges pruned for intra-class points
    –LPP/LEA; k-nn/ ε − nn pruned graph
    –as number of edges/relationship among data points
                             ( n ),
                                2                     PCA      
                             m            m
                                                               
                                                              
                           ∑            ∑
                                     nj
                | E ( X ) |=  ( 2 ), s.t. n j = n,    LDA     
                              j =1       j =1                 
                             nK ,                    LPP / LEA
                                                              


• What is a good compromise of data complexity and model power ?
IBM Research, 2012                            12                   Z. Li
Discriminant Power Co-efficient (DPC)
• Given the model power constraint:
   – w, h, appearance model luminance field size
   – d, dimensionality of A(x, q)

• How to identify a neighborhood to achieve a good balance of data
  complexity and model power ?

   - DPC, K(A(X,q)) =


              w× h× d
             | E ( X (q ) ) |

   - Need to balance DPC with
     info loss in node/edge pruning




IBM Research, 2012                         13                    Z. Li
Head Pose Recognition Performance
• Recognition rate is improved:
    – W=18, h=18, K=30




• And the cost in computation is rather modest
    – Matlab code, online local model A(X,q) learning and NN classification:




IBM Research, 2012                 14                               Z. Li
Face Recognition Performance
• Local model combination in face recognition
    – Query point drives 3 local models, A1(X, q), A2(X, q), A3(X, q)
    – Local model classification error estimation,
    – Combining the results – weighted voting




    Multiple face models with different area and scale:
              (a) Upper face model (18 × 16).
              (b) Lower face model (14 × 18).
                (c) Full face model (21 × 28).


                                                          ORL data set test: leave 1,2,3 out:


IBM Research, 2012                           15                                     Z. Li
Query Driven Solution Problems

• Optimality of the Local Model is not established
    – Parameters ² ¡ N N , k-NN, and heat kernel size determines the
      number of non-zero affinity edges in local graph
    – The choice is based on DPC, which is still heuristic

• Computational Complexity
    – Need to compute a nearest neighbor set and its affinity, as well as
      the local embedding model at run time.
    – Need extra storage to store all training data, because the local NN
      data patch is generated at run time, as function of the query point.
    – Indexing/Hashing scheme to support efficient access of training
      data.




IBM Research, 2012                 16                               Z. Li
Stiefel and Grassmannian Manifolds

• Stiefel manifolds
    – All possible p-dimensional subspaces in d-dimensional space, A pxd,
      spans Stiefel Manifold, S(p, d) in Rdxp, d > p.
                            ©          d£p            0
                                                            ª
                S(p; d) = A 2 R             ; s:t:A A = Id
    – The DoF is not pxd, rather: pd – (1/2)d(d+1)


• Grassmannian manifolds
    – G(p, d) identifies p-dimensional subspaces in d-dimensional space
    – It is stiefel manifolds but with an equivalence constraint:
         » A1 = A2, if span(A1) = span(A2), or
         » Exist othonormal dxd matrix Rd, A1=A2Rd.
    – The DoF: pd-d2. G(p, d) is the quotient space of S(p, d)/O(d)


IBM Research, 2012                    17                              Z. Li
Subspaces on Grassmannian Manifold

• The BEST subspace for identification ?
    – All possible p-dimensional subspaces in d-dimensional space, A pxd,
      spans Grassmannian Manifold, G(p, d) in Rdxp, d > p.
          » eg., G(2, 3), biz card example
    – The DoF of A is not pxd, as for,

            < aj ; ak >= 0; < aj ; aj >= 1; for AT = [a1 ; a2 ; :::; ap ];

    – Face Appearance model, typically, d=400~500, p=10~30.
    – The BEST subspace A* is somewhere on G(p, d), therefore it is
      important to figure out a way to characterize the similarity between
      subspaces in G(p, d), and give a structure of all subspace w.r.t the
      task of identification.




IBM Research, 2012                          18                               Z. Li
Grassmannian Manifold Visualization

• Consider a typical appearance modeling
    – Image size 12x10 pel, appearance space dimension d=120, model
      dimension p=8.
    – 3D visualization of all S(8, 120) and their covariance eigenvalues”
    – Grassmann Manifolds are quotient space S(8, 120)/O(8)




IBM Research, 2012                 19                                Z. Li
Principle Angles

• The principle angles between two subspaces:
    – For Y1, and Y2 in G(p, d), their principle angles are defined as

          cos(µk ) =              max                 u0 vk
                                                       k
                                                              span(A1 )   span(A2 )
                        uk 2span(A1 );vk 2span(A2 )

                          (
                           u0 uk = 1; vk vk = 1
                            k
                                        0
                     s:t:
                           u0 ui = 0; vk vi = 0
                            k
                                       0




    – Where, {uk} and {vk} are called principle dimensions for span(A1) and
      span(A2).


IBM Research, 2012                                    20                              Z. Li
Principle Angles Computing

• The principle angles between two subspaces:
    – For A1, and A2 in G(p, d), their principle dimensions and angles are
      computed by SVD:


                           [U; S; V ] = SV D(AT A2 )
                                              1




    – Where, U=[u1, u2, …, up], and V=[v1, v2, …, vp] are the principle angles.
    – The diagonal of S, [s1, s2,..., sp] are the cosine of principle angles,


                             sk = cos(µk )

IBM Research, 2012                    21                                  Z. Li
Subspace Distance on Grassmannian Manifold

• Subspace distances:
    – Projection Distance
       Def:                               Xp
                       dprj (A1 ; A2 ) = (   sin2 µi )1=2
                                           i=1
        Computing:                             p
                                               X
                       d2 (A1 ; A2 ) = p ¡
                        prj                          cos2 µi = m ¡ jjA0 A2 jj 2
                                                                      1       F
                                               i=1

    – Binet-Cauchy Distance
      Def:                                     Y
                      dbc (A1 ; A2 ) = (1 ¡        cos2 µi )1=2
                                               i
        Computing:
                                           Y
                     d2 (A1 ; A2 )
                      bc             =1¡         cos2 µi = 1 ¡ det2 (A0 A2 )
                                                                      1
                                           i




IBM Research, 2012                    22                                          Z. Li
Subspace Distance on Grassmannian Manifold

• Subspace distances
    – Arc Distance
      Def:                darc (A1 ; A2 ) = (
                                             X
                                               µi )1=2
                                                2

                                             i

        Also known as geodesic distance. It traverse the Grassmannian
        surface, and two subspace collapse into one, when all principle angles
        becomes zero.




IBM Research, 2012                     23                              Z. Li
Weighted Merging of two subspaces

• What if we need merge two subspaces ?
    – Motivation:
          » say if subspace A1 is best for data set S1, and subspace A2 is best for
            data set S2, can we find a subspace A3 that is good for both ?
    – When two subspaces are sufficiently close on Grassmannian
      manifold, we can approximate this by, A3=[t1, t2, ….]

                               n1           n2
                     tk =           uk +         vk                  +
                            n1 + n2      n1 + n2

        Where n1,2 are the size of data set S1,2
    – The new sets of basis may not be orthogonal. Can be corrected by
      Gram-Schmidt orthogonalization.




IBM Research, 2012                             24                             Z. Li
Judicious Local Models

• Data Space Partition
    – Partition the training data set by kd-tree
    – For the kd-tree height of h, we have 2h local data patch as leaf node
    – For each leaf node data patch k, build a local LDA/LPP/PCA model Ak:




IBM Research, 2012                 25                              Z. Li
Subspace Index

• Organizing the Subspace Models
    – For data index of height of h, we have 2h local models Ak: k=1..2h.
    – For a given probe data point, find its leaf node and associated local
      model, do identification. Is this good ?
    – No, because
          » Could be over-fitting, not sure what is the right size local data patch.
          » Improper neighborhood, probe data points falling on the boundary of leaf
            node:
    – Build local models at each subtree ?
          » No, the data partition does not reflect the smooth change of the local
            models.




IBM Research, 2012                      26                                  Z. Li
Model Hierarchical Tree (MHT)

• Indexing Subspaces on Grassmannian manifold
    – It is a VQ like process.
    – Start with a data partition kd-tree, their leaf nodes and associated
      subspaces {Ak}, k=1..2h
    – Repeat
          » Find Ai and Aj, if darc(Ai, Aj) is the smallest among all, and the associated
            data patch are adjacent in the data space.
          » Delete Ai and Aj, replace with merged new subspace, and update
            associated data patch leaf nodes set.
          » Compute the empirical identification accuracy for the merged subspace
          » Add parent pointer to the merged new subspace for A i and Aj .
          » Stop if only 1 subspace left.
    – Benefit:
          » avoid forced merging of subspace models at data patches that are very
            different, though adjacent.

IBM Research, 2012                          27                                  Z. Li
MHT Based Identification

• MHT operation
    – Organize the leaf nodes models into a new hierarchy, with new models
      and associated accuracy (error rate) estimation
    – When a probe point comes, first identify its leaf nodes from the data
      partition tree.
    – Then traverse the MHT from leaf nodes up, until it hits the root,
      which is the global model, and choose the best model along the path
      for identification




IBM Research, 2012                28                               Z. Li
Simulation

• The data set
    – MSRA Multimedia data set, 65k images with class and relevance
      labels:




IBM Research, 2012               29                             Z. Li
Simulation

• Data selection and features
    – Selected 12 classes with 11k images and use the original combined
      889d features from color, shape and texture
    – Performance compared with PCA, LDA and LPP modeling




IBM Research, 2012                30                               Z. Li
Simulation

• Face data set
    – Mixed data set of 242 individuals, and 4840 face images
    – Performance compared with PCA, LDA and LPP modeling




IBM Research, 2012                31                            Z. Li
Summary

• Contributions
    – The work is a piece-wise linear approximation of non-linear
      appearance manifold
    – Query driven provide suboptimal performance but still better than a
      global model.
    – It offers best local models for identification by deriving the
      subspace structure/index with metrics on Grassmannian manifold
    – Guaranteed performance gains, and the root model degenerates into
      the global linear model

• Limitations
    – Do not have a continuous characterization of Identification error
      function on the Grassmann manifold.
    – Still heavy on storage cost
    – Need to get more large scale data set to test it.

IBM Research, 2012                32                               Z. Li
Summary

• Future work
    – Grassmann Hashing – Penalize projection selection with Grassmannian
      metric, offers performance gains over LSH and spectral hashing.
    – Gradient and Newtonian optimization on Grassmannian manifold.

   Related papers
    – X. Wang, Z. Li, and D. Tao, "Subspace Indexing on Grassmann Manifold for Image Search ",
      IEEE Trans. on Image Processing, vol. 20(9), 2011.
    – X. Wang, Z. Li, L. Zhang, and J. Yuan, "Grassmann Hashing for Approx Nearest Neighbour Search in
      High Dimensional Space", Proc. of IEEE Int'l Conf on Multimedia & Expo (ICME), Barcelona, Spain,
      2011.
    – H. Xu, J. Wang, Z. Li, G. Zeng, S. Li, “Complementary Hashing for Approximate Nearest Neighbor
      Search”, IEEE Int'l Conference on Computer Vision (ICCV), Barcelona, Spain, 2011.
    – Yun Fu, Z. Li, J. Yuan, Ying Wu, and Thomas S. Huang, "Locality vs. Globality: Query-Driven Localized
      Linear Models for Facial Image Computing," IEEE Transactions on Circuits and Systems for Video
      Technology (T-CSVT), vol. 18(12), pp. 1741-1752, December, 2008.




IBM Research, 2012                              33                                            Z. Li
Acknowledgement

     • Grants:
          – The work is partially supported by;
                » a Hong Kong RGC Grant, and
                » Microsoft Research Asia faculty grant.

     • Collaborators:
                » Xinchao Wang, valedictorian of Dept of COMP, HK Polytechnic
                  University, class 2010, now PhD at EPFL

                » Dacheng Tao, Professor at Univ of Technology of Sydney.




IBM Research, 2012                      34                                  Z. Li
Q&A

• Questions please......




IBM Research, 2012         35    Z. Li
Thanks !




IBM Research, 2012   36         Z. Li

More Related Content

What's hot

TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++
TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++
TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++sleepy_yoshi
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsTerry Taewoong Um
 
(DL hacks輪読) Deep Kernel Learning
(DL hacks輪読) Deep Kernel Learning(DL hacks輪読) Deep Kernel Learning
(DL hacks輪読) Deep Kernel LearningMasahiro Suzuki
 
Stochastic Variational Inference
Stochastic Variational InferenceStochastic Variational Inference
Stochastic Variational InferenceKaede Hayashi
 
クラシックな機械学習の入門 6. 最適化と学習アルゴリズム
クラシックな機械学習の入門  6. 最適化と学習アルゴリズムクラシックな機械学習の入門  6. 最適化と学習アルゴリズム
クラシックな機械学習の入門 6. 最適化と学習アルゴリズムHiroshi Nakagawa
 
Brief intro : Invariance and Equivariance
Brief intro : Invariance and EquivarianceBrief intro : Invariance and Equivariance
Brief intro : Invariance and Equivariance홍배 김
 
Flow based generative models
Flow based generative modelsFlow based generative models
Flow based generative models수철 박
 
Bidang datar pada ruang (GAR)_Rahmah Fitri
Bidang datar pada ruang (GAR)_Rahmah FitriBidang datar pada ruang (GAR)_Rahmah Fitri
Bidang datar pada ruang (GAR)_Rahmah FitriRahmahFitri4
 
Prml3.5 エビデンス近似〜
Prml3.5 エビデンス近似〜Prml3.5 エビデンス近似〜
Prml3.5 エビデンス近似〜Yuki Matsubara
 
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems. (AAAI 2022)”
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems.  (AAAI 2022)” [DL輪読会]“Meta-Learning for Online Update of Recommender Systems.  (AAAI 2022)”
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems. (AAAI 2022)” Deep Learning JP
 
סיכום הקורס במבנים אלגבריים
סיכום הקורס במבנים אלגברייםסיכום הקורס במבנים אלגבריים
סיכום הקורס במבנים אלגברייםcsnotes
 
Soal dan pembahasan operasi biner dan teori grup dasar - mathcyber1997
Soal dan pembahasan   operasi biner dan teori grup dasar - mathcyber1997Soal dan pembahasan   operasi biner dan teori grup dasar - mathcyber1997
Soal dan pembahasan operasi biner dan teori grup dasar - mathcyber1997HabibisSaleh1
 
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1東京都市大学 データ解析入門 4 スパース性と圧縮センシング1
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1hirokazutanaka
 
The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)James McMurray
 
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous DrivingDeep Learning JP
 
Gaussian Process Regression
Gaussian Process Regression  Gaussian Process Regression
Gaussian Process Regression SEMINARGROOT
 

What's hot (20)

TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++
TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++
TokyoNLP#7 きれいなジャイアンのカカカカ☆カーネル法入門-C++
 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot Mechanics
 
(DL hacks輪読) Deep Kernel Learning
(DL hacks輪読) Deep Kernel Learning(DL hacks輪読) Deep Kernel Learning
(DL hacks輪読) Deep Kernel Learning
 
Stochastic Variational Inference
Stochastic Variational InferenceStochastic Variational Inference
Stochastic Variational Inference
 
クラシックな機械学習の入門 6. 最適化と学習アルゴリズム
クラシックな機械学習の入門  6. 最適化と学習アルゴリズムクラシックな機械学習の入門  6. 最適化と学習アルゴリズム
クラシックな機械学習の入門 6. 最適化と学習アルゴリズム
 
Brief intro : Invariance and Equivariance
Brief intro : Invariance and EquivarianceBrief intro : Invariance and Equivariance
Brief intro : Invariance and Equivariance
 
Flow based generative models
Flow based generative modelsFlow based generative models
Flow based generative models
 
Bidang datar pada ruang (GAR)_Rahmah Fitri
Bidang datar pada ruang (GAR)_Rahmah FitriBidang datar pada ruang (GAR)_Rahmah Fitri
Bidang datar pada ruang (GAR)_Rahmah Fitri
 
Euler and runge kutta method
Euler and runge kutta methodEuler and runge kutta method
Euler and runge kutta method
 
CMSI計算科学技術特論B(12) 大規模量子化学計算(1)
CMSI計算科学技術特論B(12) 大規模量子化学計算(1)CMSI計算科学技術特論B(12) 大規模量子化学計算(1)
CMSI計算科学技術特論B(12) 大規模量子化学計算(1)
 
Prml3.5 エビデンス近似〜
Prml3.5 エビデンス近似〜Prml3.5 エビデンス近似〜
Prml3.5 エビデンス近似〜
 
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems. (AAAI 2022)”
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems.  (AAAI 2022)” [DL輪読会]“Meta-Learning for Online Update of Recommender Systems.  (AAAI 2022)”
[DL輪読会]“Meta-Learning for Online Update of Recommender Systems. (AAAI 2022)”
 
סיכום הקורס במבנים אלגבריים
סיכום הקורס במבנים אלגברייםסיכום הקורס במבנים אלגבריים
סיכום הקורס במבנים אלגבריים
 
Soal dan pembahasan operasi biner dan teori grup dasar - mathcyber1997
Soal dan pembahasan   operasi biner dan teori grup dasar - mathcyber1997Soal dan pembahasan   operasi biner dan teori grup dasar - mathcyber1997
Soal dan pembahasan operasi biner dan teori grup dasar - mathcyber1997
 
P1 . norm vector space
P1 . norm vector spaceP1 . norm vector space
P1 . norm vector space
 
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1東京都市大学 データ解析入門 4 スパース性と圧縮センシング1
東京都市大学 データ解析入門 4 スパース性と圧縮センシング1
 
Prml07
Prml07Prml07
Prml07
 
The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)
 
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving
[DL輪読会]SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving
 
Gaussian Process Regression
Gaussian Process Regression  Gaussian Process Regression
Gaussian Process Regression
 

Viewers also liked

Face recognition using laplacian faces
Face recognition using laplacian facesFace recognition using laplacian faces
Face recognition using laplacian facesPulkiŧ Sharma
 
This one genre theory powerpoint
This one genre theory powerpointThis one genre theory powerpoint
This one genre theory powerpointbir
 
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detectionTaleb ALASHKAR
 
Using Topological Data Analysis on your BigData
Using Topological Data Analysis on your BigDataUsing Topological Data Analysis on your BigData
Using Topological Data Analysis on your BigDataAnalyticsWeek
 
Topological Data Analysis: visual presentation of multidimensional data sets
Topological Data Analysis: visual presentation of multidimensional data setsTopological Data Analysis: visual presentation of multidimensional data sets
Topological Data Analysis: visual presentation of multidimensional data setsDataRefiner
 

Viewers also liked (9)

Kernel Methods on Manifolds
Kernel Methods on ManifoldsKernel Methods on Manifolds
Kernel Methods on Manifolds
 
Lec16 subspace optimization
Lec16 subspace optimizationLec16 subspace optimization
Lec16 subspace optimization
 
Face recognition using laplacian faces
Face recognition using laplacian facesFace recognition using laplacian faces
Face recognition using laplacian faces
 
Multimedia Communication Lec02: Info Theory and Entropy
Multimedia Communication Lec02: Info Theory and EntropyMultimedia Communication Lec02: Info Theory and Entropy
Multimedia Communication Lec02: Info Theory and Entropy
 
This one genre theory powerpoint
This one genre theory powerpointThis one genre theory powerpoint
This one genre theory powerpoint
 
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
 
Genre theory quotes
Genre theory quotesGenre theory quotes
Genre theory quotes
 
Using Topological Data Analysis on your BigData
Using Topological Data Analysis on your BigDataUsing Topological Data Analysis on your BigData
Using Topological Data Analysis on your BigData
 
Topological Data Analysis: visual presentation of multidimensional data sets
Topological Data Analysis: visual presentation of multidimensional data setsTopological Data Analysis: visual presentation of multidimensional data sets
Topological Data Analysis: visual presentation of multidimensional data sets
 

Similar to Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification

Graph convolutional networks in apache spark
Graph convolutional networks in apache sparkGraph convolutional networks in apache spark
Graph convolutional networks in apache sparkEmiliano Martinez Sanchez
 
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsMethods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsRyan B Harvey, CSDP, CSM
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problemsjfrchicanog
 
JOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJordan Open Source Association
 
Machine learning for medical imaging data
Machine learning for medical imaging dataMachine learning for medical imaging data
Machine learning for medical imaging dataliyiou
 
Machine Learning Foundations for Professional Managers
Machine Learning Foundations for Professional ManagersMachine Learning Foundations for Professional Managers
Machine Learning Foundations for Professional ManagersAlbert Y. C. Chen
 
Neo4j MeetUp - Graph Exploration with MetaExp
Neo4j MeetUp - Graph Exploration with MetaExpNeo4j MeetUp - Graph Exploration with MetaExp
Neo4j MeetUp - Graph Exploration with MetaExpAdrian Ziegler
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problemsjfrchicanog
 
Compressing Graphs and Indexes with Recursive Graph Bisection
Compressing Graphs and Indexes with Recursive Graph Bisection Compressing Graphs and Indexes with Recursive Graph Bisection
Compressing Graphs and Indexes with Recursive Graph Bisection aftab alam
 
MLIP - Chapter 6 - Generation, Super-Resolution, Style transfer
MLIP - Chapter 6 - Generation, Super-Resolution, Style transferMLIP - Chapter 6 - Generation, Super-Resolution, Style transfer
MLIP - Chapter 6 - Generation, Super-Resolution, Style transferCharles Deledalle
 
Matrix Computations in Machine Learning
Matrix Computations in Machine LearningMatrix Computations in Machine Learning
Matrix Computations in Machine Learningbutest
 
Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)Mumbai Academisc
 
Recent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationRecent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationChristopher Morris
 
Elementary Landscape Decomposition of the Test Suite Minimization Problem
Elementary Landscape Decomposition of the Test Suite Minimization ProblemElementary Landscape Decomposition of the Test Suite Minimization Problem
Elementary Landscape Decomposition of the Test Suite Minimization Problemjfrchicanog
 
Software tookits for machine learning and graphical models
Software tookits for machine learning and graphical modelsSoftware tookits for machine learning and graphical models
Software tookits for machine learning and graphical modelsbutest
 

Similar to Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification (20)

Graph convolutional networks in apache spark
Graph convolutional networks in apache sparkGraph convolutional networks in apache spark
Graph convolutional networks in apache spark
 
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data SetsMethods of Manifold Learning for Dimension Reduction of Large Data Sets
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problems
 
JOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured Data
 
Lec17 sparse signal processing & applications
Lec17 sparse signal processing & applicationsLec17 sparse signal processing & applications
Lec17 sparse signal processing & applications
 
Machine learning for medical imaging data
Machine learning for medical imaging dataMachine learning for medical imaging data
Machine learning for medical imaging data
 
Machine Learning Foundations for Professional Managers
Machine Learning Foundations for Professional ManagersMachine Learning Foundations for Professional Managers
Machine Learning Foundations for Professional Managers
 
Neo4j MeetUp - Graph Exploration with MetaExp
Neo4j MeetUp - Graph Exploration with MetaExpNeo4j MeetUp - Graph Exploration with MetaExp
Neo4j MeetUp - Graph Exploration with MetaExp
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problems
 
Compressing Graphs and Indexes with Recursive Graph Bisection
Compressing Graphs and Indexes with Recursive Graph Bisection Compressing Graphs and Indexes with Recursive Graph Bisection
Compressing Graphs and Indexes with Recursive Graph Bisection
 
MLIP - Chapter 6 - Generation, Super-Resolution, Style transfer
MLIP - Chapter 6 - Generation, Super-Resolution, Style transferMLIP - Chapter 6 - Generation, Super-Resolution, Style transfer
MLIP - Chapter 6 - Generation, Super-Resolution, Style transfer
 
Lec14 eigenface and fisherface
Lec14 eigenface and fisherfaceLec14 eigenface and fisherface
Lec14 eigenface and fisherface
 
Class01
Class01Class01
Class01
 
Matrix Computations in Machine Learning
Matrix Computations in Machine LearningMatrix Computations in Machine Learning
Matrix Computations in Machine Learning
 
Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)Face recognition using laplacianfaces (synopsis)
Face recognition using laplacianfaces (synopsis)
 
Presentation on machine learning
Presentation on machine learningPresentation on machine learning
Presentation on machine learning
 
Recent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationRecent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph Classification
 
"Let us talk about output features! by Florence d’Alché-Buc, LTCI & Full Prof...
"Let us talk about output features! by Florence d’Alché-Buc, LTCI & Full Prof..."Let us talk about output features! by Florence d’Alché-Buc, LTCI & Full Prof...
"Let us talk about output features! by Florence d’Alché-Buc, LTCI & Full Prof...
 
Elementary Landscape Decomposition of the Test Suite Minimization Problem
Elementary Landscape Decomposition of the Test Suite Minimization ProblemElementary Landscape Decomposition of the Test Suite Minimization Problem
Elementary Landscape Decomposition of the Test Suite Minimization Problem
 
Software tookits for machine learning and graphical models
Software tookits for machine learning and graphical modelsSoftware tookits for machine learning and graphical models
Software tookits for machine learning and graphical models
 

More from United States Air Force Academy

More from United States Air Force Academy (11)

Lec-08 Feature Aggregation II: Fisher Vector, AKULA and Super Vector
Lec-08 Feature Aggregation II: Fisher Vector, AKULA and Super VectorLec-08 Feature Aggregation II: Fisher Vector, AKULA and Super Vector
Lec-08 Feature Aggregation II: Fisher Vector, AKULA and Super Vector
 
Lec07 aggregation-and-retrieval-system
Lec07 aggregation-and-retrieval-systemLec07 aggregation-and-retrieval-system
Lec07 aggregation-and-retrieval-system
 
Lec11 object-re-id
Lec11 object-re-idLec11 object-re-id
Lec11 object-re-id
 
Lec12 review-part-i
Lec12 review-part-iLec12 review-part-i
Lec12 review-part-i
 
Lec11 rate distortion optimization
Lec11 rate distortion optimizationLec11 rate distortion optimization
Lec11 rate distortion optimization
 
Lec-03 Entropy Coding I: Hoffmann & Golomb Codes
Lec-03 Entropy Coding I: Hoffmann & Golomb CodesLec-03 Entropy Coding I: Hoffmann & Golomb Codes
Lec-03 Entropy Coding I: Hoffmann & Golomb Codes
 
ECE 4490 Multimedia Communication Lec01
ECE 4490 Multimedia Communication Lec01ECE 4490 Multimedia Communication Lec01
ECE 4490 Multimedia Communication Lec01
 
Mobile Visual Search: Object Re-Identification Against Large Repositories
Mobile Visual Search: Object Re-Identification Against Large RepositoriesMobile Visual Search: Object Re-Identification Against Large Repositories
Mobile Visual Search: Object Re-Identification Against Large Repositories
 
Tutorial on MPEG CDVS/CDVA Standardization at ICNITS L3 Meeting
Tutorial on MPEG CDVS/CDVA Standardization at ICNITS L3 MeetingTutorial on MPEG CDVS/CDVA Standardization at ICNITS L3 Meeting
Tutorial on MPEG CDVS/CDVA Standardization at ICNITS L3 Meeting
 
Light Weight Fingerprinting for Video Playback Verification in MPEG DASH
Light Weight Fingerprinting for Video Playback Verification in MPEG DASHLight Weight Fingerprinting for Video Playback Verification in MPEG DASH
Light Weight Fingerprinting for Video Playback Verification in MPEG DASH
 
Scaled Eigen Appearance and Likelihood Prunning for Large Scale Video Duplica...
Scaled Eigen Appearance and Likelihood Prunning for Large Scale Video Duplica...Scaled Eigen Appearance and Likelihood Prunning for Large Scale Video Duplica...
Scaled Eigen Appearance and Likelihood Prunning for Large Scale Video Duplica...
 

Recently uploaded

TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clashcharlottematthew16
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 

Recently uploaded (20)

TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clash
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 

Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification

  • 1. Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification Zhu Li Media Networking Lab FutureWei (Huawei) Technology, USA Bridgewater, NJ IBM Research, 2012 1 Z. Li
  • 2. Outline • Short Self-Intro • Large Scale Visual Analytics – Applications – Key Technical Challenges – Query-Driven Local Subspaces – Indexed Subspaces on Grassmannian Manifold – Simulation – Conclusion & Future Work IBM Research, 2012 2 Z. Li
  • 3. About Me: http://users.eecs.northwestern.edu/~zli • Bio: – Sr. Staff Researcher, Core Networks R&D, Huawei Tech USA, 2010.10~ to date – Asst Prof, HK Polytechnic Univ, CTO, Mudi Tech, 2008.04~2010.09 – Senior, Senior Staff, and then Principal Staff Researcher, Multimedia Research Lab, Motorola Labs, USA, 2000-08. • Research Interests: – Biometrics, large scale fingerprint identification, face identification, behaviour biometrics – Visual search, Compact Visual Descriptor for Mobile Search. Check out Mudi's QbC visual search solution for dangdang.com: http://i.joyton.com/download/DangDang2.0.apk – Large scale audio/visual data analysis, storage and indexing, search and mining, Hadoop based map-reduce scheme for large scale visual analytics. IBM Research, 2012 3 Z. Li
  • 4. The Large Scale Visual Analytics Problems • Face Recognition – Identify face from 7 million HK ID face data set • Image Search – Find out the category of given images IBM Research, 2012 4 Z. Li
  • 5. The Problem • Identification – Given a set of training image data and label{fk, lk}, and a probe p, identify the unique label associated with p. • Why is it difficult ? – When the number of unique labels, m, and training data n are large.... X = f(I) IBM Research, 2012 5 Z. Li
  • 6. Appearance Modeling • Find a “good” f() – Such that after projecting the appearance onto the subspace, the data points belong to different classes are easily separable X = f(I) IBM Research, 2012 6 Z. Li
  • 7. Global Linear LPP Models: f(X) = AX • LPP (Xiaofei He, et.al): - Minimizing weighted distance (a graph) after projection X min wj;k jjAxj ¡ Axk jj2 A j;k -Solve by: X T T XLX A = ¸XDX A; s:t:L = D ¡ W; Dk;k = wj;k j - Embed a graph with pruned edges ( wj;k = e¡®jjxj ¡xk jj ; if jjxj ¡ xk jj · ² 0; else IBM Research, 2012 7 Z. Li
  • 8. Global Linear LDA Models: f(X)=AX • LDA: - Maximizing inter-class scatter over intra A = arg max | AT S B A |, s.t. | AT SW A |= 1 A n n s B = ∑ nk ( X k − X )( X k − X ) T sW = ∑ ∑ ( X j − X k )( X j − X k )T k =1 k =1 P ( X j )=k -Solve by: S B A = λSW A - Embedding a graph with no edges among inter-class points ( 1 wj;k = mi ; if xj ; xk 2 class i 0; else IBM Research, 2012 8 Z. Li
  • 9. Graph Embedding Interpretation • Find the best embedding – LDA: » preserve the affinity matrix that has zero affinity for data points pairs that are not belonging to the same class – LPP: » Have more flexibility in modeling affinity w jk. LPP Affinity LDA Affinity IBM Research, 2012 9 Z. Li
  • 10. Non-Linear Models • Appearance manifolds are non-linear in nature – Global linear models will suffer • Non-Linear Solutions: – Kernel method: e.g K-PCA, K-LDA, K-LPP, SVM » Evaluate inner product <xj,xk> with a kernel function k(xj, xk), which if satisfy the conditions in Mercer’s Theorem, implicitly maps data via a non-linear function. » Typically involves a QP problem with a Hessian of size n x n, when n is large, not solvable. – LLE /Graph Laplacian: » An algorithm that maps input data {xk} to {yk} that tries to preserve an embedded graph structure among data points. » The mapping is data dependent and has difficulty handling new data outside the training set, e.g., a new query point • How to compromise ? – Piece-wise Linear Approximation IBM Research, 2012 10 Z. Li
  • 11. Piece-wise Linear : Query Driven • Query-Driven Piece-wise Linear Model – No pre-determined structure on the training data – Local neighborhood data patch identified from query point q, – Local model built with local data, A(X, q) X Local Graph Embedding Projection A(X,q) q Y=A(X,q)X + Local data: N(X, q) + IBM Research, 2012 11 Z. Li
  • 12. Local Model Discriminating Power Criteria • What is a good N(X, q) ? • Model power of a linear model: – A: Dxd, D=wxh • Data Complexity: Graph Embedding Interpretation: –PCA: a fully connected graph –LDA: a graph with edges pruned for intra-class points –LPP/LEA; k-nn/ ε − nn pruned graph –as number of edges/relationship among data points ( n ), 2 PCA  m m    ∑ ∑ nj | E ( X ) |=  ( 2 ), s.t. n j = n, LDA   j =1 j =1  nK , LPP / LEA   • What is a good compromise of data complexity and model power ? IBM Research, 2012 12 Z. Li
  • 13. Discriminant Power Co-efficient (DPC) • Given the model power constraint: – w, h, appearance model luminance field size – d, dimensionality of A(x, q) • How to identify a neighborhood to achieve a good balance of data complexity and model power ? - DPC, K(A(X,q)) = w× h× d | E ( X (q ) ) | - Need to balance DPC with info loss in node/edge pruning IBM Research, 2012 13 Z. Li
  • 14. Head Pose Recognition Performance • Recognition rate is improved: – W=18, h=18, K=30 • And the cost in computation is rather modest – Matlab code, online local model A(X,q) learning and NN classification: IBM Research, 2012 14 Z. Li
  • 15. Face Recognition Performance • Local model combination in face recognition – Query point drives 3 local models, A1(X, q), A2(X, q), A3(X, q) – Local model classification error estimation, – Combining the results – weighted voting Multiple face models with different area and scale: (a) Upper face model (18 × 16). (b) Lower face model (14 × 18). (c) Full face model (21 × 28). ORL data set test: leave 1,2,3 out: IBM Research, 2012 15 Z. Li
  • 16. Query Driven Solution Problems • Optimality of the Local Model is not established – Parameters ² ¡ N N , k-NN, and heat kernel size determines the number of non-zero affinity edges in local graph – The choice is based on DPC, which is still heuristic • Computational Complexity – Need to compute a nearest neighbor set and its affinity, as well as the local embedding model at run time. – Need extra storage to store all training data, because the local NN data patch is generated at run time, as function of the query point. – Indexing/Hashing scheme to support efficient access of training data. IBM Research, 2012 16 Z. Li
  • 17. Stiefel and Grassmannian Manifolds • Stiefel manifolds – All possible p-dimensional subspaces in d-dimensional space, A pxd, spans Stiefel Manifold, S(p, d) in Rdxp, d > p. © d£p 0 ª S(p; d) = A 2 R ; s:t:A A = Id – The DoF is not pxd, rather: pd – (1/2)d(d+1) • Grassmannian manifolds – G(p, d) identifies p-dimensional subspaces in d-dimensional space – It is stiefel manifolds but with an equivalence constraint: » A1 = A2, if span(A1) = span(A2), or » Exist othonormal dxd matrix Rd, A1=A2Rd. – The DoF: pd-d2. G(p, d) is the quotient space of S(p, d)/O(d) IBM Research, 2012 17 Z. Li
  • 18. Subspaces on Grassmannian Manifold • The BEST subspace for identification ? – All possible p-dimensional subspaces in d-dimensional space, A pxd, spans Grassmannian Manifold, G(p, d) in Rdxp, d > p. » eg., G(2, 3), biz card example – The DoF of A is not pxd, as for, < aj ; ak >= 0; < aj ; aj >= 1; for AT = [a1 ; a2 ; :::; ap ]; – Face Appearance model, typically, d=400~500, p=10~30. – The BEST subspace A* is somewhere on G(p, d), therefore it is important to figure out a way to characterize the similarity between subspaces in G(p, d), and give a structure of all subspace w.r.t the task of identification. IBM Research, 2012 18 Z. Li
  • 19. Grassmannian Manifold Visualization • Consider a typical appearance modeling – Image size 12x10 pel, appearance space dimension d=120, model dimension p=8. – 3D visualization of all S(8, 120) and their covariance eigenvalues” – Grassmann Manifolds are quotient space S(8, 120)/O(8) IBM Research, 2012 19 Z. Li
  • 20. Principle Angles • The principle angles between two subspaces: – For Y1, and Y2 in G(p, d), their principle angles are defined as cos(µk ) = max u0 vk k span(A1 ) span(A2 ) uk 2span(A1 );vk 2span(A2 ) ( u0 uk = 1; vk vk = 1 k 0 s:t: u0 ui = 0; vk vi = 0 k 0 – Where, {uk} and {vk} are called principle dimensions for span(A1) and span(A2). IBM Research, 2012 20 Z. Li
  • 21. Principle Angles Computing • The principle angles between two subspaces: – For A1, and A2 in G(p, d), their principle dimensions and angles are computed by SVD: [U; S; V ] = SV D(AT A2 ) 1 – Where, U=[u1, u2, …, up], and V=[v1, v2, …, vp] are the principle angles. – The diagonal of S, [s1, s2,..., sp] are the cosine of principle angles, sk = cos(µk ) IBM Research, 2012 21 Z. Li
  • 22. Subspace Distance on Grassmannian Manifold • Subspace distances: – Projection Distance Def: Xp dprj (A1 ; A2 ) = ( sin2 µi )1=2 i=1 Computing: p X d2 (A1 ; A2 ) = p ¡ prj cos2 µi = m ¡ jjA0 A2 jj 2 1 F i=1 – Binet-Cauchy Distance Def: Y dbc (A1 ; A2 ) = (1 ¡ cos2 µi )1=2 i Computing: Y d2 (A1 ; A2 ) bc =1¡ cos2 µi = 1 ¡ det2 (A0 A2 ) 1 i IBM Research, 2012 22 Z. Li
  • 23. Subspace Distance on Grassmannian Manifold • Subspace distances – Arc Distance Def: darc (A1 ; A2 ) = ( X µi )1=2 2 i Also known as geodesic distance. It traverse the Grassmannian surface, and two subspace collapse into one, when all principle angles becomes zero. IBM Research, 2012 23 Z. Li
  • 24. Weighted Merging of two subspaces • What if we need merge two subspaces ? – Motivation: » say if subspace A1 is best for data set S1, and subspace A2 is best for data set S2, can we find a subspace A3 that is good for both ? – When two subspaces are sufficiently close on Grassmannian manifold, we can approximate this by, A3=[t1, t2, ….] n1 n2 tk = uk + vk + n1 + n2 n1 + n2 Where n1,2 are the size of data set S1,2 – The new sets of basis may not be orthogonal. Can be corrected by Gram-Schmidt orthogonalization. IBM Research, 2012 24 Z. Li
  • 25. Judicious Local Models • Data Space Partition – Partition the training data set by kd-tree – For the kd-tree height of h, we have 2h local data patch as leaf node – For each leaf node data patch k, build a local LDA/LPP/PCA model Ak: IBM Research, 2012 25 Z. Li
  • 26. Subspace Index • Organizing the Subspace Models – For data index of height of h, we have 2h local models Ak: k=1..2h. – For a given probe data point, find its leaf node and associated local model, do identification. Is this good ? – No, because » Could be over-fitting, not sure what is the right size local data patch. » Improper neighborhood, probe data points falling on the boundary of leaf node: – Build local models at each subtree ? » No, the data partition does not reflect the smooth change of the local models. IBM Research, 2012 26 Z. Li
  • 27. Model Hierarchical Tree (MHT) • Indexing Subspaces on Grassmannian manifold – It is a VQ like process. – Start with a data partition kd-tree, their leaf nodes and associated subspaces {Ak}, k=1..2h – Repeat » Find Ai and Aj, if darc(Ai, Aj) is the smallest among all, and the associated data patch are adjacent in the data space. » Delete Ai and Aj, replace with merged new subspace, and update associated data patch leaf nodes set. » Compute the empirical identification accuracy for the merged subspace » Add parent pointer to the merged new subspace for A i and Aj . » Stop if only 1 subspace left. – Benefit: » avoid forced merging of subspace models at data patches that are very different, though adjacent. IBM Research, 2012 27 Z. Li
  • 28. MHT Based Identification • MHT operation – Organize the leaf nodes models into a new hierarchy, with new models and associated accuracy (error rate) estimation – When a probe point comes, first identify its leaf nodes from the data partition tree. – Then traverse the MHT from leaf nodes up, until it hits the root, which is the global model, and choose the best model along the path for identification IBM Research, 2012 28 Z. Li
  • 29. Simulation • The data set – MSRA Multimedia data set, 65k images with class and relevance labels: IBM Research, 2012 29 Z. Li
  • 30. Simulation • Data selection and features – Selected 12 classes with 11k images and use the original combined 889d features from color, shape and texture – Performance compared with PCA, LDA and LPP modeling IBM Research, 2012 30 Z. Li
  • 31. Simulation • Face data set – Mixed data set of 242 individuals, and 4840 face images – Performance compared with PCA, LDA and LPP modeling IBM Research, 2012 31 Z. Li
  • 32. Summary • Contributions – The work is a piece-wise linear approximation of non-linear appearance manifold – Query driven provide suboptimal performance but still better than a global model. – It offers best local models for identification by deriving the subspace structure/index with metrics on Grassmannian manifold – Guaranteed performance gains, and the root model degenerates into the global linear model • Limitations – Do not have a continuous characterization of Identification error function on the Grassmann manifold. – Still heavy on storage cost – Need to get more large scale data set to test it. IBM Research, 2012 32 Z. Li
  • 33. Summary • Future work – Grassmann Hashing – Penalize projection selection with Grassmannian metric, offers performance gains over LSH and spectral hashing. – Gradient and Newtonian optimization on Grassmannian manifold. Related papers – X. Wang, Z. Li, and D. Tao, "Subspace Indexing on Grassmann Manifold for Image Search ", IEEE Trans. on Image Processing, vol. 20(9), 2011. – X. Wang, Z. Li, L. Zhang, and J. Yuan, "Grassmann Hashing for Approx Nearest Neighbour Search in High Dimensional Space", Proc. of IEEE Int'l Conf on Multimedia & Expo (ICME), Barcelona, Spain, 2011. – H. Xu, J. Wang, Z. Li, G. Zeng, S. Li, “Complementary Hashing for Approximate Nearest Neighbor Search”, IEEE Int'l Conference on Computer Vision (ICCV), Barcelona, Spain, 2011. – Yun Fu, Z. Li, J. Yuan, Ying Wu, and Thomas S. Huang, "Locality vs. Globality: Query-Driven Localized Linear Models for Facial Image Computing," IEEE Transactions on Circuits and Systems for Video Technology (T-CSVT), vol. 18(12), pp. 1741-1752, December, 2008. IBM Research, 2012 33 Z. Li
  • 34. Acknowledgement • Grants: – The work is partially supported by; » a Hong Kong RGC Grant, and » Microsoft Research Asia faculty grant. • Collaborators: » Xinchao Wang, valedictorian of Dept of COMP, HK Polytechnic University, class 2010, now PhD at EPFL » Dacheng Tao, Professor at Univ of Technology of Sydney. IBM Research, 2012 34 Z. Li
  • 35. Q&A • Questions please...... IBM Research, 2012 35 Z. Li
  • 36. Thanks ! IBM Research, 2012 36 Z. Li