Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Exact Recovery of Low-Rank Plus Compressed Sparse Matrices Morteza Mardani, Gonzalo Mateos and Georgios Giannakis ECE Department, University of Minnesota Acknowledgments: MURI (AFOSR FA9550-10-1-0567) grant Ann Arbor, USA August 6, 2012 1 Context Occam’s razor Image processing Among competing hypotheses, pick one which makes the fewest assumptions and thereby offers the simplest explanation of the effect. Network cartography x 10 7 4 |xf,t| 3 2 1 0 0 100 200 300 Time index (t) 400 500 2 Objective F T L Goal: Given data matrix and low rank and compression matrix , identify sparse . 3 Challenges and importance Both rank( ) and support of Seriously underdetermined generally unknown LT X + FT A (F ≥ L) >> LT Y Important special cases R = I : matrix decomposition [Candes et al’11], [Chandrasekaran et al’11] X = 0 : compressive sampling [Candes et al’05] A = 0: (with noise) PCA [Pearson’1901] 4 Unveiling traffic anomalies Backbone of IP networks Traffic anomalies: changes in origin-destination (OD) flows Failures, transient congestions, DoS attacks, intrusions, flooding Anomalies congestion limits end-user QoS provisioning Measuring superimposed OD flows per link, identify anomalies 5 Model Graph G (N, L) with N nodes, L links, and F flows (F=O(N2) >> L) (as) Single-path per OD flow zf,t 1 0.9 f2 0.8 0.7 Packet counts per link l and time slot t l 0.6 0.5 Anomaly f1 0.4 0.3 0.2 0.1 0 0 0.2 0.4 0.6 0.8 1 є {0,1} Matrix model across T time slots LxT LxF 6 Low rank and sparsity Z: traffic matrix is7 low-rank [Lakhina et al‘04] X: low-rank x 10 4 |xf,t| 3 2 1 0 0 100 200 300 Time index (t) 400 500 A: anomaly matrix is sparse across both time and flows 8 8 x 10 4 |af,t| |af,t| 4 2 0 0 200 400 600 Time index(t) 800 1000 x 10 2 0 0 50 Flow index(f) 100 7 Criterion Low-rank sparse vector of SVs nuclear norm || ||* and l1 norm (P1) Q: Can one recover sparse and low-rank exactly? A: Yes! Under certain conditions on 8 Identifiability Y = X0+ RA0 = X0 + RH + R(A0 - H) X1 Problematic cases but For low-rank and A1 sparse , and r = rank(X0), low-rank-preserving matrices RH Sparsity-preserving matrices RH 9 Incoherence measures ΩR θ=cos-1(μ) Ф Incoherence among columns of R , Local identifiability requires , Exact recovery requires , Incoherence between X0 and R 10 Main result Theorem: Given and , if every row and column of k non-zero entries and , then imply Ǝ has at most for which (P1) exactly recovers M. Mardani, G. Mateos, and G. B. Giannakis,``Recovery of low-rank plus compressed sparse matrices with application to unveiling traffic anomalies," IEEE Trans. Info. Theory, submitted Apr. 2012.(arXiv: 1204.6537)11 Intuition Exact recovery if r and s are sufficiently small Nonzero entries of A0 are “sufficiently spread out” Columns (rows) of X0 not aligned with basis of R (canonical basis) R behaves like a “restricted” isometry Interestingly Amplitude of non-zero entries of A0 irrelevant No randomness assumption Satisfiability for certain random ensembles w.h.p 12 Validating exact recovery 50 0.7 0.6 40 R rank(X ) (r) Setup L=105, F=210, T = 420 R=URSRV’R~ Bernoulli(1/2) X = V’R WZ’, W, Z ~ N(0, 104/FT) aij ϵ {-1,0,1} w. prob. {ρ/2, 1-ρ, ρ/2} 0.5 30 0.4 0.3 20 0.2 10 0.1 Relative recovery error 0.1 0 2.5 4.5 6.5 8.5 10.5 12.5 % non-zero entries () 13 Real data Abilene network data Dec. 8-28, 2008 N=11, L=41, F=121, T=504 ---- True ---- Estimated 0.8 [Lakhina04], rank=1 [Lakhina04], rank=2 [Lakhina04], rank=3 Proposed method [Zhang05], rank=1 [Zhang05], rank=2 [Zhang05], rank=3 0.6 0.4 0.2 0 0 0.2 0.4 0.6 0.8 False alarm probability 6 1 Anomaly amplitude Detection probability 1 5 4 3 2 1 0 Pf = 0.03 Pd = 0.92 Qe = 27% 100 500 400 300 Flow 50 0 Data: http://internet2.edu/observatory/archive/data-collections.html 100 0 200 Time 14 Synthetic data Random network topology 1 0.8 0.6 N=20, L=108, F=360, T=760 Minimum hop-count routing 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 1 Detection probability 0.8 ---- True ---- Estimated 0.6 0.4 PCA-based method, r=5 PCA-based method, r=7 PCA-based method, r=9 Proposed method, per time and flow 0.2 0 0 0.2 0.4 0.6 False alarm probability 0.8 1 Pf=10-4 Pd = 0.97 15 Distributed estimator Centralized (P2) Network: undirected, connected graph ? ? ? ? ? ? ? n ? Challenges Nuclear norm is not separable Global optimization variable A Key result [Recht et al’11] M. Mardani, G. Mateos, and G. B. Giannakis, "In-network sparsity-regularized rank minimization: Algorithms and applications," IEEE Trans. Signal Process., submitted Feb. 2012.(arXiv: 1203.1570) 16 Consensus and optimality (P3) Consensus with neighboring nodes Alternating-directions method of multipliers (ADMM) solver Highly parallelizable with simple recursions Low overhead for message exchanges n Claim: Upon convergence attains the global optimum of (P2) 17 Online estimator Streaming data: (P4) ATLA--HSTN CHIN--ATLA 5 4 0 DNVR--KSCY 20 10 0 HSTN--ATLA 20 Anomaly amplitude Link traffic level 2 0 WASH--STTL 40 20 0 WASH--WASH 30 20 10 10 0 0 Time index (t) ---- estimated ---- real 0 1000 2000 3000 4000 5000 6000 Time index (t) o---- estimated ---- real Claim: Convergence to stationary point set of the batch estimator M. Mardani, G. Mateos, and G. B. Giannakis, "Dynamic anomalography: Tracking network anomalies via 18 sparsity and low rank," IEEE Journal of Selected Topics in Signal Processing, submitted Jul. 2012.