Advanced Techniques for Massive Graph Analysis

all distances sketches revisited hip estimators n.w
1 / 26
Embed
Share

Explore the cutting-edge methodology of All-Distances Sketches (ADS) and HIP Estimators for analyzing massive graphs efficiently. These techniques offer compact size, scalability, and various applications in areas like viral marketing, recommendations, and more, making them essential for understanding complex network structures.

  • Graph Analysis
  • HIP Estimators
  • Massive Graphs
  • Data Analytics
  • Advanced Techniques

Uploaded on | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. All-Distances Sketches, Revisited: HIP Estimators for Massive Graphs Analysis Edith Cohen Microsoft Research Presented by: Thomas Pajor Microsoft Research

  2. Very Large Graphs Model many types of relations and interactions Call detail data, email exchanges Web crawls Social Networks: Twitter, Facebook, linkedIn Web searches, Commercial transactions, Need for scalable analytics: Centralities/Influence (power/importance/coverage of a node or a set of nodes): Viral marketing, Similarities/Communities (how tightly related are 2 or more nodes): Recommendations, Advertising, Marketing

  3. All-Distances Sketches (ADS) [C 94] Summary structures: For each node ? ? : ???(?) samples the distance relations of ? to all other nodes. Useful for queries involving a single node: Neighborhood cardinality and statistics Sketches of different nodes are coordinated: related in a way that is useful for queries that involve multiple nodes (similarities, influence, distance)

  4. All-Distances Sketches (ADS) [C 94] Basic properties ? edges, ? nodes, parameter ? 1 which controls trade-off between sketch size and information ADSs work for directed or undirected graphs Compact size: ?[ ??? ? ] ? ln ? Scalable Computation: ?? ln? edge traversals to compute ??? ? for all nodes ? Many applications

  5. All-Distances Sketches: Definition ???(?)is a list of pairs of the form (?,???) Draw a random permutation of the nodes: ? ? [?] ? ??? ? ?(?) < kthsmallest rank amongst nodes that are closer to ? than? This is a bottom-? ADS, it is the union of bottom-? MinHash sketches (?smallest rank) of all neighborhoods. There are other ADS flavors , vary by the rank distribution ? (e.g. can use ?(?) ?[0,1] ) or sketch structure.

  6. ADS example SP distances: 10 16 13 14 15 15 10 0 5 6 7 17 17 0.28 0.77 0.07 5 0.49 1 0.14 3 3 4 6 7 5 4 4 4 3 10 10 5 10 3 0.91 0.70 2 0.35 10 10 10 7 0.21 5 3 4 6 0.63 0.84 4 Random permutation of nodes 0.56 0.42

  7. ADS example ? = 1 All nodes sorted by SP distance from 0.63 0.42 0.56 0.84 0.07 0.35 0.49 0.77 0.91 0.21 0.28 0.14 0.70 ? = 1: 0.63 0.42 0.07

  8. ADS example ? = 2 Sorted by SP distances from 0.63 0.42 0.56 0.84 0.07 0.35 0.49 0.77 0.91 0.21 0.28 0.14 0.70 ? = 2: 0.63 0.42 0.56 0.07 0.35 0.21 0.14

  9. Basic use of ADSs (90s 2013) Extract MinHash sketch of the ? neighborhood of ?, ??? , from ADS(?) : bottom-?{? ???(?)|???< ? } From MinHash sketches, we can estimate: Cardinality |??? | Estimate has CV ? ? information in the MinHash sketch) Jaccard similarity of ??(?) and ??(?), Other relations of ??(?) and ??(?), 1 ? 2 (optimally uses the

  10. Historic Inverse Probability (HIP) inclusion probability & estimator For each node ?, we estimate the presence of ? with respect to ?:?? ?(=1 if ? ?, 0 otherwise) Estimate is ???= ?if ? ADS(?). If ? ADS(?), we compute the probability ? that it is included, conditioned on fixed rank values of all nodes that are closer to ? than ?. We then use the inverse-probability estimate ???=? ?. [HT52] This is unbiased (when ? > 0):

  11. Bottom-? HIP For bottom-? and ? ?[0,1] ? = kth{?(?)|? ADS ? ???< ???} HIP can be used with all flavors of MinHash sketches. Over distance (ADS) or time (Streams)

  12. Example: HIP estimates Bottom-2 ADS of 0.63 0.42 0.56 0.07 0.35 0.21 0.14 ?: 0.63 0.56 0.42 0.35 0.21 1 1 ? =1 1.59 1.79 2.38 2.86 4.76 1 1 ?: ?:2nd smallest r value among closer nodes

  13. HIP cardinality estimate Bottom-2 ADS of 0 5 6 10 10 15 17 distance: ?: 1 1 1.59 1.79 2.38 2.86 4.76 Query: ??? = { ? ??? ? } | = ????? ? ??(?) = ???= 1 + 1 + 1.59 = 3.59 ?,??? ??? ? | ??? ?

  14. Quality of HIP cardinality Estimate Lemma: The HIP neighborhood cardinality estimator ??(?) = ??? ?,??? ??? ? | ??? ? has CV ? 1 ? 2? 2 This is 2 improvement over basic estimators, which have CV ? ? 1 ? 2 See paper for the proof

  15. HIP versus Basic estimators Basic X 2 HIP

  16. HIP: applications Querying ADSs: Cardinality estimation: 2 gain in relative error over basic (MinHash based) estimates More complex queries: closeness centrality with topic awareness (gain can be polynomial) Estimating relations (similarities, coverage) of pairs (sets) of nodes . Streaming: Approximate distinct counting on streams.

  17. Topic-aware Distance-decay Closeness Centrality ?(???)?(?) ??= ? ? non increasing; ? some filter Centrality with respect to a filter ?(?): : Topic, interests, education level, age, community, geography, language, product type Applications for filter: attribute completion, targeted advertisements

  18. .Closeness Centrality ?(???)?(?) ??= ? ? non increasing; ? some filter Polynomial (Harmonic) decay: ? ? =1 Exponential decay ? ? = ? ? Threshold ( ??? ): ? ? = 1 ? ? ?

  19. HIP estimates of Centrality ?(???)?(?) ??= ? ? non increasing; ? some filter ??? ?(???) ?(?) ?? = ? ???(?)

  20. HIP estimates: closeness to good/evil Bottom-2 ADS of 0 5 6 10 10 15 17 distance: ?: 1 1 1.59 1.79 2.38 2.86 4.76 ?: ? ? ? ?.? ?.? ? ?.? Filter: ? 0,1measures goodness Distance-decay: ? ??? ????(?)? ???= ? ?+ ? ?+ ?.?? ??+ ?? = ? ???(?)

  21. Counting Distinct Elements on Data Stream 32, 12, 14, 32, 7, 12, 32, 7, 6, 12, 4, Elements occur multiple times, we want to count the number of distinct elements approximately with small storage, about O(log log ?) Best practical and theoretical algorithms maintain a MinHash sketch. Cardinality is estimated by applying an estimator to sketch [Flajolet Martin 85], Best (in practice) is the HyperLogLog (HLL) algorithm and variations. [Flajolet + FGM 2007],

  22. Counting Distinct Elements with HIP We maintain a MinHash sketch and an approximate counter -- variation on [Morris77]. The counter explicitly maintains an approximate distinct count. Each time the sketch is updated (E ?ln ? times ), we increase the counter (add the HIP estimate for the inserted new distinct element) The approximate counter can be represented with few bits (e.g., can be a relative correction to sketch- based estimate or share its exponent ) This works with any MinHash sketch. In experiments, for comparison, we use the same sketch as HyperLogLog (HLL).

  23. HLL vs. HIP (on HLL sketch)

  24. Conclusion ADS: old but a very versatile and powerful tool for (scalable approximate) analytics on very large graphs: distance/similarity oracles, distance distribution, closeness, coverage, influence, tightness of communities HIP: simple and practical technique, applicable with ADSs and streams Further ADS+HIP applications: closeness similarity (using ADS+HIP) [CDFGGW COSN 2013] Timed-influence oracle

  25. Thank you!!

  26. Ninjago Legends of Chima Nya Samurai X Acidicus Cragger Laval Kai Ninja of Fire Rascal Lloyd The Green Ninja Eris Star Wars Sensei Wu R2-D2 Darth VaderLuke Yoda Skywalker

Related


More Related Content