Model-Independent Online Learning for Influence Maximization Sharan - - PowerPoint PPT Presentation

model independent online learning for influence
SMART_READER_LITE
LIVE PREVIEW

Model-Independent Online Learning for Influence Maximization Sharan - - PowerPoint PPT Presentation

Model-Independent Online Learning for Influence Maximization Sharan Vaswani 1 , Branislav Kveton 2 , Zheng Wen 2 , Mohammad Ghavamzadeh 3 , Laks Lakshmanan 1 , Mark Schmidt 1 1 University of British Columbia 2 Adobe Research 3 Deepmind 1


slide-1
SLIDE 1

Model-Independent Online Learning for Influence Maximization

Sharan Vaswani1, Branislav Kveton2, Zheng Wen2, Mohammad Ghavamzadeh3, Laks Lakshmanan1, Mark Schmidt1

1 University of British Columbia 2 Adobe Research 3 Deepmind

1

slide-2
SLIDE 2

Influence Maximization

Underlying principle: Influence propagates through ‘word of mouth’ in a social network Idea: Give discounts to ‘influential’ users who will trigger off word-of-mouth epidemics Aim: Find a subset of users (‘seed set’) who will influence maximum people to become aware of a product

2

slide-3
SLIDE 3

Influence Maximization

Weighted Graph Set of feasible seed sets Stochastic diffusion Model E.g:

Input:

3

Key Property: For common diffusion models, is submodular in S Objective: Find

Pr (S influences target node v) Expected number of nodes influenced by S

slide-4
SLIDE 4

Influence Maximization

4

Challenges to using IM in practice:

[1] Du, Nan, et al. "Influence function learning in information diffusion networks." ICML, 2014 [2] Goyal, Amit, et al. "Learning influence probabilities in social networks." WSDM, 2010

  • Challenge 2: Learning model parameters requires considerable data, often

unavailable to a new marketer.

  • Challenge 1: IM is not robust to the choice of the diffusion model [1] nor

its model parameters [2].

slide-5
SLIDE 5

Model Independent Formulation

5

Assumption 1: F(S) is monotonic in S. Key Idea: Parametrize the problem in terms of pairwise reachability probabilities

Pr(u influences v under a diffusion model)

Advantages:

  • Common parametrization for all progressive models.
  • Guaranteed approximation.
  • Surrogate objective is submodular irrespective of the

diffusion model. Surrogate Objective: Find

v u2 u1

slide-6
SLIDE 6

Influence Maximization

6

Challenges to using IM in practice:

  • Challenge 2: Learning model parameters requires considerable data, often

unavailable to a new marketer.

  • Challenge 1: IM is not robust to the choice of the diffusion model [1] nor

to the model parameters [2].

slide-7
SLIDE 7

Online Influence Maximization

7

New marketer who has no past data to learn the reachability probabilities

Setting: Idea: Perform IM while simultaneously learning through trial and error across multiple

rounds.

Basic Protocol:

Diffusion occurs according to an underlying diffusion model.

  • Update parameter estimates
  • submodular optimization subroutine

probability estimates

Observe semi-bandit feedback.

  • size n binary vector. each entry = 1 iff that node is influenced by the seed u
slide-8
SLIDE 8

Online Influence Maximization

8

Challenge 1: Learn n

2 reachability probabilities

d dimensional feature describing a target node (Eigenbasis features, node2vec [3]) Vector to be learnt for every source node.

Advantages:

  • Reduces the number of parameters from O(n2) to O(dn).
  • In each round, mean estimates of can be updated by solving K regression problems.

[3] Grover, Aditya, et al. "node2vec: Scalable feature learning for networks." KDD, 2016.

slide-9
SLIDE 9

Online Influence Maximization

9

Basic Idea:

Use the Upper Confidence Bound algorithm i.e. use overestimate (mean + variance) of reachability probabilities as input to the oracle.

Computational Complexity: Per-round time:

+

  • racle computation

UCB computation

Challenge 2: Trade off exploration and exploitation

slide-10
SLIDE 10

Online Influence Maximization

10

Performance metric:

Regret after T rounds Optimal seed set in hindsight To account for the approximation in ORACLE Selected seed set

Regret Bound:

near optimal dependence Surrogate approximation factor best dependence on network size ORACLE approximation factor standard linear bandit dependence standard combinatorial bandit dependence

slide-11
SLIDE 11

Experiments on Facebook dataset

11

1000 2000 3000 4000 5000 2 4 6 8 10 12 14 16 18 x 10

5

Number of rounds Cumulative Regret

CUCB(10) TAB(10) I(10,50) L(10,50)

IC model

1000 2000 3000 4000 5000 2 4 6 8 10 12 14 x 10

5

Number of rounds Cumulative Regret

CUCB(10) TAB(10) I(10,50) L(10,50)

LT model

slide-12
SLIDE 12

Conclusion

12

Future Work:

  • Extend the framework to different feedback models and bandit algorithms.
  • Generalization across source nodes for better statistical efficiency.

Contributions:

  • Developed a model-independent parametrization for IM and proposed a

surrogate objective function.

  • Proposed and analyzed a UCB based algorithm for model-independent
  • nline IM.