Gradient Method for Continuous Influence Maximization with Budget-Saving Considerations
Continuous influence maximization (CIM) generalizes the original influence maximization by incorporating general marketing strategies: a marketing strategy mix is a vector $\boldsymbol x = (x_1,\dots,x_d)$ such that for each node $v$ in a social network, $v$ could be activated as a seed of diffusion...
Saved in:
Main Authors | , , |
---|---|
Format | Journal Article |
Language | English |
Published |
20.11.2019
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.1911.09100 |
Cover
Loading…
Summary: | Continuous influence maximization (CIM) generalizes the original influence
maximization by incorporating general marketing strategies: a marketing
strategy mix is a vector $\boldsymbol x = (x_1,\dots,x_d)$ such that for each
node $v$ in a social network, $v$ could be activated as a seed of diffusion
with probability $h_v(\boldsymbol x)$, where $h_v$ is a strategy activation
function satisfying DR-submodularity. CIM is the task of selecting a strategy
mix $\boldsymbol x$ with constraint $\sum_i x_i \le k$ where $k$ is a budget
constraint, such that the total number of activated nodes after the diffusion
process, called influence spread and denoted as $g(\boldsymbol x)$, is
maximized. In this paper, we extend CIM to consider budget saving, that is,
each strategy mix $\boldsymbol x$ has a cost $c(\boldsymbol x)$ where $c$ is a
convex cost function, we want to maximize the balanced sum $g(\boldsymbol x) +
\lambda(k - c(\boldsymbol x))$ where $\lambda$ is a balance parameter, subject
to the constraint of $c(\boldsymbol x) \le k$. We denote this problem as
CIM-BS. The objective function of CIM-BS is neither monotone, nor DR-submodular
or concave, and thus neither the greedy algorithm nor the standard result on
gradient method could be directly applied. Our key innovation is the
combination of the gradient method with reverse influence sampling to design
algorithms that solve CIM-BS: For the general case, we give an algorithm that
achieves $\left(\frac{1}{2}-\varepsilon\right)$-approximation, and for the case
of independent strategy activations, we present an algorithm that achieves
$\left(1-\frac{1}{e}-\varepsilon\right)$ approximation. |
---|---|
DOI: | 10.48550/arxiv.1911.09100 |