Approximation Algorithms for the Submodular Hitting Set Problem

Shaojing Du1, Bo Hou1, Wen Liu1
1School of Mathematical Sciences, Hebei Normal University, Shijiazhuang, 050024, P. R. China.

Abstract

In this paper, we study the submodular hitting set problem (SHSP), which is a variant of the hitting set problem. In the SHSP, we are given a supergraph H=(V,C) and a nonnegative submodular function on the set 2V. The objective is to determine a vertex subset to cover all hyperedges such that the cost of submodular covering is minimized. Our main work is to present a rounding algorithm and a primal-dual algorithm respectively for the SHSP and prove that they both have the approximation ratio k, where k is the maximum number of vertices in all hyperedges.

Keywords: submodular, hitting set, rounding, primal-dual, approximation algorithm

1. Introduction

The hitting set problem involves a supergraph H=(V,C), where V is a finite set with elements called vertices and C is a family of subsets of V called hyperedges. Let k be the maximum size of all hyperedges and Δ be the maximum degree of all vertices. A hitting set of H, also called a vertex cover of H, is a subset XV such that each hyperedge in C contains at least one vertex from X. The hitting set problem is to select a vertex subset of minimum size. For each vertex uV, if there is an associated nonnegative weight cost w(u), we obtain the weighted version of the problem, the goal of which is to select a vertex subset of minimum covering cost. The minimum hitting set problem is NP-hard [] and it is one of the fundamental combinatorial optimization problems. In this case, it is conventional that for any subset X of V the covering cost of X is the sum of the covering cost of all vertices in X, that is, w(X)=uXw(u). Then this function w() induces a linear function from 2V to R+.

Most results of decision problems in combinatorial optimization, such as trees or cuts, vertex cover in graphs, facility location, set cover, are in general asking for the linear costs. However, nonlinear costs are more realistic in many settings and there have existed some works in the area of combinatorial optimization problems with submodular costs. Assume E is a nonempty set and 2E is the power set of E. Let F():2ER+ be a nonnegative-valued set function. We call F() submodular if for any X,Y2E,F(XY)+F(XY)F(X)+F(Y) and F() modular if the equality holds. It is obvious that all modular functions are submodular and so all linear functions. We call F() monotone if F(X)F(Y) for each pair of subsets X,Y2E with XY. Submodular functions occur very often in a variety of subjects such as economics, engineering, computer science and management science. There is a greedy approximation which achieves an approximation ratio of H(Δ) for the hitting set problem with linear costs, where H(Δ) is the Δ-th harmonic number [1]. Moreover, using the greedy approximation, Wan et al. [2] gave some similar results for the hitting set problem with submodular costs.

One of the most fundamental problems related to submodular functions is the submodular function minimization problem. In this problem, we are given a finite set E and a submodular function F() defined on 2E, then the goal is to select a subset X of E minimizing F(X) among the power set of E, i.e., to select a minimizer of submodular function F() over 2E. It is known that this problem can be solved in polynomial time. Recently, constrained variants of the submodular function minimization problem have been extensively studied in various fields ([3-6]). For instance, Hochbaum [4] studied the submodular minimization problem with linear constraints having at most two variables per inequality. Kamiyama [5] studied the the submodular function minimization problem with covering type linear constraints. Koufogiannakis and Young [3] studied the monotone submodular function submodular minimization problem with general covering constrains. Iwata and Nagano [6] studied the submodular function minimization problem with set cover constraints. In this paper, we are motivated to study the submodular hitting set problem (SHSP). That is, given a nonnegative submodular function f():2VR+, the goal of the SHSP is to select a vertex subset XV to cover all the hyperedges that minimizes the submodular cost f(X). We present a rounding algorithm and a primal-dual program for the SHSP and prove that they both have the approximation ratio k.

The paper is organized as follows. In Section 2, we give some definitions about submodular function. In Section 3, we present a rounding algorithm and a primal-dual program respectively for the SHSP and prove that they both have the approximation ratio k, where k is the maximum number of vertices in all hyperedges.

2. Lovász Extension of a Submodular Function

In this section, we introduce the Lova´sz extension of a submodular function. Let F():2ER+ be a submodular function.

Assume |E|=n. Then for any vector zzRn and any non-empty subset XE, we define zz(X)=iXzz(i), where zz(i) is the i-th coordinate value of vector zz. It is easy to see that zz() is a modular function from 2ER.

For any XE, define χχX{0,1}n as follows: χχX(i)={1,if iX,0,otherwise, called the characteristic vector of X.

Given a submodular function F(), denote P(F)={zzzzRn,zz(X)F(X),XE}. We call P(F) a submodular polyhedron and its elements subbases. For any subbase zzP(F), a subset X of E is called zz-tight whenever zz(X)=F(X). There is a property for the collection of all zz-tight subsets.

Proposotion 1. <stron For any given subbase zzP(F), the collection of all zz-tight subsets is closed under union and intersection.

Proof. Since F() is submodular and zz() is modular, the result holds. ◻

Now we introduce the Lova´sz extension of a submodular function and its properties. Given a vector ppR+n and suppose p1,p2,,pl are all distinct coordinate values of pp with p1>p2>>pl. For each j=1,2,,l, denote Nj={ipp(i)pj}. Then we define a vector function over R+n as follows: F^(pp)=j=1l(pjpj+1)F(Nj), where pl+1=0. The function F^() is said to be the Lova´sz extension of F(). It is easy to see that for any subset X of E, F^(χχX)=F(X) and F^(00)=F().

It is known from [7] that (1)F^(pp)=max{i=1npp(i)zz(i)zzP(F)}. The problem (1) is a linear optimization over the submodular polyhedron, which is efficiently solved by Edmonds [8]. Combining (1) with the primal-dual theory, we get that F^(pp) is equal to the optimal value of the following program: (2)minXEF(X)ζ(X)s.t.XEiXζ(X)=pp(i),iE,ζ(X)0,XE.

There are some properties of the Lova´sz extension function and we list them here for later use.

Lemma 1. ([7]) A set function F() is submodular if and only if F^() is convex.

Lemma 2. [9]

Assume F^() is the Lova´sz extension function of submodular function F(), then we have

  1. (positively homogeneous) for any real number α>0 and any vector vv00, F^(αvv)=αF^(vv);

  2. the function F^() is a monotonically increasing convex function.

3. Approximation Algorithms

In this section, we construct a rounding algorithm and a primal-dual algorithm for the SHSP.

For the given supergraph H=(V,C), assume |V|=n, C={T1,T2,,Tm} and V=i=1m{Ti}. Denote k=max{|Ti|TiC}. Assume there is a submodular function f():2VR+ with f()=0, where f(X) means the covering cost of X for any subset X of V.

Firstly we define a binary variable ζ(X) for any subset XV as follows: ζ(X)={1,if X is selected to cover some hyperedges,0,otherwise. Therefore, the integer linear program for the SHSP is

(3)minXVf(X)ζ(X)s.t. XVXTiζ(X)1,TiC,ζ(X){0,1},XV.

In the above program, the first constraint ensures that each hyperedge TiC is covered by some vertex subset X. Since f() is a submodular function, there must exist a unique subset X of V such that ζ(X)=1 in the optimal solution of (3).

Recall that f^(χχX)=f(X), the convex programming for the SHSP is: (4)min f^(xx)s.t. uTixx(u)1,TiC,xx(u){0,1},uV. The relaxation of (4) is: (5)min f^(xx)s.t. uTixx(u)1,TiC,xx(u)0,uV.

3.1. A Rounding Algorithm for the SHSP

For a submodular function f(), we define a set function f():2VR+ as f(X)=min{f(Z)XZV},XV. Then f() is monotone and submodular [10]. Moreover, this function f^() has the following properties.

Lemma 3. ([6])

  1. suppose X is the exactly minimal subset Z satisfying XZV and f(Z)=f(X). Then f(X)=f(X) and f(X)f(X) for every subset XV;

  2. for each xxR+n, f^(xx)f^(xx);

  3. f^() is monotonically increasing and positively homogeneous.

Since the program (5) can be solved in polynomial time via the ellipsoid method, we denote xxR+n an optimal solution of (5). Based on the program (5), we give the following rounding algorithm.

]

Step 0: Solve the program (5) and denote xxR+n as the optimal solution.
Step 1: Define the subset AV by A={uVxx(u)1k}. Step 2: Define A be the exactly minimal subset of V such that AA and f(A)=f(A).
Step 3: Return A.

Recall that k=max{|Ti|TiC}. For every hyperedge TiC, since uTixx(u)1, at least one of xx(u) must be greater than 1k, which ensures A is a hitting set. Since AA, A is also a hitting set. Moreover, the hitting set A can be obtained in polynomial time by submodular function minimization theory. Therefore, Algorithm 1 can be implemented in polynomial time and the obtained A is a hitting set. Finally, we analyze the approximation ratio of Algorithm 1.

Theorem 1. With reference to f(),f() and A defined above, then we have f(A)kf(M) for any hitting set M, where k=max{|Ti|TiC}.

Proof. Since xxR+n is an optimal solution of (5) and f^(χχM)=f(M) for any hitting set M, we find f^(xx)f^(χχM)=f(M). Lemma 3(2) guarantees f^(xx)f^(xx) and hence (6)f^(xx)f(M). The fact χχAkxx and Lemma 3(3) ensure that f^(χχA)f^(kxx)=kf^(xx). Note that f^(χχA)=f(A)=f(A). Then we have (7)f(A)kf^(xx). Combining (6) with (7), we obtain f(A)kf(M) as desired. ◻

3.2. A Primal-dual Algorithm

In this subsection, we give a primal-dual algorithm for the SHSP. For any subset XV, we denote SX={TiCTiX} and we call X is a hitting set if SX=C.

Recall that for any given vector ppR+n, f^(pp) is equal to the optimal value of (2). Combining (2) with (5), we consider the following relaxation program for the SHSP : (8)minXVf(X)ζ(X)s.t.uTixx(u)1,TiC,XVuXζ(X)=xx(u),uV,ζ(X)0,XV. We do not need to add the constraint that xx(u)0 for the second constraint of (8) ensures it for all uV. And the dual program for (8) is: (9)maxTiCy(Ti)s.t.TiCuTiy(Ti)=zz(u),uV,zzP(f),y(Ti)0,TiC. Before proposing the primal-dual algorithm for the SHSP, we give some necessary facts.

Given a subbase zzP(f). Let (fzz)():2VR+ be defined by (fzz)(X)=f(X)zz(X),XV. Then (fzz)() is a submodular function. Observe that the definition of (fzz)() is related to the modularity of zz(). So we get the function (fzz)() for any subbase zzP(f) and there is a proposition for the function.

Proposotion 2. Given a subbase zzP(f), there exists the unique maximal subset X of V for the function (fzz)() such that f(X)=zz(X).

Proof. Note that zzP(f), it is clear that (fzz)(X)0,XV. Moreover, we get that minXV(fzz)(X)=(fzz)()=0. This implies that the existence of X. According to Proposition 1, we see that for every pair of minimizers X,Y of function (fzz)(), XY is also a minimizer. Thus, there exists the unique maximal subset XV for the function (fzz)() such that f(X)=zz(X). ◻

Lemma 4. Let zzP(f) and A be the unique maximal subset of V satisfying f(A)=zz(A). Assume CSA and a hyperedge TjCSA. If we define α=minXVχχTj(X)0f(X)zz(X)χχTj(X) and zz=zz+αχχTj, then we have

  • zzP(f);

  • furthermore, we define A as the maximal subset of V such that f(A)=zz(A). Then we have AA.

Proof. To prove zzP(f), it suffices to prove zz(X)f(X) for every subset X of V. The assumption zzP(f) implies that zz(X)f(X),XV. On the one hand, for every subset X of V such that χχTj(X)=0, we have zz(X)=zz(X)+α0=zz(X)f(X). On the other hand, for every subset X of V such that χχTj(X)0, we have zz(X)=zz(X)+αχχTj(X)zz(X)+f(X)zz(X)χχTj(X)χχTj(X)=f(X), where the inequality follows from the definition of α. The result follows.

Next we prove AA. The assumption TjCSA indicates ATj=. The definition of zz implies zz(u)=zz(u) for every element uA. Moreover, we get zz(A)=zz(A)=f(A). The definition of A implies AA. Suppose Z is a subset of V which satisfies both χχTj(Z)0 and α=f(Z)zz(Z)χχTj(Z). Then we have f(Z)=zz(Z). The definition of A implies ZA. Since χχTj(Z)0, we have TjZ0. Combining ATj= with TjZ0, we obtain AZ. Since AZ,AA and ZA, it is immediate that AA. ◻

Based on the dual program (9), we give the following primal-dual algorithm.

Algorithm 2(H). Step 0: Set t=1. Initialize y1=0,zz1=0. Define A1 as the unique maximal subset of V such that f(A1)=zz1(A1).
Step 1: If CSAt=, go to step 2. If not, select a hyperedge TjCSAt.

Define the real number αt by αt=minXVχχTj(X)0f(X)zzt(X)χχTj(X) Define yt+1 by yt+1(Ti)={yt(Ti)+αt,if Ti=Tj ,yt(Ti),otherwise and let zzt+1=zzt+αtχχTj.Define At+1 to be the unique maximal subset of V such that zzt+1(At+1)=f(At+1).
Set t=t+1.

Step 2: Let T=t and A=AT. Return A.

To obtain the ratio of Algorithm 2, we first prove the following lemmas.

Lemma 5. Algorithm 2 can be implemented in polynomial time.

Proof. Combining the definition of A with Lemma 4(2), we get that the algorithm iterates after at most |V|+1 times. That is, T is at most |V|+1. At every iteration, we can compute αt in polynomial time [11]. The Proposition 2 guarantees there exists the unique maximal At+1V such that zzt+1(At+1)=f(At+1) and we can compute it in polynomial time [7]. Therefore, Algorithm 2 can be implemented in polynomial time. ◻

Algorithm outputs {(y1,zz1),(y2,zz2),,(yT,zzT)}. Next we will prove it is a feasible region of the dual program (9).

Lemma 6. For each t{1,2,,T}, (yt,zzt) is a feasible solution of the dual program (9).

Proof. Need to show for each t{1,2,,T}, (yt,zzt) satisfies all constrains of (9). We prove this result by using induction on t.

Since (y1,zz1)=00 and f() is a nonnegative function, we obtain (y1,zz1) is a feasible solution of (9). Assume that this lemma holds when t=k. Suppose t=k+1 and TjCSAk. Firstly, according to Lemma 4 (1), it is obvious that zzk+1P(f). Combining zzkP(f) with the definition of αk, we have αk0, which implies yk+10. Next, we prove that for each uV, (yk+1,zzk+1) satisfies (10)TiC:uTiyk+1(Ti)=zzk+1(u). Since (yk,zzk) is a feasible solution of (9), we have (11)TiC:uTiyk(Ti)=zzk(u),uV. Note that V=Tj(VTj) and zzk+1=zzk+αkχχTj. For each uVTj, we obtain (12)zzk+1(u)=zzk(u) and (13)TiC:uTiyk+1(Ti)=TiC:uTiyk(Ti). From (11)-(13), it is immediate that (10) holds for each uVTj. For each uTj, we have zzk+1(u)zzk(u)=αk and TiC:uTiyk+1(Ti)TiC:uTiyk(Ti)=yk+1(Tj)yk(Tj)=αk, which implies that (10) holds for each uTj. Therefore, we have (10) for each uV. ◻

We now analyze the approximation ratio of Algorithm 2.

Theorem 2. Algorithm 2 is a k-approximation algorithm, where k=max{|Ti|TiC}.

Proof. Algorithm 2 returns a hitting set A and the final feasible solution (yT,zzT) for the dual program (9). Let OPT denote the optimal objective value for the SHSP. In what follows, we prove that f(A)kOPT. Based on Lemma 6 and the dual program (9), we get (14)f(A)=zzT(A)=uAzzT(u)=uATiCuTiyT(Ti). Using the definitions of k and A, it is immediate that (15)uATiCuTiyT(Ti)kTiCyT(Ti). The final feasible solution (yT,zzT) for (9) indicates that (16)TiCyT(Ti)OPT. From (14)-(16), we get f(A)kOPT. ◻

Acknowledgment

The authors would like to thank Professor Ding-Zhu Du for his many valuable advices during their study of approximation algorithm.

This work is supported by the NSF of China (No. 11971146), the NSF of Hebei Province (No. A2019205089, No. A2019205092), Hebei Province Foundation for Returnees (CL201714) and Overseas Expertise Introduction Program of Hebei Auspices (25305008).

References:

  1. Du, D., Ko, K.I. and Hu, X., 2012. Design and analysis of approximation algorithms (Vol. 62). New York: Springer.
  2. Wan, P.J., Du, D.Z., Pardalos, P. and Wu, W., 2010. Greedy approximations for minimum submodular cover with submodular cost. Computational Optimization and Applications, 45(2), pp.463-474.
  3. Koufogiannakis, C. and Young, N.E., 2013. Greedy γ-approximation algorithm for covering with arbitrary constraints and submodular cost. Algorithmica, 66, pp.113-152.
  4. Hochbaum, D.S., 2010. Submodular problems-approximations and algorithms. arXiv preprint arXiv:1010.1945.
  5. Kamiyama, N., 2018. A note on submodular function minimization with covering type linear constraints. Algorithmica, 80, pp.2957-2971.
  6. Iwata, S. and Nagano, K., 2009, October. Submodular function minimization under covering constraints. In 2009 50th Annual IEEE Symposium on Foundations of Computer Science (pp. 671-680). IEEE.
  7. Lovász, L., 1983. Submodular functions and convexity. Mathematical Programming The State of the Art: Bonn 1982, pp.235-257.
  8. Edmonds, J., 2003, January. Submodular functions, matroids, and certain polyhedra. In Combinatorial Optimization—Eureka, You Shrink! Papers Dedicated to Jack Edmonds 5th International Workshop Aussois, France, March 5–9, 2001 Revised Papers (pp. 11-26). Berlin, Heidelberg: Springer Berlin Heidelberg.
  9. Li, Y., Du, D., Xiu, N. and Xu, D., 2015. Improved approximation algorithms for the facility location problems with linear/submodular penalties. Algorithmica, 73, pp.460-482.
  10. Fujishige, S., 2005. Submodular functions and optimization. Elsevier.
  11. Murota, K., 2003. Discrete Convex Analysis, volume 10 of SIAM monographs on discrete mathematics and applications. SIAM.