Paper title:
Paper theme:
This paper studies a keyword-aware influential community query (KICQ) to find the communities with the highest influence in the network. The authors argue that existing top-r k-influential communities approach in [12] has the limitation that an appropriate k (k-core) is hard to choose for a given keyword query, and setting the influential score of a community to be the minimum weight among all members cannot accurately represent the influence of the communities. To overcome this, the authors measure the influential score of communities by combing the cohesion factor k (the cohesion factor of a connected k-core is k) and the sum of influential weight of individuals in the community. Specifically, they study two kinds of the relation types of the keywords ”And” and “OR”. To solve this problem, they first propose a baseline algorithm to find the top-r influential communities from the whole network. Then they further devise a bi-level indexing method to partition the whole graph and then find and merge the communities of the partitioned graph to obtain the final communities. The experimental result shows that bi-level indexing approach outperforms the baseline algorithm, but the improvement is only about several times which is not significant. Case studies to compare the proposed method and existing work [12] are also not provided.

Strong points:
S1: This problem is well motivated and a new scoring function is provided to find the influential communities.

S2: A baseline algorithm and optimization techniques are proposed with detailed examples to clearly illustrate the process of the algorithms clearly.

S3: A wide scope of the related worked are introduced and position of the proposed method is very clear.

Weak Points:
W1: Some aspects about the measuring score and proposed method are not very well explained which it not very convincible. I will show more in the detailed comments.

W2: The experiments are not adequate. Only one real dataset DBLP are tested and it is unclear how the proposed methods perform on other widely used datasets in previous methods. Case studies on real datasets are also lacked to show the difference between the proposed methods with existing work [12].

W3. The figures (e.g., Figures 8, 10, 11) for comparing the performance of bi-level indexing method and baseline method are not plotted in log-scale, which make it very hard for readers to see how many magnitudes faster of the improved algorithm.

Detailed comments:

  1. The underlying reason of defining the influence factor of the as IF(k) =ln(1+k) is not given. It is also unknown how different will the result be if we simply set IF(k) as k.

  2. In Definition 2, for AND query, the influential score of a node is defined as the minimum score of the associated keywords. But for the OR query, the influential score is defined as the average value. Why not define it as the maximum score?

  3. The authors propose Heuristics 1 for pruning the unpromising candidates, but it does not have any correctness guarantee. Moreover, how the parameter affect the final performance is not evaluated in the experiments.

There are also a number of error expressions in the paper. I listed several below.

  1. Page 3, bottom of the left column, $$V{H} \in V –> V{H} \subseteq V$$.
  2. Page 4, right column,$$ V{H} \in V –> V{H} \subseteq V$$.
  3. Page 7, left column, {n1, n2, n4}. for an AND query -> left column, {n1, n2, n4}. For an AND query.

<———————————–分割线——————————————–>

Paper title: <>
Paper theme:
This paper studies A dual graph contains a physical graph and a conceptual graph, both of which have the same vertex set. Given that there exist no previous studies on the k-core in dual graphs, the authors formulate a k-connected core (k-CCO) model in dual graphs. A k-CCO is a k-core in the conceptual graph, and also connected in the physical graph. Given a dual graph and an integer k, the authors propose a polynomial time algorithm for computing all k-CCOs. Thy also propose three algorithms for computing all maximum-connected cores (MCCO). The authors conduct extensive experiments on six real-world datasets and several synthetic datasets. The experimental results demonstrate the effectiveness and efficiency of threir proposed algorithms.

Strong points:
S1: This problem is well motivated and three algorithms are provided to find the maximum-connected cores(MCCO).

S2: The algorithms are understood clearly, and through experimental comparison, compared with the original algorithm DCS have been greatly improved which is significant.

Weak Points:
W1: There are some grammatical mistakes in the article, I will show more in the detailed comments.

W2: The physical graph and conceptual graph are not very well explained which it is not very convincible.

Detailed comments:
1、Page 5、6、7, there are six words wrong, the word definition less write a letter i.
2、In Algorithm 4, the name of the algorithms is wrong, should named TD-MCCO。And in line 4, it just is operation and instead of operation or.
3、Fig.4 (c) and Fig.4 (f) seem to be the same picture,but Epinions and Gowalla datasets are very big difference in graph size.

<———————————–分割线——————————————–>

Paper title: <>
Paper theme:
In this paper , the authors deal with the overlapping community detection problem. They propose a new algorithm, named Fast Community Detection based on Neighborhood Influence(FCD-NI), which leverages the overlap between neighbors to accelerate the search process. Besides, they use Metric Backbone (a minimum subgraph that preserves the shortest paths information) in place of the original graph to run the algorithms. The FCD-NI can find community not only on the undirected, unweighted graph, it also can be extended to the directed, weighted graph. The authors perform extensive experiments on both synthetic benchmark and a variety of real world networks to evaluate the effectiveness and efficiency of the proposed methods. Compared with some state-of-the-art community detection techniques, FCD-NI outperforms in running time and accuracy. In addition, utilizing Metric Backbone reduce the graph size dramatically. For some large graphs, it save up to 50% storage.

Strong points:
S1: This problem is well motivated and the idea of the algorithm is illustrated clearly.In addition, the authors utilize Metric Backbone reduce the graph size dramatically.

S2: The experiments are adequate. it is clear that the proposed methods perform on other widely used datasets in previous methods have been greatly improved which is significant.

Weak Points:
W1: There are some grammatical mistakes in the article, I will show more in the detailed comments.

W2: Lacking of specific examples to introduce algorithms which it is not very convincible to understand the algorithms.

Detailed comments:

  1. Page 6, in section 4.1, 6 column, that sentence should be that $$\l \omega$$

There are also a number of error expressions in the paper. I listed several below.

  1. Page 11, line 6-7 in the algorithm 4, remove(x, v) –> remove(\x, \upsilon ) and remove(v, x) –> remove(\upsilon, \x)

<———————————–分割线——————————————–>

Paper title: A Feature Derived from Communities

Paper theme:
In this paper, the authors map some communities into vector space as community feature by cluster affiliation model. Firstly, the authors generate vertex-community matrix from network based on Poisson distribution. Secondly, they generate community feature vertex-community matrix. Finally, they conduct experiments about user relation prediction problem on microblogging data. the precision and AUC core are improved up to 7.2% and 5.7%.

Strong points:
S1: The concerned problem is well motivated and the various parts of the algorithm is illustrated clearly.

Weak points:
W1: There are some grammatical mistakes in the article, I will show more in the detailed comments.

W2: Related works section maybe write a little less, and I think the algorithms are easy, most use current machine learning algorithms by minor improvement.

W3: Fig.1 is the main framework, it have three results but in the experiments have two models’ results.

Detailed comments:

  1. Fig.1, the picture is a bit rough.
  2. Algorithm 1, some concepts of variable are fuzzy.
  3. Fig.4, one variable is wrong.
  4. Table 2, word ‘refers’ should be ‘refer’, and word ‘algorithmk’ should be ‘algorithm’

20181222

Paper title: Efficient Search of the Most Cohesive Co-Located Community in Attributed Networks

Paper theme:
In this paper, the authors study the problem of searching the most chesive colocated community, which returns a community of structural cohesiveness and spatial co-location. Firstly, the authors propose an index structure called DkQ-tree to integrate the spatial information and the local structure information together to accelerate the query processing. Then, they develop two efficient algorithms based on this index structure. The extensive experiments conducted on both real and synthetic datasets demonstrate the efficiency and effectiveness of the
proposed methods.

Strong points:
S1: The experiments are adequate. it is clear that the proposed methods perform on other widely used datasets in baseline methods have been greatly improved which is significant.

Weak points:
W1: There are some grammatical mistakes in the article, I will show more in the detailed comments.
W2: The authors propose two algorithms, but they only given one algorithm code. And the authors didn’t give time complexity and space complexity of the algorithms.
W3: The paper is short of the section of conclusions.
W4: The name of two algorithms which the authors propose are same, that makes confusing for us.

Detailed comments:

  1. Table 1, there is no need to do.
  2. Fig. 2, node G should be in right grid which includes node H and I.
  3. subtutle 6.1 and 6.2, word ‘lg’ is inconsistent with the word ‘LG’ of comment.
  4. Algorithm 1, in line 1 and 7, word ‘d’ should be ‘D’. And in line 13, symbol ‘&’ should be ‘&&’.
  5. Fig. 5, both node D and G are the boundary vertex, but they are different in show.
  6. Parameters and Query generation of section 7.1, in line 4, parameter ‘s’ is unclear.
  7. section 7.2 and 7.3, the structure is wrong.

20190311