Professional Documents
Culture Documents
Knowledge Science, Engineering and Management
Knowledge Science, Engineering and Management
Knowledge Network
1 Introduction
Knowledge graphs, which belong to the field of knowledge engineering, are pro-
posed by Google in 2012. A knowledge graph is a multi-relational graph com-
posed of entities as nodes and relations as edges with different types. Knowledge
graphs are a kind of knowledge representation form, and they extract domain
entities, attributes and their relationships from a large amount of text data to
produce structured knowledge. The major advantage of knowledge graphs is the
ability to express knowledge of complex relationships accurately and graphi-
cally, which is in line with human learning habits and can help people learn key
knowledge and relationships more quickly. Therefore, knowledge graphs play an
c Springer Nature Switzerland AG 2019
C. Douligeris et al. (Eds.): KSEM 2019, LNAI 11775, pp. 628–639, 2019.
https://doi.org/10.1007/978-3-030-29551-6_56
Building Chinese Legal Hybrid Knowledge Network 629
2 Related Work
Our work is related to the work of knowledge graph construction, especially in
the legal domain.
With the rapid development of open knowledge graphs, researchers draw atten-
tion to knowledge graphs in specific domains, like in legal domain. Since knowl-
edge graphs need a certain data representation so that it can be used, Erwin Filtz
proposes a method to represent the legal data of Australia, mainly legal norms
Building Chinese Legal Hybrid Knowledge Network 631
1
https://baike.baidu.com.
2
http://www.baike.com.
3
https://zh.wikipedia.org.
632 S. Bi et al.
Fig. 1. The framework shows the general steps to develop the Chinese legal hybrid
knowledge network. The top part is a typical process to construct a network of legal
terms from Chinese encyclopedia, and the bottom part is the process to develop legal
knowledge graph from legal judgments online. The lower right part indicates the final
legal hybrid knowledge network.
Table 1. The statistical result of entity extracted from different Chinese encyclopedia.
between legal terms and common terms. What is worse, it is worthwhile to note
that legal knowledge is not original data from these raw texts but should be
extracted carefully from them. However, the entities extracted from Chinese
encyclopedia are not clearly classified and it is to manually classify them because
the size of the set of entities is too big. To solve this problem, we build a classifier
to find professional legal terms from unprofessional websites.
We get legal terms in encyclopedia information by category. If the category
of terms belongs to “legalese”, “jurist” or “laws and articles”, we consider these
terms as legal related. However, these data are still not enough. To enlarge the
number of legal terms, we take legal terms as seed and acquire terms that are
not related to law automatically. We use these data to train a classifier.
Intuitively, the closer to legal term with the internal link, the more likely
they are relevant. Assuming that the order of neighborhoods of legal terms is
smaller than four, these neighborhoods are legal related and may be legal terms.
In other words, the order of neighborhoods of legal entities is bigger than three,
these neighborhoods have no relation with legal entities and can be regarded
as negative samples. Therefore, we consider first-order, second-order and third-
order neighborhoods of legal entities as possible legal entities to be classified,
and other neighborhoods of legal entities as negative samples. Figure 3 indicates
the structure of different order entity. For example, as for legal entities in Baidu
Building Chinese Legal Hybrid Knowledge Network 633
Fig. 2. The example of building network of legal terms from Chinese encyclopedia.
Firstly, we obtain legal terms from three largest Chinese encyclopedia sites. Then we
calculate the similarities between these legal terms. Finally, these legal terms are linked
together by similarities to form a network of legal terms.
Fig. 3. We use seed as positive sample, and the neighborhoods which order bigger than
three as negative sample, the rest as potential legal domain related.
After obtained enough legal entities through our classifier, we use these enti-
ties to construct our network of legal terms. In our network of legal terms, each
entity is regarded as a point and entities are connected by the similarities between
these entities. We use SimRank to compute similarity. The weight between entity
a and entity b is computed as follows:
|I(a)| |I(b)|
C
s(a, b) = s(Ii (a), Ij (b)) (1)
|I(a)||I(b)| i=1 j=1
where s(a, b) is the similarity of point a and point b. Note that when a = b,
s(a, b) = 1. Ii (a) is the i − th in-neighbor of point a. When I(a) = ∅ or I(b) = ∅,
s(a, b) = 0. Parameter C is damped coefficient, and C ∈ (0, 1). The description
of this formula is that the similarity between a and b is equal to the average of
the similarities between in-neighbors of a and in-neighbors of b.
In this subsection, we show the details of building our legal knowledge graph.
According to the definition, a knowledge graph is a special graph where nodes
are entities and edges are relations. Knowledge graphs represent knowledge by
using RDF-style triples (h, r, t) which describe the relation r between the first
head entity h and the second tail entity t. For instance, “Beijing is the capital of
China” can be represented as (Beijing, capitalOf, China). Therefore, we should
extract RDF-style triples from legal judgments to build a legal knowledge graph.
Considering the specific standard format of legal judgments, we extract triples
based on several simple manual rules, which are listed in Table 2. Although the
rule-based method extracts some necessary information like the plaintiff and
defendant, there is a lot of information that is too complex to be extracted by
rules simply.
To tackle this problem, we adopt other named entity recognition (NER)
method to define entities hidden in sentences. NER is a subtask of information
extraction that seeks to locate and classify named entities in text into pre-defined
categories such as the names of persons, organizations, locations, expressions of
times, quantities, monetary values, and percentages. Currently, the most pop-
ular method of NER is using Conditional Random Field (CRF) [7] and Long
Short-Term Memory (LSTM), which is used in this paper. Moreover, we obtain
keywords and abstracts of legal judgments by the means of TextRank [10] to
enrich our legal knowledge graph. TextRank is a graph-based ranking model for
text processing which can be used to find the most relevant sentences in text and
keywords as well. Figure 4 indicates the triple extraction and knowledge graph
building process.
Building Chinese Legal Hybrid Knowledge Network 635
Fig. 4. The example of extracting triples and building knowledge graph from legal
judgments. The left part is a part of legal judgment and the right part is a knowledge
graph consisted of triples extracted from the legal judgment. Different relations are
labelled in different colors. (Color figure online)
4 Evaluation
In this section, we evaluated three algorithms used in constructing the Chinese
Legal Hybrid Knowledge Network.
Fig. 5. The example of the Chinese legal hybrid knowledge network. We link the legal
entities which are in legal judgment to network of legal terms.
built 5 classifiers totally and each classifier was trained with different features.
To be specific, these features were one hot, tf-idf [15], LDA [3], doc2vec [8],
word2vec [11] respectively. As mentioned above, in classification, we used SVM
to perform binary classification. The training dataset included 29706 positive
samples and 29706 negative samples. Note that 29706 negative samples were
randomly selected from all negative samples and we updated the whole negative
samples by deleting those selected samples each time. The performances of these
five classifiers with different features are presented in Fig. 6.
As is shown in Fig. 6, all of our classifiers have high accuracy, recall and F1
values, which verifies the effectiveness of our methodology.
where S1 was the entity set which was linked by our method, and T1 was the
labelled entity set. The results were shown in Table 3.
Precision Recall F1
86.21% 89.69% 87.91%
As was shown in Table 3, FEL had good performance in precision, recall and
F1 measure and was of great effect.
References
1. Auer, S., Bizer, C., Kobilarov, G., Lehmann, J., Cyganiak, R., Ives, Z.: DBpedia:
a nucleus for a web of open data. In: Aberer, K., et al. (eds.) ASWC/ISWC -2007.
LNCS, vol. 4825, pp. 722–735. Springer, Heidelberg (2007). https://doi.org/10.
1007/978-3-540-76298-0 52
2. Benjamins, V.R., Casanovas, P., Breuker, J., Gangemi, A.: Law and the Semantic
Web: Legal Ontologies, Methodologies, Legal Information Retrieval, Andapplica-
tions, vol. 3369. Springer, Heidelberg (2005). https://doi.org/10.1007/b106624
3. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn.
Res. 3(Jan), 993–1022 (2003)
4. Bollacker, K., Evans, C., Paritosh, P., Sturge, T., Taylor, J.: Freebase: a collabo-
ratively created graph database for structuring human knowledge. In: Proceedings
of the 2008 ACM SIGMOD International Conference on Management of Data, pp.
1247–1250. ACM (2008)
Building Chinese Legal Hybrid Knowledge Network 639
5. Do, P.K., Nguyen, H.T., Tran, C.X., Nguyen, M.T., Nguyen, M.L.: Legal ques-
tion answering using ranking SVM and deep convolutional neural network. arXiv
preprint arXiv:1703.05320 (2017)
6. Filtz, E.: Building and processing a knowledge-graph for legal data. In: Blomqvist,
E., Maynard, D., Gangemi, A., Hoekstra, R., Hitzler, P., Hartig, O. (eds.) ESWC
2017. LNCS, vol. 10250, pp. 184–194. Springer, Cham (2017). https://doi.org/10.
1007/978-3-319-58451-5 13
7. Lafferty, J., McCallum, A., Pereira, F.C.: Conditional random fields: probabilistic
models for segmenting and labeling sequence data (2001)
8. Le, Q., Mikolov, T.: Distributed representations of sentences and documents. In:
International Conference on Machine Learning, pp. 1188–1196 (2014)
9. Lenat, D.B.: CYC: a large-scale investment in knowledge infrastructure. Commun.
ACM 38(11), 33–38 (1995)
10. Mihalcea, R., Tarau, P.: TextRank: bringing order into text. In: Proceedings of the
2004 Conference on Empirical Methods in Natural Language Processing (2004)
11. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word repre-
sentations in vector space. arXiv preprint arXiv:1301.3781 (2013)
12. Miller, G.A.: WordNet: a lexical database for English. Commun. ACM 38(11),
39–41 (1995)
13. Montiel-Ponsoda, E., Gracia, J., Rodrı́guez-Doncel, V.: Building the legal knowl-
edge graph for smart compliance services in multilingual Europe. In: CEUR work-
shop proceedings No. ART-2018-105821 (2018)
14. Niu, X., Sun, X., Wang, H., Rong, S., Qi, G., Yu, Y.: Zhishi.me - weaving chinese
linking open data. In: Aroyo, L., et al. (eds.) ISWC 2011. LNCS, vol. 7032, pp.
205–220. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25093-
4 14
15. Ramos, J., et al.: Using TF-IDF to determine word relevance in document queries.
In: Proceedings of the First Instructional Conference on Machine Learning, Pis-
cataway, NJ, vol. 242, pp. 133–142 (2003)
16. Sánchez A, V.D.: Advanced support vector machines and kernel methods. Neuro-
computing 55(1–2), 5–20 (2003)
17. Suchanek, F.M., Kasneci, G., Weikum, G.: YAGO: a core of semantic knowledge.
In: Proceedings of the 16th International Conference on World Wide Web, pp.
697–706. ACM (2007)
18. Wang, Z., et al.: XLore: a large-scale English-Chinese bilingual knowledge graph.
In: International semantic web conference (Posters & Demos), vol. 1035, pp. 121–
124 (2013)
19. Xu, B., et al.: CN-DBpedia: a never-ending Chinese knowledge extraction system.
In: Benferhat, S., Tabia, K., Ali, M. (eds.) IEA/AIE 2017. LNCS (LNAI), vol.
10351, pp. 428–438. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-
60045-1 44
Machine Learning