Abstract
Traditional methods only consider topic information in English vocabulary information extraction, lose the statistical feature information of the keywords themselves, and easily ignore the semantic information of the words. In order to improve the extraction efficiency of English keyword information, based on the CAD mesh model, this paper adds constraint factors such as vertex neighborhood flatness, vertex degree, side length, and flatness on both sides of the side on the basis of the original QEM quadratic error simplification algorithm, and it incorporates a smoothing effect into the edge folding cost function. Moreover, based on the proposed normal vector-based QEM mesh simplification algorithm, the point selection after the edge folding operation is fixed as the vertices of the original edge, and it is applied to the mesh parameterization. In addition, the algorithm solves the local parameterization problem of partially deleted vertices after the simplification operation of each layer is completed. After the model is constructed, the performance of the model is verified through experiments. The research shows that the English keyword information extraction model constructed in this paper is effective.
1. Introduction
Keyword extraction technology is a key technology in the field of text information mining. This paper takes a single document as the research object and proposes an algorithm for extracting keywords from English text based on a complex network [1]. Based on complex networks, the algorithm uses natural sentences as window units, words as nodes, co-occurrence relationship between words as edges, and word co-occurrence degree C (wxwy) based on the co-occurrence frequency as the weight of the edges to construct a text word co-occurrence network model based on a complex network [2]. Moreover, this paper constructs the comprehensive eigenvalue formula of nodes based on the degree centrality, eigenvector centrality, and node betweenness centrality of network nodes. The network nodes are output in descending order of comprehensive feature values, single word nodes are removed, and the first K words are extracted as text keywords [3].
The main task of automatic keyword extraction is to extract words that can express the main idea, main content, and author’s opinion of the article from a specific article [4]. When we use a computer to extract keywords from a specific document, we can clearly understand the theme and main content of this document in a short time, so that we can obtain valuable information and improve our overall grasp of the content of this document [5]. The characteristic of English is that there is no obvious word boundary, and each word can appear independently. However, only by forming words with letters can we express specific meanings. If we want to get a single word, we need to perform English word segmentation first [6]. At the same time, texts in oriental traditional languages such as Japanese and Korean have the same characteristics. Word segmentation processing in these languages is also the key to intelligent text information technology. Currently, there are many word segmentation methods, and word segmentation algorithms are also different [7]. However, these methods are based on mechanical participle and comprehensive participle as the basic principles, and the final processing results are also different. English keywords are automatically extracted, and the process of English word segmentation cannot be avoided. At present, we still use English word segmentation technology for pre-processing [8]. However, English words are ever-changing, and different combinations of words and words have different meanings. This makes the processing of English word segmentation more complicated. Therefore, more manpower and material resources are needed to improve this work.
Keyword extraction algorithm based on statistics refers to an algorithm that uses the frequency of statistical words to perform related processing [9]. It judges whether a word is a keyword based on the statistical frequency of a word. Words and characters form a word. If a few consecutive words appear more often in the same statistical time period, they are likely to be a word. Therefore, the relationship between characters, such as frequency of occurrence, and the combination probability reflect whether they become a word [10].
2. Related Work
In order to further improve the effect and quality of keyword extraction, many scholars have improved the above algorithm. The literature proposed a text key phrase extraction method based on LDA and TextRank [11]. The literature uses a method that combines word position and word span to improve TFIDF weight, or uses word frequency and word position features to perform weighted analysis by using semantic coherence [12]. In addition, some scholars introduce the method of information entropy. However, these methods have certain problems in the application process, such as high computational complexity, and they need a considerable scale in terms of article type and corpus scale. There are also some scholars who combine article comprehensive information and quote news category factors, combined with other characteristic information for weighting, which can solve the problem of word frequency dependence to a certain extent [13]. However, these methods do not consider the impact of keywords’ part of speech and keyword coverage. The literature proposed a method of keyword extraction based on graphs [14]. The method considers the context of the word, the position of the word, the centrality of the word, the part of speech, and other characteristics, and the initial weight of the word is modified. Meanwhile, it has obtained a good extraction effect. Keyword extraction methods based on relationship frequency have been widely used in foreign academic circles. Due to the characteristics of this method in the generation of horizontal networks, most applications eliminate the relationships that do not meet the threshold requirements and their corresponding keywords according to the set frequency (weight) threshold, and then expand the analysis based on the generated horizontal network [15].
3. The Theoretical Basis of Graph Theory of Plane Graph
If a graph G can be represented in a two-dimensional space, and the edges on the two-dimensional space do not intersect except for the common endpoints, the graph is a plane graph (plane embedding), denoted by . Graph G is called a planable graph. The floor plan divides the two-dimensional space into two regions, a bounded domain and an unbounded domain, and the boundary of the two domains (Jordan curve). As shown in Figure 1, Figure 1(a) is a plane graph, and Figure 1(b) is a plan embedding graph.

(a)

(b)
If the graph is embedded in the plane to obtain , then the graph G is called the original graph of the plane graph . For mapping (immersion) , , and are regarded as interior points. The topological mapping (immersion) of is an edge (straight line) of . The boundary topology of the original image G is mapped to a Jordan line segment.
For any , and simply intersect with individual common points. If and are not adjacent, they are recorded as .(1) is a bijection (one-to-one mapping);(2), .
4. Theoretical Basis of Differential Geometry
According to the basis of differential geometry, the mapping from the plane domain to is
It satisfies(1)Each component function is infinite order continuous differentiable;(2)Vectorare linearly independent, that iswhere r is called a surface (patches) of and is called the parameters (coordinates) of the surface.
We assume that there is a surface , its parameter representations is , and the tangent vectors and the tangent vector form a tangent plane at any point on the surface P. The arbitrary tangent vector on the surface P can be expressed as
The square of the length of the tangent vector is expressed as
Among them, we set is the length of the tangent vector , is the length of the tangent vector , is the cosine of the angle formed by the tangent vector and the tangent vector , and is the metric tensor.
We assume that the parametric forms of P surface and surface are and , respectively, and the first basic row of the two surfaces are, respectively,
Under the corresponding mapping,is obtained and the existence of mapping is a sufficient and necessary condition for conformal mapping:
Among them,
Furthermore, we assume that is the tangent vector of the P surface at point Q, and set a curve of P with point Q as the starting point, namely
Then, is the curve on the surface , and its tangent vector at is expressed as
The tangent vector depends on the mapping and the P surface tangent vector , so the mapping between the tangent planes of the two surfaces can be derived from the mapping between the surface P and the surface , which is also called the tangent mapping:
In addition to the intrinsic properties of the surface, the triangular mesh model also relies on some other properties, of which the measurement of deformation is the most important. For a better understanding, we assume that exists on the surface. When an infinitesimal movement occurs at a point in the parameter domain, the infinitesimal parameter displacement is defined as , and the first-order Taylor expansion on the surface corresponding to the new point is
The linear function maps all the vertices in the neighborhood to the tangent plane at the point , and stretches the circle centered at to an ellipse at the point p (as shown in Figure 2).

The expansion is rewritten as
Among them, is the Jacobian matrix of f, that is, the partial derivative matrix of f. Use singular value decomposition for the partial derivative matrix.
Among them, the singular value is , and the orthogonal matrix is , and . U, V decompose and column vectors respectively. After that, the linear function is decomposed (as shown in Figure 3):(1)Deformation first rotates all vertices around the center , so that the vector and the vector are located on the u v axis, and normalization is performed;(2)The deformation is stretched at the ratio of on the u v axis;(3)Deformation U finally maps the two vectors to the tangent plane at point p.

The circle with as the center and radius r is mapped to an ellipse with point p as the center and semi-axis lengths and , respectively, and orthogonal vectors , are mapped to orthogonal vectors , , respectively.
The deformation of the circle stretched to an ellipse is called the local metric deformation of the triangular mesh model, which shows how the function f affects the parameter points of the neighborhood and the corresponding surface point . All local metric deformation information is hidden between the singular values . When f is isometric mapping or length-preserving mapping, ; When f is conformal mapping, ; When f is the area-preserving mapping, .
5. Merge Super Face
The general idea of the merge super-surface algorithm is to merge the adjacent triangular surfaces of the original model to form the so-called “super-surface.” In this way, one hole after another is formed on the surface of the original model, and then these holes are re-triangulated to obtain a new approximate model output. Compared with the original model, the new approximate model has fewer vertices and triangles, which can simplify the mesh.
The so-called hypersurface is a curved surface area composed of multiple polygonal surfaces, and the boundary of the curved surface is surrounded by the vertices of these polygonal surfaces, and the boundary of each polygonal surface forms a large non-planar polygon.
The algorithm has the following characteristics:(1)Within the given tolerance range, the output simplified mesh model is similar to the original mesh model. That is, the Euclidean distance between the vertex in the original mesh and the corresponding vertex of the output simplified model does not exceed the value set by the user. The opposite is the same, that is, the distance between the vertex of the simplified model and the vertex corresponding to the original mesh does not exceed .(2)The algorithm is very efficient. It is very practical to simplify very large grid models, such as those from medical CT and MRI data.(3)The topology information of the original mesh model is maintained.(4)The algorithm is domain-independent, it does not need to extract the original data to perform simplified operations. The vertex set of the hyperplane boundary is a proper subset of the vertex set of the original mesh model, so this algorithm is particularly suitable for the hierarchical representation of the original polygon mesh model.
The three steps of the grid simplification algorithm are as follows:(1)Create a hypersurface: The patch merging step mainly divides the patch set of the original model into hypersurface regions.(2)Boundary straightening: The boundary of the hyperplane area is simplified by merging boundary sets.(3)Triangulation of the hypersurface: After the triangulation of the hypersurface, the new vertex set is determined. In this step, a single hypersurface region will be decomposed into multiple hypersurfaces, and each sub-hypersurface has its own boundary and new triangulated vertices.
Among them, there are several rules for merging triangles:(1)The rule of planarity: the distance from all vertices in the triangular face to the plane must not exceed . That is, for there are That is, a pair of linear constraints on :(2)Face axis rule: the direction of the triangular face must be the same as the plane . Among them, is the outward unit normal vector of the triangular face , and is the maximum angle between the two normal vectors.(3)No folding rule: the triangular face must not be folded on the super face.
Hoppe proposed a series of edge shrinking operation mesh model simplification algorithm, as shown in Figure 4. This method records a series of detailed information and intermediate processes of edge contraction. After a period of iterative operation, the multi-resolution model (LOD) formed to meet the needs of different applications, and the algorithm is reversible, that is, the simplified final coarse mesh can record information according to intermediate steps.

The original high-resolution mesh model is obtained by performing the opposite point splitting operation, namely
6. Model Building
This article uses the local coherence model to calculate local coherence based on the relationship between adjacent sentences. First, this paper uses dependency syntax analysis to obtain the syntactic components of the words in the sentence, builds a text graph model with sentences as nodes based on this, and then calculates the local coherence of the sentences in the text. Secondly, this article uses the text word co-occurrence network model based on the complex network in the text keyword extraction described above to obtain the comprehensive feature value of the network node to meet the basic information requirements of the text in the text summarization process. Third, in order to improve the semantic information of the text, a text LDA model is constructed in units of sentences. By combining the three to construct a comprehensive scoring function, the sentences in the text are scored. Finally, this article sets a threshold (this article is limited to 10%), and selects text sentences within the threshold range as the text summary. The method flow is shown in Figure 4.
There are many keyword extraction algorithms. Common keyword extraction algorithms can be summarized into three categories: semantic-based algorithms, machine learning-based algorithms, and statistical model-based algorithms. Semantic-based keyword extraction algorithms rely on background knowledge bases, dictionaries, vocabularies, etc. Therefore, when extracting keywords from texts with many new online words and transliterated words, words or phrases that are not included in the knowledge base cannot be extracted. The keyword extraction algorithm based on machine learning relies on the selected algorithm model, and training takes a long time. The keyword extraction algorithm based on the statistical model has a simple principle, does not require training samples, does not rely on the knowledge base, and can extract words or phrases not included in the knowledge base. Aiming at the advantages and disadvantages of the above algorithms, this paper proposes a keyword group or phrase extraction algorithm based on the length of the common substring and the frequency of occurrence in the text under the English text. This algorithm is a keyword extraction algorithm based on statistical models. First, it extracts words or phrases that appear frequently in the text. A piece of English text is composed of multiple sentences, and the words or phrases with higher frequency in a piece of text can be extracted by extracting the common substrings of the sentences that make up the text. This method can extract words or phrases that appear twice or more in the text, and then combine the length and frequency of the extracted words or phrases to filter out keywords. This algorithm improves accuracy and does not require training samples and building models and can extract words or phrases that are not included in the knowledge base.
The English text is composed of sentences, so the English text can be processed by clauses to obtain the set of clauses corresponding to the text. Sentences are composed of words or phrases. Words or phrases are composed of characters. There are no spaces between words or phrases and the length is variable. Therefore, each clause can be regarded as a string of characters, and the two or two clauses can be improved. The common substring extraction algorithm can extract words or phrases that appear in both clauses, and finally get words or phrases with a frequency greater than or equal to two in the entire text. The length of the word or phrase extracted by this method is not limited, and the new words and transliterated words existing in the text can be accurately extracted without the limitation of the dictionary.
In this section, the improved longest common substring algorithm is used to extract words or phrases with higher frequency from short English texts. The main idea is to segment the short text and use the improved longest common substring algorithm to extract the words or phrases that appear in both clauses based on the obtained clauses, and to finally use the string matching method to calculate the frequency of the extracted words or phrases in the text, and filter out the words or phrases with a frequency greater than the threshold. This method extracts words or phrases that appear frequently in the text. The longer a word or phrase in an English text is, the more it has practical meaning and the more it can represent the theme of the text. Therefore, according to the frequency of the word or phrase in the text and the length of the word or phrase, the keywords in the word or phrase can be filtered out. In summary, the keyword extraction in this chapter is implemented based on statistical models, without training samples, nor relying on the knowledge base.
In the cluster center point selection algorithm based on the similarity between texts, how to select the cluster center point is also the difficulty of the algorithm after the similarity between the texts is obtained. Common methods for selecting clustering center points based on the similarity between texts include: selecting texts with less mutual similarity as the clustering center points, and selecting texts that are related to more texts as the clustering center Point, and selecting texts with greater similarity with part of the text as the cluster center point. Selecting the texts with less mutual similarity as the cluster center text: when the cluster center text selected by this method is used for clustering, the similarity between the clusters obtained is small. The disadvantage is that the randomness of the central text usually selected is relatively strong, and there may be “outliers.” Selecting the text that has correlation with more texts as the cluster center text: When the cluster center text selected by this method is used for clustering, the clustering result will not appear local optimal, and the clustering result is optimized. The disadvantage is that multiple cluster center points may belong to the same category, resulting in a category being divided into multiple categories. Select the text with greater similarity with part of the text as the cluster center text: when the cluster center text selected by this method is used for clustering, the clustering result will not appear locally optimal, and the clustering result is better. The disadvantage is that multiple clustering center points may also belong to the same category, resulting in a category being divided into multiple categories.
In this paper, combined with actual R&D projects, a planning and design scheme of text platform software is proposed. The system architecture is shown in Figure 5.

Next, this article constructs an experiment to extract information from the English key words of the model. This article has set a total of 102 sets of texts. The keywords have been extracted by manual labeling. Therefore, the key word information can be extracted directly through the model proposed in this article, and the results can be compared with the manual labeling results. The model in this paper is named CAD, and the model in this paper is compared with the neural network model, and the neural network model is named NN. The keyword extraction results obtained on this basis are shown in Figure 6.

It can be seen from Figure 6 that the recognition rate of the model constructed in this paper can reach over 85%, and the highest can reach 95%, while the recognition rate of the neural network model is distributed between 65% and 85%. Next, in order to identify whether the 11 features proposed in the CAD model-based keyword extraction method can identify keywords, we conduct a feature analysis experiment. The experimental results of the feature analysis are shown in Table 1 and Figure 7.

From the above analysis results, we can see that the model constructed in this paper meets the expected requirements of the model constructed in this paper.
7. Conclusion
As one of the core technologies of natural language processing, keyword extraction from English text plays an important role in the application of natural language processing technology.
Mesh simplification and mesh parameterization are hot issues in the research of computer graphics. In recent years, rapid progress has been made in minimizing deformation energy and reducing the complexity of the algorithm itself. Based on the proposed normal vector-based QEM mesh simplification algorithm, this paper fixes the point selection after the edge folding operation to the vertex of the original edge, and uses it on the mesh parameterization. After the simplification operation of each layer of the grid is completed, the algorithm solves the local parameterization problem of partially deleted vertices, which avoids the complex calculations caused by solving large matrix systems. Successfully used in parametric-based heavy meshing and texture mapping applications. The experimental research results also prove that the model constructed in this paper has a certain effect.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The author has declared to have no competing interests.
Acknowledgments
This work was supported by Shenyang Urban Construction University.