Study on the detection methods of S&T frontier from the multi-dimensional perspective①
2022-04-07ZENGWenZHENGJiaWANGDaweiXIONGShulingZHANGLeiWEIXiaoqi
ZENG Wen(曾 文), ZHENG Jia, WANG Dawei, XIONG Shuling, ZHANG Lei, WEI Xiaoqi
(Institute of Scientific and Technical Information of China, Beijing 100038, P.R.China)
Abstract The development of network and information technology has brought changes to the information environment. The sources of information are becoming more diverse,and intelligence acquisition will be more complicated. The intelligence reflected by different dimensions of scientific and technological (S&T) data will have their own focuses. It has become inevitable to carry out the multi-dimensional research of S&T frontier, which is also a current research hotspot. This paper uses quantitative and qualitative research methods to conduct research and analysis of S&T frontier detection from three dimensions including S&T research projects,S&T papers and patents,and proposes related research methods and development tools. This work analyzes the S&T frontiers in the field of artificial intelligence and draws conclusions based on the analysis results of real and effective S&T data in three dimensions.
Key words: multi-dimensional, scientific and technological (S&T), frontier, artificial intelligence
0 Introduction
The detection of scientific and technological (S&T)frontiers is one of the main tasks of intelligence research to serve the national S&T strategic decisionmaking. Its purpose is to support the S&T strategic decision-making departments to study and predict S&T frontiers, grasp the general trend, seize opportunities,and make strategic planning and deployment of science and technology as soon as possible. The existing research methods basically use quantitative or qualitative research methods. The problems of quantitative research methods are mainly content analysis of data,which has strong objectivity. However, due to the single data type,the results of detection are prone to deviations. Forecasting research or activities such as technology foresight,research institutions and implementers adopts qualitative analysis method based on expert experience. Such methods are usually based on macropolicy data, and the data are limited. The limitations of the expert field background will inevitably lead to subjective biases of results. Based on this, the multidimensional data is used as a research object, and a combination of quantitative and qualitative research method is adopted to form a S&T frontier research method and system.
1 Related work
Through the investigation of S&T literature at home and abroad, it is found that the research on the S&T frontiers is mostly focused on the detection of research frontiers or technological frontier. The detection of frontier mostly uses S&T papers as the main source of intelligence[1], and the detection of technological fronts uses patents as the main source of intelligence.Most of the detection methods are based on quantitative or qualitative analysis methods. The detection of quantitative analysis methods focuses on the quantitative research and judgment of hotspot and high-level S&T frontier research or technology, and the detection of qualitative analysis methods focuses on subjective S&T frontier research or technology predict.
1.1 Review of quantitative analysis methods
(1)Methods based on citations: there are mainly methods based on co-citation[2-3], document coupling[4], direct citation[5-6]and the main path analysis method of citations[7-8]. The problem of above methods is the delay of citation time. For these methods it takes a certain time for papers to reach a certain citation frequency or high citation, so it is impossible to detect and identify emerging research hotspots and research trends.
(2)Methods based on topic words (keywords):there are mainly burst word monitoring methods[9-10],co-word method[11], and topic model based on latent Dirichlet allocation (LDA)[12]. These methods do not have the problem of the delay of citation time. They use natural language processing technology and topic statistical models to detect important research topics(keywords) from text data, but because words can express different meanings in different contexts, the expression meaning of a single word is not specific, and the users need to judge whether it is the frontier research of the field based on their own domain knowledge. Therefore, the interpretation of the detection results varies from person to person, that is, these methods have more subjective factors.
(3) Compound methods based on citation and subject terms: there are mainly methods of combining common words and citations[13], combining subject terms and citations[14], and combining full text and citations[15]. Although these methods make up for the respective shortcomings of the citation-based and topicbased methods, this method becomes complicated,cumbersome, and poor in practice when it is applied with a large amount of data.
(4)Method based on patents measurement: design and use patents evaluation indicators to determine technology frontiers in a certain field through citation analysis or cluster analysis[16]. Citation analysis has a certain delay in patent documents, and cluster analysis is affected by the number of clusters or cluster spacing,which makes the quality of clustering poor.
(5)Other methods: mainly include neural network model method[17], knowledge map method[18],essential science indicators (ESI) method[19], alternative metrology method[20], computer processing technology, and quantitative analysis method combined with the development of visualization tools. These research methods have improved the first four methods to a certain extent, but the effect is still limited.
1.2 Review of qualitative analysis methods
The realization of S&T frontier detection is supported by expert knowledge and cognition, and experts in related fields are organized by international organizations, governments, authoritative institutions or departments, S&T research management institutions, and S&T funding institutions to conduct on-site discussions, online discussions, communication surveys, interviews and other methods. It gathers the knowledge and wisdom of experts in different fields, and uses qualitative methods such as Delphi and forward-looking methods to form conclusions of prediction.
2 Main research methods and related tool development
2.1 Main research methods
This paper identifies three types of S&T data,i.e., S&T project, papers, and patents. Among them, S&T projects are the concrete implementation and concrete manifestation of a country’s strategic layout in science and technology frontier, and have a certain industry or social application demand orientation,and represent the future direction of frontier. The S&T data in the field of artificial intelligence are used in this paper. Judging from the hotspot of the global research field, artificial intelligence is the core driving force of the current global industrial intelligent development,and it is also the cause of a new round of technological change and industrial revolution. In terms of research methods, this paper uses a combination of quantitative analysis and qualitative judgment. The quantitative analysis method uses word frequency statistics and topic model method (LDA) to quantitatively analyze multidimensional S&T data. The quantitative analysis method is mainly through the fusion of term frequency-inverse document frequency (TF-IDF), LDA, and other algorithms, as shown in Fig.1. Functions, such as data splitting and data clustering, statistics and analysis of data in different dimensions, are used to analyze the development trend of S&T frontier research in the field of artificial intelligence with the help of expert opinions.
LDA is a topic generation model proposed in Ref.[12], which represents text based on the implicit semantics of text data, and is a semantic dimensionality reduction technology. The LDA method assumes that text data is composed of multiple topics, and each topic is composed of multiple words. The generation of a text is to select a topic with a certain probability, then selects a word under this topic with a certain probability and keep looping, finally gets a text. The use of LDA is a reverse process of generation. The process is to find the topic distribution of the text and the word distribution corresponding to the topic based on a text.The LDA topic model is a three-layer Bayesian network model, including a three-layer structure of text, topic,and words. The core idea is that each document corresponds to a topic vector that obeys the Dirichlet distribution, and each topic corresponds to a word that obeys the Dirichlet distribution.The model is shown in Fig.2.
Fig.1 The basic process of quantitative analysis method
Fig.2 Framework of LDA topic model
The explanation of Fig.2 is shown below.
Mis the total amount of text data.
Kis number of topics set.
Nis the number of words in a selected text.
θis the topic vector distribution of a text.
φis the topic distribution of each text, usually is set independently.
αis hyperparmeters of the prior Dirichlet distribution of the topic distribution of each paper, usually is set independently.
βis hyperparmeters of the prior Dirichlet distribution for each topic word distribution, usually is set independently.
Zrefers to the theme given by the word in the text, which is a hidden variable.
Wrefers to the word of the text.
For each document, the topic generation process is as follows.
(1) Randomly assign values to all data texts and topics.
(2) For then-th word in them-th text, randomly assign its topicZk,and calculate model probability.
(3) Try to enumerate all topicsZ, select a topic for then-th word in them-th document based on the results of these probabilities, use Gibbs Sampling formula to sample, and select the topicZk.
(4) If the topicZkchosen at this time is different from the randomly assigned at the beginning, it will have an impact onθandφ,θandφwill in turn affect the calculation of model probability.
(5) Repeat Steps (3) and (4) until Gibbs Sampling converges and ends.
(6) Output the finalθandφ.
2.2 Development of quantitative analysis software
The quantitative analysis software for data is developed by using the separation of front and back ends,and is mainly divided into three main functional modules, i. e. data processing and analysis module, data preprocessing and interface generation module,and data display module. In the data processing and analysis module, Python language is mainly used to develop and integrate statistical analysis algorithms such as TFIDF and LDA. This module mainly implements functions such as text splitting and data clustering, and provides the underlying data support for the entire software. In the data preprocessing and interface generation module, Java language is mainly used for research and development, which mainly realizes the functions of data preprocessing, data storage, and interface generation. The data display module adopts the currently popular VUE framework to realize the visual display of data. An example of the software structure of the tool is shown in Fig.3.
Data quantitative analysis software can currently realize the management and analysis of data such as S&T strategies, papers, patents, projects in the field of artificial intelligence and life sciences. It realizes the basic visual display function. Taking the field of artificial intelligence as an example, there are 4 submenus under the menu, i.e. S&T strategies, S&T papers, S&T patents, and S&T projects. The module interface in the field of artificial intelligence is shown in Fig.4 and Fig.5.
Fig.3 Example of software structure
Fig.5 Data display interface of sub-menu
The display interface of data analysis result is shown in Fig.6. In the display interface, relevant information after data analysis will be displayed. The analysis results of vocabulary in each type of data are arranged and displayed numerically according to the score automatically calculated by the software from high to low. In the display interface, there is a paging display function, and the users can choose to display a page or choose the number of data records displayed on each page.
Fig.6 Data display interface of analysis result
3 Research results
3.1 Frontier analysis of artificial intelligence based on the dimensions of papers and patents
In terms of S&T papers, this paper selects top international conference papers for analysis. In the field of artificial intelligence, in addition to academic journal papers, researchers also attach importance to top international conferences. These top conference papers can reflect hotspot research directions, the latest methods and technologies, and represent the latest advances and trends of research or technology in the field of artificial intelligence. Based on the above, through network investigation and experts consultation, this work selects the international journals of Engineering Village Complex Index Library, the Association for the Advance of Artificial Intelligence (AAAI), IEEE Conference on Computer Vision and Pattern Recognition(CVPR), IEEE International Conference on Computer Vision (ICCV), International Conference on Machine Learning (ICLM), and International Joint Conference on Artificial Intelligence (IJCAI) as data sources, and the actual span of the papers published is from 2011 to 2018. After data correlation calculation[21]and data cleaning, a total of 16 871 valid data are finally obtained.
In 2019, the World Intellectual Property Organization (WIPO), in order to understand the development of the field of artificial intelligence, defines artificial intelligence in three aspects including artificial intelligence technology, artificial intelligence application, artificial intelligence application field based on patents data. This work compares and analyzes results of the quantitative calculation about patents and papers, and obtains the following conclusions.
(1) The main frontier of artificial intelligence technology is machine learning, among which deep learning, latent representation, unsupervised learning,support vector machine, neural network, reinforcement learning, multi-task learning, etc. are the frontier hotspot technologies. In terms of artificial intelligence applications, the application direction of frontier is computer vision, among which biometric recognition,scene understanding, general computer vision, predictive reality,image and video segmentation,feature recognition, target tracking, etc. are frontier application directions and hotspots.
(2) The most common combinations in patents are deep learning and computer vision,computer vision and transportation, telecommunications and security,ontology engineering and natural language processing,life science research and machine learning.
(3) Machine learning is the main artificial intelligence technology, and more than one-third of all identified inventions are machine learning. The industries involved in the application fields are transportation,life sciences, personal equipment, computing and human-computer interaction (HCI), banking, security,industrial manufacturing, agriculture, and networking.
3.2 Frontiers of artificial intelligence in the dimension of S&T projects
This paper selects the open project database of 4 research institutions including the National Science Foundation (NSF), the European Commission, The Engineering and Physical Sciences Research Council(EPSRC) and the UK Research and Innovation Agency (UKRI) as the data source,and the search involves project data of artificial intelligence. The span of project selected from 2016 to 2019. After data correlation calculation[21]and data cleaning, a total of 2102 project data related to artificial intelligence is finally obtained. Through quantitative analysis, it is found that the frontiers of artificial intelligence with the most project investment are semantics, robotics, multi-modality, electronics, space technology, wireless technology,life and medical sciences, transportation, the Internet and other application fields.
3. 3 The results of quantitative and qualitative analysis on frontiers about artificial intelligence
(1) Image recognition based on deep neural network
Image recognition is a recent hotspot topic in the field of artificial intelligence. In particular, deep neural network algorithms based on convolutional neural networks and recurrent neural networks are very active in the field of image recognition. At present, the hotspot topics of this research mainly focus on the two aspects of human facial recognition and action recognition, by means of deep learning related algorithms to realize the calculation of human movement, posture and spatial features. From the perspective of technical application, related researches mainly focus on pedestrian re-recognition technology, image segmentation,image retrieval, visual tracking, and 3D perspective reconstruction.
(2) Natural language processing technology based on semantic understanding
Natural language processing is developing from the understanding of word meaning and sentence meaning to the direction of multiple semantics and knowledge discovery. Hot topics involve learning methods for natural language processing tasks, interpretable natural language processing based on semantic analysis, knowledge and common sense, context modeling, and multiple rounds of semantic understanding. From the perspective of specific application scenarios, the application of basic researches such as machine translation based on neural networks, natural communication between humans and machines using natural language,machine reading comprehension, information retrieval,knowledge question and answer, automatic abstracting,etc., are relatively research hotspots.
(3) Unsupervised learning algorithm
Machine learning methods have presented two major development directions, supervised learning and unsupervised learning. In terms of unsupervised learning, automatic classification of multi-dimensional data,automatic data indexing, and hidden structure of learning data are research hotspots. From the perspective of learning characteristics of data, there are two evolutionary characteristics of machine learning, the transformation of existing labeled data to unlabeled data,the transformation of synthetic data to real data. At present, because the accumulation of real basic data is far from meeting the needs of model training, synthetic data is the data basis for unsupervised machine learning. The field of semantic segmentation algorithms,fully convolutional adaptive networks and unsupervised adaptive semantic segmentation are frontier hotspots.
(4) Brain-like intelligence
Research on brain-like intelligence based on bionics has long been incorporated into the artificial intelligence development strategy of various countries for the key layout. From an overall point of view, global brain-like intelligence has two frontier areas, including brain-like computing and brain-like bionic control.Among them, the research of brain-like computing mostly starts from the two directions of neural network and neuron, and mostly starts from the perspective of brain mechanism and functional structure. Brain chips are the main frontier hotspot of brain-like computing in hardware. The brain-computer interface is also one of the frontier hotspots. The human-computer hybrid system is controlled through the brain-computer interface.Most of the frontier research in brain-like bionic control comes from the field of intelligent robots. Research on multi-modal perception and brain-like autonomous decision-making based on perceptual information is one of the frontier hotspots. In addition, the development of new materials, new transmission devices, new materials and new transmission systems have enabled a series of breakthroughs in muscle-like drivers.
4 Conclusion
From the perspective of information research methods, this paper combines quantitative analysis and qualitative analysis methods, extends the single-dimensional detection to multi-dimensional detection of S&T frontier, breaks through the limitations of the existing detection method framework, and applies it to frontier research in the field of artificial intelligence. Research results show that artificial intelligence technology has shifted from theoretical research to commercial product and service applications focuses on the field of machine learning and computer vision. What needs to be pointed out is that there is currently no unified standard or model for the reliability or validity of the detection results of S&T frontiers, and most of researches use expert evaluation methods. For the credibility of the detection results of the frontiers of S&T, the paper adopts the method combining quantitative analysis of data and expert opinions. In terms of obtaining multi-source data,the paper adopts the research method proposed in Ref.[21], and combined with the terms of artificial intelligence provided by domain experts. These information is used to filter or retrieve data to ensure the validity of the multi-dimensional data source and at the same time ensure the validity of the data analysis results. In the future, we will conduct further research on the evaluation methods of the S&T frontiers.
杂志排行
High Technology Letters的其它文章
- Deep convolutional adversarial graph autoencoder using positive pointwise mutual information for graph embedding①
- A joint optimization scheme of resource allocation in downlink NOMA with statistical channel state information①
- Personalized movie recommendation method based on ensemble learning①
- An improved micro-expression recognition algorithm of 3D convolutional neural network①
- Resonance analysis of single DOF parameter-varying system of magnetic-liquid double suspension bearing①
- Computation offloading and resource allocation for UAV-assisted IoT based on blockchain and mobile edge computing①