Loading...

Table of Content

    14 May 2019, Volume 0 Issue 05
    An Android Malicious Code Detection Mechanism Based on Native Layer
    SUN Bing-lin, ZHUANG Yi
    2019, 0(05):  1.  doi:10.3969/j.issn.1006-2475.2019.05.001
    Asbtract ( 251 )   PDF (1186KB) ( 114 )  
    References | Related Articles | Metrics
    Android’s existing malicious code detection mechanism is mainly for the bytecode layer codes. This means that malicious code embedded in Native layer can’t be detected. The latest research shows that 86% of popular Android APPs contain Native layer code. In order to solve this problem, this paper proposes an Android malicious code detection mechanism based on Native layer, which converts smali code and so file into assembly code, generates control flow graph then optimizes it. Through comparing with malware library by subgraph isomorphism method, the similarity values are calculated and compared with the given thresholds to determine whether the software under test contains malicious code. The experimental results show that compared with the others the method can detect malicious code of Native layer and has higher accuracy and detection rate.
    File Scheduling Algorithm Based on Magneto-optical Virtual Storage System
    WANG Zi-xuan, WEI Li, ZHANG Yu-ping
    2019, 0(05):  7.  doi:10.3969/j.issn.1006-2475.2019.05.002
    Asbtract ( 214 )   PDF (971KB) ( 98 )  
    References | Related Articles | Metrics
    The Hadoop distributed file system (HDFS CD-ROM database) based on CD-ROM database meets the current requirements of large data storage in terms of unit storage cost, data security and service life, etc., but it is not suitable for storing a large number of small files and real-time data reading. To better apply HDFS CD-ROM database in more big data storage scenarios, this paper proposes a magneto-optical virtual storage system (MOVS) more suitable for big data storage, which adds disk cache between HDFS CD-ROM database and users, and merges small files in disk cache into large files suitable for HDFS CD-ROM storage through file label classification, virtual storage, small file merging and other technologies, improving the data transmission speed. The system also uses file scheduling algorithm such as file pre-fetching and cache replacement to dynamically update the files in disk cache, so as to minimize the number of HDFS CD-ROM database accesses. The results of experiment show that MOVS can greatly improve the response time and data transmission speed compared with HDFS CD-ROM database.
    An Optimized Kernel State File Sending Method
    TU Xue-zhen
    2019, 0(05):  13.  doi:10.3969/j.issn.1006-2475.2019.05.003
    Asbtract ( 172 )   PDF (1061KB) ( 117 )  
    References | Related Articles | Metrics
    The traditional Linux kernel protocol stack can no longer meet the higher performance requirements of large-scale data processing systems for network transmission. A lot of researches have been done to move the protocols and interfaces from the kernel state to the user state, but there are few studies on kernel state optimization. Based on the analysis of the Linux kernel state file sending interface sendfile( ) processing flow and management mechanism, this paper proposes a kernel state file sending optimization method, using automatic load balancing fixed-length memory pool management, CPU affinity technologies, the kernel state interface sendfile( ) has been optimized and modified. It solves the problems of memory fragmentation, memory exhaust and CPU jitter under high load conditions, and improves data transmission performance effectively. Experimental results show that in the high concurrent and high throughput scenario, the system using this optimization method runs more stably, and the kernel state CPU occupancy rate drops by 50%.
    Stylometry-based Analysis of Literature Texts
    LI Yan-li, LI Wan-rong, LIAO Xin, LI Jing-juan, TANG Lu, LIU Xi-ping
    2019, 0(05):  19.  doi:10.3969/j.issn.1006-2475.2019.05.004
    Asbtract ( 244 )   PDF (1175KB) ( 122 )  
    References | Related Articles | Metrics
    This study compares literary works from the perspective of stylometry. At present, the research on literature is mainly qualitative and subjective analysis, and there are few quantitative studies and empirical analysis. A total number of 225 literary works are collected in the study, including Internet literary works and classical literary works, which are divided into three subsets, corresponding to the “excellent”, “good” and “poor”. For each work, a lot of features regarding article length, part of speech, rhythm, vocabulary, etc. are extracted. Based on these features, classifiers such as decision trees, neural networks and Bayesian are constructed. The models are utilized to find the key differences among the three datasets. The study found that the three datasets have obvious differences in stylometry statistics, and for different pair of datasets, the features have different discriminative power.
    A Micro-service-based Oil Big Data Mining Platform
    GUO Yi1, ZHANG Wei-shan1, XU Liang2, ZHAI Jia3
    2019, 0(05):  25.  doi:10.3969/j.issn.1006-2475.2019.05.005
    Asbtract ( 215 )   PDF (1252KB) ( 166 )  
    References | Related Articles | Metrics
     In order to promote the rapid fusion and application of big data technology in oilfield, a multi-functional big data processing platform covering the whole life cycle of big data processing is proposed. The platform combines various big data analysis frameworks and machine learning frameworks to design data mining functions that can support real-time and off-line processing in the oilfield. Based on Docker containers, it encapsulates all kinds of computing frameworks and algorithms, and  based on Kubernetes framework, it completes container arrangement and scheduling. In order to ensure the reliability and extensibility of the business system services, the system adopts the micro-service-based architecture, which decomposes the application of different technology stacks into a single service module independently. This allows enterprise data analysts to focus on business data analysis issues without spending a lot of time learning the details of framework deployment and other big data mining technologies.
    Financial Risk Control System Based on Rule Engine
    WANG Wen-jing, ZHANG Cheng-dian
    2019, 0(05):  30.  doi:10.3969/j.issn.1006-2475.2019.05.006
    Asbtract ( 415 )   PDF (822KB) ( 130 )  
    References | Related Articles | Metrics
    Financial technology enterprises have introduced a new type of financial model, P2P network lending model, which is dominated by small cash loans. Since the introduction of cash loan products, a large number of customers have been accumulated in a very short period of time. How to formulate a fast and effective financial risk control strategy, improve the efficiency of customer information data processing, timely forecast and prevent credit and fraud risks in business has become an urgent problem for financial enterprises to solve. In this regard, this paper proposes a financial risk control model based on Java rule engine, which realizes the decoupling of risk control rule strategy and program hard coding. On this basis, the design of feature factors and feature model is also carried out, which plays a great role in promoting the automatic approval of the credit system of financial science and technology enterprises.
    Vehicle Detection Method Based on SSD
    WU Shui-qing, WANG Yu, SHI Yan
    2019, 0(05):  35.  doi:10.3969/j.issn.1006-2475.2019.05.007
    Asbtract ( 280 )   PDF (3448KB) ( 112 )  
    References | Related Articles | Metrics
    The traditional vehicle target detection algorithm needs to select appropriate features for different image scenes, resulting in poor generalization ability. Therefore, an image detection method based on SSD (Single Shot MultiBox Detector) for vehicle is proposed. The method detects the vehicle by predicting the convolutional feature maps of multiple scales, and improves the detection accuracy of the vehicle to a certain extent. The small defects of the original SSD method in the training process are found, and the loss function is improved to optimize the training speed. Finally, the KITTI data set is used for training. The experimental results show that the method has a higher recognition rate for vehicles and the effect is better than traditional algorithms.
    Application of Gray-level Hybrid Method in Recognition of Container Number
    ZHANG Chao, LI Xiao-ping
    2019, 0(05):  41.  doi:10.3969/j.issn.1006-2475.2019.05.008
    Asbtract ( 151 )   PDF (2064KB) ( 113 )  
    References | Related Articles | Metrics
    A recognition method of container number based on machine vision is studied. For the gray level method in the process of color image preprocessing of containers, the traditional grays cale algorithm can not effectively make up for the defilement or other missing information in the image. So, a hybrid method combining principal component analysis (PCA) with the gray change rate of Bayes threshold estimation is proposed to optimize the gray level of the image. It can make up for the missing information and effectively determine the edge features after judging the change rate of the gray value of a certain point in the image and the gray value of the neighboring pixel points, thus greatly improving the character recognition accuracy of the subsequent sequence. Finally, a set of intelligent detection system for port container is designed and implemented by using the algorithm model. Through the Matlab test for the identification of 50 port container number images, compared with the ordinary mean method and the weighted average method, the accuracy is better by using the mixed gray method proposed in this paper, and the accuracy rate of the single character can reach 96%, the accuracy rate of the container number recognition can reach 92%.
    Box Dimension Measurement and Packing Optimization Based on RGBD
    QIN Wen-xiang, GUO Ling, LIN Shu-hong
    2019, 0(05):  46.  doi:10.3969/j.issn.1006-2475.2019.05.009
    Asbtract ( 255 )   PDF (1312KB) ( 148 )  
    References | Related Articles | Metrics
    In view of the serious waste of package materials in domestic logistics compared with foreign logistics, and the situation of the huge requirement of packing optimization, a new type of bin packing scheme is proposed, and a packing optimization method based on Genetic Algorithm is designed. The objective is to select a package specification with minimum surface area but can accommodate a number of cuboid-shaped items from existing package specifications. In the method, genetic codes involving packing sequence and orientation, along with a space segmentation method, is used to find a packing solution. Considering the high cost of traditional laser scanning measuring equipment, a box dimension measurement method based on RGBD is proposed to adapt to the management mode of domestic storage, and to provide dimension of items for packing optimization. Experiment results demonstrate that the proposed method is effective on finding package specification with minimum surface area and can measure box dimensions quickly and accurately.
    A Single Traffic Image Fast Dehazing Method with Sky Segmentation #br# and Local Transmittance Optimization
    LI Xi-ying1,2,3, ZHU Ken-gang1,2,3
    2019, 0(05):  51.  doi:10.3969/j.issn.1006-2475.2019.05.010
    Asbtract ( 176 )   PDF (6398KB) ( 145 )  
    References | Related Articles | Metrics
    Traffic photographs taken in hazy weather are degraded, due to the suspended particles in the air and scatter light. Since the scattered environment light is mixed into the light accepted by the observer, the contrast and sharpness of the hazy traffic image decrease, and the difficulties of subsequent processing and analysis increase. These problems, directly influence the full play of the circuit television surveillance system utility. Therefore, fast and effective traffic image dehazing has important application value. In the existing dehazing algorithm, the transmittance estimation deviates quite greatly from the actual situation. Especially when dealing with the sky area, it is easily lead to problems such as color distortion and halo effect. On the basis of the dark channel prior theory, this paper puts forward a fast haze removal method integrating sky segmentation with local transmittance optimization. First, the original traffic image is segmented into sky area and non-sky area by OTSU. Secondly, on the basis of the dark channel prior, the transmittance of the non-sky area is optimized by the maximum filtering and guided filtering, and the transmittance of the sky area is corrected by adaptive parameter adjustment method. In the end, the restored image is adjusted by Contrast Limited Adaptive Histogram Equalization (CLAHE) to improve the image brightness. Experimental results show that the proposed algorithm can effectively reduce the phenomenon of color distortion and halo effect in the sky area, and obtain a more natural and clear restored result. For the non-sky area, the clarity and contrast of the restored result are higher, and besides, the proposed algorithm keeps high efficiency. What’s more, compared with the dark channel prior algorithm, Tarel algorithm, Meng algorithm, Zhu algorithm and Berman algorithm, the proposed algorithm does better in terms of variance, average gradient, image information entropy and other indicators. This proposed algorithm can effectively and quickly restore the haze traffic image and reduce the color distortion and halo effect in the sky area. The restored image has good clarity and color revivification degree, and obtains better image sharpness and contrast enhancement. The proposed algorithm can provide good theoretical and technical support for the road traffic supervision.
    Optimization Method of Pipeline Based on BIM
    LI Chang-hua, TAO Jun-jie, LI Zhi-jie, GONG Yue-jun
    2019, 0(05):  59.  doi:10.3969/j.issn.1006-2475.2019.05.011
    Asbtract ( 213 )   PDF (1505KB) ( 150 )  
    References | Related Articles | Metrics
    An important function of Building Information Modeling(BIM) technology in integrated pipeline design is to solve pipeline collision problem. However, the efficiency of manually drawing and adjusting the pipeline is low, so this paper combines Revit secondary development technology to realize the rapid adjustment and optimization of BIM-based pipelines. Firstly, the primitive information in the integrated pipeline model is extracted, and a hybrid collision detection algorithm based on AABB and spatial geometry is constructed to identify the pipeline that generates the collision. Then, the integrated pipeline engineering design specification is parameterized, and the pipeline adjustment optimization algorithm is proposed to determine the pipeline to be adjusted and the adjustment range. Thereby, the automatic adjustment and optimization of the collision pipeline is realized, and the engineering design efficiency is improved.
    Application of LSTM in Typhoon Path Prediction
    XU Gao-yang, LIU Yao
    2019, 0(05):  64.  doi:10.3969/j.issn.1006-2475.2019.05.012
    Asbtract ( 379 )   PDF (1213KB) ( 217 )  
    References | Related Articles | Metrics
    The typhoon paths are essentially curves on a two-dimensional plane. According to the similarity of the two typhoon path curves, the numerical similarity and morphology similarity can be judged. Therefore, the dynamic time warping algorithm can be used to select the typhoon path similar to the target typhoon from the historical typhoon database. Meanwhile, considering the time correlation of typhoon path information, we put forword a long short-term memory model. The paper uses the latitude and longitude information of historical typhoon to predict the position information of the typhoon in the next 6 hours. Compared with the traditional method of predicting typhoon path by similarity, the long short-term memory model can effectively improve the prediction accuracy of typhoon paths, and the model is more stable and efficient.
    Design of Simulation Deploying Tools for Distributed Simulation Environment
    FANG Wei, XU Tao, YAN Wen-jun, ZHANG Bing-qiang
    2019, 0(05):  69.  doi:10.3969/j.issn.1006-2475.2019.05.013
    Asbtract ( 181 )   PDF (986KB) ( 141 )  
    References | Related Articles | Metrics
     The construction of simulation environment is the most complicated problem in the large-scale distributed simulation system, and it is easy to be influenced by people subjective factors. There are issues of many aspects such as difficult environmental reconstruction, low degree of automation and difficult data resources management. The Simulation Deploying Tool(SimDeploying), based on Client/Server structure for distributed environment, is discussed. By defining standardized object system, standardized interactive interface and unified data resource scheduling, it realizes the centralized management of simulation resources under the distributed environment, optimizes the deployment environment, simplifies the deployment process. Simulation shows that it can fastly realize the reconstruction of the deployment environment.
    Sequence Control System in Substation Based on Video Confirmation
    TIAN Yu1, LI Yu2, XIE Jia1
    2019, 0(05):  74.  doi:10.3969/j.issn.1006-2475.2019.05.014
    Asbtract ( 189 )   PDF (1457KB) ( 97 )  
    References | Related Articles | Metrics
    Currently, the substation sequence control system is manually operated with low efficiency and low degree of automation. Because State Grid Corporation has put forward high requirements on the correct use of the substation sequence control system, the single anti-misoperation mechanism in the sequence control system cannot meet the high security requirements that are set by State Grid Corporation. Based on the original fail-safe logic, this paper provides a method to identify the current operating status of the device through the video, using the template feature to match video analysis algorithms, and to recognize operating status of the device such as the disconnecting link and on-and-off device with high accuracy and reliability. As a result, different signal sources of the sequence control system are double-checked. The ways to avoid misoperation are increased. The false positive rate of the sequence control system is greatly decreased. The number of workers who operate the sequence control system is decreased. The work efficiency and automation degree of the substation sequence control system are enhanced. The safe operation of State Grid Corporation is guaranteed.
    Frame Rate Conversion Based on Motion Vector Projection
    ZHANG Di1, HUANG Qian1, CHEN Si-si2
    2019, 0(05):  80.  doi:10.3969/j.issn.1006-2475.2019.05.015
    Asbtract ( 281 )   PDF (1693KB) ( 91 )  
    References | Related Articles | Metrics
    This paper presents a frame rate conversion algorithm based on motion vector projection. In the motion estimation stage, the continuous elimination algorithm SEA is adopted, which is combined with the full search algorithm to optimize the calculation process of block matching criteria, so as to ensure the image quality and reduce the computational complexity at the same time. In the process of motion vector field projection, a new selection criterion of motion vector is defined, and the position information of the block is added on the basis of the matching criterion. Compared with the traditional standard, this criterion can better represent the true motion of the interpolated block, with higher accuracy. In the phase of motion compensation, the adaptive weighted compensation interpolation algorithm is used to consider the motion information of all overlapped projection blocks. The cavitation phenomenon is filled by the median filtering algorithm of motion vector. Experimental results show that the algorithm can reduce the loss of motion information and the interpolation effect is more accurate.
    Solving 0-1 Knapsack Problem Based on Weighted Greedy Firefly Algorithm
    REN Jing-min, PAN Da-zhi
    2019, 0(05):  86.  doi:10.3969/j.issn.1006-2475.2019.05.016
    Asbtract ( 238 )   PDF (865KB) ( 161 )  
    References | Related Articles | Metrics
    According to the characteristics of firefly algorithm, a weighted greedy firefly algorithm is proposed by combining adaptive weight, improved greedy algorithm, mutation operator with basic firefly algorithm. By adding adaptive weight and mutation operator, the global searching ability of the algorithm could be improved. The addition of greedy algorithm can improve the convergence speed of the algorithm to a certain extent. On the whole, the improved firefly algorithm improves the algorithm performance. Through the simulation experiment, the improved algorithm was compared with some basic algorithms. The experimental results show that the algorithm has obvious improvement in the speed and precision of solving 0-1 knapsack problem.
    Deletion Strategy Based on Learning Clause Length and LBD
    LIU Yao1,2, SONG Zhen-ming1
    2019, 0(05):  92.  doi: 10.3969/j.issn.1006-2475.2019.05.017
    Asbtract ( 119 )   PDF (675KB) ( 75 )  
    References | Related Articles | Metrics
    The deletion of the learning clause is very important in the composition of solver. The “good or bad” of the learning clause deletion strategy not only affects the efficiency of the BCP, but also affects the memory occupation. In order to avoid these problems, many scholars have done a lot of work and put forward a lot of good learning clause deletion strategies. However, the current learning clause deletion strategy has a disadvantage: it is possible that there is a great effect in the subsequent search process when deleting the learning clause, because we cannot guarantee that the clause that we delete each time has no “value”. Based on the length of learning clause and the decision-making layer of variable, this paper proposes a deletion strategy based on learning clause length and LBD-LLBD strategy. Then, the proposed strategy is embedded in the famous solver Glucose. Finally, experiments show that the proposed strategy can solve more examples. The efficiency of the solver is also improved, indicating that the strategy has certain advantages.
    Detection of QRS Waves Based on Wavelet Transform and Hilbert Envelope Analysis
    ZHANG Yi-fan, WANG Hao-ren, SHI Hao-tian, LIU Cheng-liang
    2019, 0(05):  96.  doi:10.3969/j.issn.1006-2475.2019.05.018
    Asbtract ( 184 )   PDF (1144KB) ( 123 )  
    References | Related Articles | Metrics
    This paper presents an algorithm for QRS detection using biorthogonal wavelet transform and Hilbert transform. First, this algorithm eliminates the high frequency noise and highlights the R peak position through biorthogonal wavelet transform decomposition and reconstruction, and constructs the detection layer which is beneficial to QRS waves detection. Then a technique with differentiation and Hilbert transform is applied on the re-composed signal in order to suppress the effects of P and T waves and decrease the low frequency noise. Finally, R peak is detected according to the adaptive threshold and decision rules. The MIT-BIH arrhythmia dataset is used to verify the performance of the detection method. The accuracy of QRS wave detection results is 99.01%, and the algorithm has good robustness and real-time performance.
    Short-term Prediction of Fresh Cut Flower Price Index Based on Time Series Neural Network
    PENG Wei
    2019, 0(05):  101.  doi:10.3969/j.issn.1006-2475.2019.05.019
    Asbtract ( 217 )   PDF (1609KB) ( 142 )  
    References | Related Articles | Metrics
     The fresh cut flower price index is a trend indicator reflecting the current status of the fresh cut flower market. It is of great significance to study the change of fresh cut flower price index and grasp the dynamics and regularity of flower market. Aiming at the sequence characteristics of fresh cut flower price index, this paper constructs the cut flower price index short-term forecasting model based on the L-M optimization algorithm of BP model. The model uses tansig and purelin as the transfer function between layers, uses the time series analysis method to determine the number of input layer neurons, and by comparison with experimental data to determine the number of hidden layer neurons. Three evaluation indexes of mean absolute error, mean relative error and root mean square error are used to test the prediction accuracy of the model. The experimental results show that the model is effective and has practical application value.
    Self-balancing Follow-up Robot Target Recognition and Loss Re-selection Strategy
    DU Wen-hao, HU Wei-ping, ZHANG You-xian, YU Jian-tao
    2019, 0(05):  108.  doi:10.3969/j.issn.1006-2475.2019.05.020
    Asbtract ( 182 )   PDF (1442KB) ( 84 )  
    References | Related Articles | Metrics
    In order to realize the adaptive following control of two-wheel self-balancing robot in a simpler and more effective way, the Apriltag algorithm of monocular vision target recognition can passively solve the relative position of camera in the three-dimensional coordinates, thereby positioning the target. In the tracking and selection of target recognition, the discriminant method, the threshold method and the sliding window method are used to define, and the polynomial curve fitting method is used to track and predict the next state of the target movement, and the motion trend is determined. In the case of losing the target, according to the design of a set of search programs, the compensation follows and regains the target. Tests show that the method proposed in this paper has a significant effect in regaining the target after target recognition and loss.
    Improved K-medoids Algorithm Based on Similarity Calculation Formula
    HAN Bing, JIANG He
    2019, 0(05):  113.  doi:10.3969/j.issn.1006-2475.2019.05.021
    Asbtract ( 149 )   PDF (806KB) ( 94 )  
    References | Related Articles | Metrics
    In the traditional K-medoids clustering algorithm, similarity is generally measured only by distance. This metric is based on independent and identically distributed attributes of data objects. But most real data object attributes are associated. Therefore, this article introduces the non-independent and identical distribution calculation formula. The traditional distance calculation similarity method is replaced. At the same time, since the non-independent and identical distribution formulas are calculated according to the frequency of the attribute values, but numerical data are not sensitive to frequency, so, numerical data are clustered and replaced by attribute columns before the introduction of formulas. Experimental results show that this method can improve the clustering accuracy of algorithm.
    An Improved Subspace Segmentation Method Based on Least Squares Regression
    CAI Xiao-yun1,2, YIN He-feng1, FU Wen-jin1, ZHAO Hang-tao1,3
    2019, 0(05):  118.  doi:10.3969/j.issn.1006-2475.2019.05.022
    Asbtract ( 187 )   PDF (648KB) ( 113 )  
    References | Related Articles | Metrics
     Least Squares Regression (LSR) is a common approach for subspace segmentation, it is very efficient due to a closed form solution. However, spectral clustering is exploited in LSR to obtain the final segmentation results. The drawback of spectral clustering is that it randomly initializes the cluster centers, which may undermine the subsequent clustering performance. In order to tackle this problem, this paper presents an improved LSR algorithm (LSR-DC) based on two characteristics of cluster centers, i.e. local density and distance. Experimental results on the Extended Yale B database show that LSR-DC is robust and is superior to the existing LSR subspace segmentation methods.
    Particle Swarm Optimization Algorithm Based on Population Division and Variation Strategy
    ZHANG Xiao-yan1, HE Jun-min2, LIU Wen-ying1, LIN Ya-lin1
    2019, 0(05):  122.  doi:10.3969/j.issn.1006-2475.2019.05.023
    Asbtract ( 149 )   PDF (1217KB) ( 108 )  
    References | Related Articles | Metrics
    Particle swarm optimization algorithm is of simple in form, flexible in parameter setting, easy to operate, and capable of fast convergence, so it has attracted many attentions. However, the traditional particle swarm algorithm also has its drawbacks: the slow convergence rate and the vulnerability to local optimization. In order to solve this problem, this paper uses the niche method to initialize the population, in the initial stages of evolution, dividing the initial population into subpopulations, and using different mutation strategies for different subpopulations; in the process of evolution, different inertia weighting factors are set for different subpopulations, in order to enhance the global search ability and the local search ability. The results of test functions show that the algorithm has faster convergence than the traditional particle swarm algorithm, the global optimal solution is closer to the real solution set, and the convergence accuracy is also higher.