Top Read Articles

    Published in last 1 year |  In last 2 years |  In last 3 years |  All
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Graph Neural Network-based Multi-agent Reinforcement Learning for Adversarial Policy Detection Algorithm
    SUN Qining1, GUI Zhiming1, LIU Yanfang2, FAN Xinxin3, LU Yunfeng4
    Computer and Modernization    2025, 0 (04): 42-49.   DOI: 10.3969/j.issn.1006-2475.2025.04.007
    Abstract566)      PDF(pc) (2811KB)(141)       Save
    In a multi-agent environment, the reinforcement learning model has security vulnerabilities in coping with adversarial attacks and is susceptible to adversarial attacks, of which adversarial policy-based adversarial attacks are more difficult to defend against because they do not directly modify the victim’s observations. To solve this problem, this paper proposes a graph neural network-based adversarial policy detection algorithm, which aims to effectively identify malicious behaviors among agents. This paper detects adversarial policy by training the graph neural network as an adversarial policy detector by employing alternative adversarial policies during the collaboration process of the agents, and calculates the trust scores of the other agent based on the local observations of the agents. The detection method in this paper provides two levels of granularity; adversarial detection at the game level detects adversarial policies with very high accuracy, and time-step level adversarial detection allows for adversarial detection at the early stage of the game and timely detection of adversarial attacks. This paper conducts a series of experiments on the StarCraft platform, whose results show that the detection method proposed in this paper can achieve an AUC value of up to 1.0 in detecting the most advanced adversarial policy-based adversarial attacks, which is better than the state-of-the-art detection methods. The detection method in this paper can detect adversarial policy faster than existing methods, and can detect the adversarial attack at the 5th time step at the earliest. Applying this paper’s detection method to adversarial defense improves the win rate of the attacked game by up to 61 percentage points. In addition experimental results show that the algorithm in this paper is highly generalizable and the detection method in this paper does not need to be trained again and can be directly used to detect observation-based adversarial attacks. Therefore, the method proposed in this paper provides an effective adversarial attack detection mechanism for reinforcement learning models in a multi-agent environment. 
    Reference | Related Articles | Metrics | Comments0
    Cross Border E-commerce Data Security Sharing Model Based on Zero Trust and Blockchain
    LI Xiaomeng1, 2, JIANG Rong1, LIANG Zhihong3, YU Yimin1
    Computer and Modernization    2025, 0 (04): 111-118.   DOI: 10.3969/j.issn.1006-2475.2025.04.017
    Abstract476)      PDF(pc) (8326KB)(118)       Save
    With the rapid development of cross-border e-commerce industry, the explosive growth of related data in cross-border e-commerce has made data security issues increasingly prominent. Traditional cross-border e-commerce platforms often face issues of data leakage, tampering, and untrustworthiness, leading to a decrease in the trust of merchants and users to the platform. To address this issue, this paper proposes a cross-border e-commerce data security sharing model based on zero trust and blockchain (EDSM-BZT). Firstly, a zero trust based multi factor identity authentication scheme is designed, which combines time and knowledge factors to dynamically verify each visitor. Secondly, a trust evaluation strategy is added to the scheme to calculate the trust value of the accessing subject and achieve more fine-grained control over user access behavior. Finally, an on chain and off chain collaborative storage sub model in the model is built to ensure the secure transmission and storage of cross-border e-commerce data. The experimental results show that the model can achieve dynamic control of access authorization and efficient storage of cross-border e-commerce data, with good performance and scalability.
    Reference | Related Articles | Metrics | Comments0
    Survey on Intelligent Optimization Algorithm for Feature Selection
    QI Haochun
    Computer and Modernization    2025, 0 (02): 33-43.   DOI: 10.3969/j.issn.1006-2475.2025.02.05
    Abstract367)      PDF(pc) (985KB)(141)       Save
     Feature selection, as one of the main techniques in data preprocessing, can effectively identify key features, thereby reducing dimensionality and effectively addressing the issue of “curse of dimensionality”. Feature selection is a typical NP-hard problem, and intelligent optimization algorithm have been widely employed in feature selection due to their remarkable global search ability. Firstly, this paper summarizes methods for evaluating feature importance and parameters updating. The former is used for evaluating the relevance and redundancy of features, while the latter is used for updating algorithm parameters. These two methodologies are both applicable to various crucial steps of intelligent optimization algorithm for feature selection. Then, the strategic design of three core steps in the process, namely algorithm initialization, population search, and objective function design, is introduced. The initialization strategy is summarized from the perspectives of decision space initialization and population initialization, with an analysis of the advantages and limitations of different strategies. Based on the population quantity, a detailed classification of search strategies for single population and multiple population is provided. According to the different metrics applied in the objective function, a categorization of objective function design can be summarized. Finally, it discusses future work for intelligent optimization algorithm to feature selection.
    Reference | Related Articles | Metrics | Comments0
    Construction of Depression Recognition Model Based on Multi-Feature Fusion
    HOU Menghan, WEI Changfa
    Computer and Modernization    2025, 0 (03): 1-5.   DOI: 10.3969/j.issn.1006-2475.2025.03.001
    Abstract348)      PDF(pc) (1144KB)(350)       Save
     In recent years, depression has become the primary problem of global mental health burden. In order to identify it, this paper proposes a depression recognition model combining BERT, BiLSTM and ConvNeXt. Firstly, the BERT model is used to generate feature vectors with rich semantics. Secondly, the BiLSTM, and ConvNeXt model is used to obtain the context information and the local features of the text, respectively. Thirdly, to alleviate the loss of semantic information in the feature extraction process, the context and local learned by BiLSTM and ConvNeXt models are fused through residual connections. Finally, depression is recognized according to the fused feature information. The experimental results show that the proposed model improves the accuracy, recall and F1 value compared with other deep learning models, which  can effectively extract the depression features of the text and improve the accuracy of depression recognition.
    Reference | Related Articles | Metrics | Comments0
    AMDFF-Net: Adaptive Multi-dimensional Feature Fusion Network for Tiny Object Detection
    LIU Yaokai, REN Dejun, LIU Chongyi, LU Yudong
    Computer and Modernization    2025, 0 (03): 106-112.   DOI: 10.3969/j.issn.1006-2475.2025.03.016
    Abstract324)      PDF(pc) (3369KB)(242)       Save
     Tiny object detection is a huge challenge in object detection research because tiny objects take up fewer pixels in the image, which results in a lack of feature information. To address this issue, an adaptive multi-dimensional feature fusion network (AMDFF-Net) for tiny target detection is designed to improve the accuracy of tiny object detection. Firstly, by integrating pooling layers and attention mechanisms, this paper constructs a pooling attention module, enabling the model to achieve a larger receptive field to enable self-adaptive and long-range correlations in self-attention. Secondly, an adaptive selection multi-dimensional feature fusion(ASMFF) module is designed, and an adaptive multi-dimensional feature pyramid network is designed based on the ASMFF module. This network adaptively fuses image features at different scales to enhance the information about tiny objects. To verify the performance and generalization of the model, experiments are conducted on the VisDrone2019, AI-TOD, and TinyPerson datasets. The experimental results show that AMDFF-Net improves the accuracy of tiny target detection, and the effectiveness of the proposed model in tiny target detection is verified by comparing with other mainstream algorithms.
    Reference | Related Articles | Metrics | Comments0
    Method for Underwater Robot Cage Inspection Control Based on Model Predictive Control
    ZHANG Jiaxu, LIU Xiaoyang, HONG Shengcheng, SHENG Yifan, WANG Mingyang
    Computer and Modernization    2025, 0 (03): 38-44.   DOI: 10.3969/j.issn.1006-2475.2025.03.006
    Abstract281)      PDF(pc) (3523KB)(147)       Save
    Aiming at the high cost and low efficiency of manual periodic inspection of aquaculture cages during the aquaculture process, a method based on MPC for ROV (Cable Remote Control Underwater Robot) aquaculture cage inspection trajectory tracking control is proposed. Firstly, the physical and kinematic constraints of the ROV based on actual conditions are included, and combining the ROV kinematic and dynamic models, a speed controlled MPC controller is designed. Using the traditional PID (Proposal Integration Difference) control algorithm as the baseline, a model-free PID controller is designed. Secondly, the simulation experiment of two-dimensional horizontal track and three-dimensional space track tracking in cage aquaculture environment is carried out and compared. Finally, the experimental results show that the proposed method has advantages such as good trajectory tracking performance, stable ROV operation, and small fluctuations. The method proposed in this paper for solving the low efficiency of aquaculture cage inspection provides an advanced inspection solution for the aquaculture industry.
    Reference | Related Articles | Metrics | Comments0
    A New Method of Pavement Disease Detection Based on Improved YOLOv8
    HE Feixiong1, XIE Haiwei1, PU Chao2, ZOU Chuanming2, JIA Yixuan1
    Computer and Modernization    2025, 0 (02): 108-113.   DOI: 10.3969/j.issn.1006-2475.2025.02.015
    Abstract260)      PDF(pc) (2770KB)(192)       Save
    As the operating time of a road increasing, the repeated effects of traveling loads and natural factors lead to deterioration of the road condition, and impacting its service life and quality. Therefore, In this paper, an improved YOLOv8 network is proposed for pavement disease detection. Firstly, targeted data enhancement techniques such as image flipping, lighting conditions change, and motion blur operation are applied, considering the characteristics of road disease images. Secondly, the loss function Wise-IoU is employed, which adopts a dynamic nonlinear focusing mechanism to evaluate the quality of the anchor box with outliers instead of IoU, and the wise gradient gain allocation strategy is provided to balance the differences in the number of samples among disease categories and improve the overall performance of the detector. Additionally, part of the C2F modules are replaced with DCNv3, and convolutional neuron weights are shared to reduce computational complexity and better learn features in pavement disease images. At the same time, multiple mechanisms are introduced, Softmax normalization along the sampling points enhances the model’s ability to understand road disease images. The experimental results show that the improved YOLOv8 road disease detection algorithm can achieve an accuracy of 77.3% in testing the network model, which is 3.9 percentage points higher than YOLOv8. mAP@50 reaches 76.9%, which is 3.4 percentage points higher than YOLOv8. This model can detect road diseases accurately and precisely, which is superior to the existing road disease detection algorithms and can applicate in engineering. 
    Reference | Related Articles | Metrics | Comments0
    ICS-ResNet: A Lightweight Network for Maize Leaf Disease Classification 
    JI Zhengjie, WEI Linjing
    Computer and Modernization    2025, 0 (04): 19-28.   DOI: 10.3969/j.issn.1006-2475.2025.04.004
    Abstract260)      PDF(pc) (7723KB)(136)       Save
    Accurate identification of maize leaf diseases plays a crucial role in preventing crop diseases and improving maize yield. However, plant leaf images are often affected by various factors such as complex backgrounds, climate conditions, lighting, and imbalanced sample data. To enhance recognition accuracy, this study proposes a lightweight convolutional neural network named ICS-ResNet, which is based on the ResNet50 backbone network and incorporates improved spatial and channel attention modules along with depthwise separable residual structures. The residual connections in the ResNet architecture prevent gradient vanishing during deep network training. The improved channel attention module (ICA) and spatial attention module (ISA) fully leverage semantic information from different feature layers to precisely localize key network features. To reduce the number of parameters and computational costs, traditional convolution operations are replaced with depthwise separable residual structures. Additionally, a cosine annealing learning rate strategy is employed to dynamically adjust the learning rate, mitigating training instability, enhancing the model's convergence ability, and preventing it from getting trapped in local optima.Finally, experiments were conducted on the Corn dataset from PlantVillage, comparing the proposed lightweight network with six other popular networks, including CSPNet, InceptionNet_v3, EfficientNet, ShuffleNet, and MobileNet. The results demonstrate that the ICS-ResNet model achieves an accuracy of 98.87%, outperforming the other six networks by 5.03, 3.18, 1.13, 1.81, 1.13, and 0.68 percentage points, respectively. Moreover, compared to the original ResNet50, the parameter size and computational cost are reduced by 16.27 MB and 2.25 GB, respectively, significantly improving the efficiency of maize leaf disease classification.
    Reference | Related Articles | Metrics | Comments0
    Dense Pedestrian Detection Algorithm Based on Improved YOLOv8
    DUAN Jingwei1, CHEN Liang1, LI Xue¹, LIU Mengmeng¹, LIU Jinyu²
    Computer and Modernization    2025, 0 (10): 24-31.   DOI: 10.3969/j.issn.1006-2475.2025.10.005
    Abstract238)      PDF(pc) (2315KB)(78)       Save
    Abstract: To address the issues of missed and false detections in dense pedestrian scenarios caused by complex backgrounds, high crowd density, low-light conditions, and partial occlusions, this paper proposes an optimized dense pedestrian detection algorithm based on YOLOv8n. The algorithm replaces the original convolutional blocks in the backbone network with efficient GSConv convolutions, reducing the model’s computational load while maintaining recognition accuracy. Additionally, GSConv convolutions enable the model to run efficiently on standard GPUs. The feature fusion network is replaced with the SlimNeck lightweight feature fusion module, which reduces the number of feature channels, thereby improving the model’s detection precision and speed. An EMA attention mechanism is embedded in the feature extraction network to enhance the model’s ability to capture both global and local information, thereby reducing false and missed detections in dense pedestrian scenarios. The algorithm also incorporates the Repulsion Loss function to better handle overlaps and occlusions among adjacent pedestrians in dense pedestrian detection, reducing interference between targets and optimizing bounding box regression. Training and validation on the CrowdHuman dataset demonstrate that the improved YOLOv8 model yields a 4.5 percentage points increase in mAP over the baseline. Furthermore, the model exhibits superior performance in dense crowds, occlusions, small-object detection, and low-light conditions, thereby offering an efficient and robust solution for dense pedestrian detection.

    Reference | Related Articles | Metrics | Comments0
    Infrared and Visible Image Fusion Based on Twin Axial-attention and Dual-discriminator Generative Adversarial Network
    WANG Lidan1, ZHAO Huaici2, PAN Duotao1, FANG Jian2, YUAN Decheng1
    Computer and Modernization    2025, 0 (04): 89-95.   DOI: 10.3969/j.issn.1006-2475.2025.04.014
    Abstract233)      PDF(pc) (5722KB)(84)       Save
    For the same scene, the fused image of infrared and visible can preserve the thermal radiation information of the foreground target and the background texture details at the same time, and the description is more comprehensive and accurate. However, many classical fusion algorithms based on deep learning usually have the defects of insufficient information retention and unbalanced feature fusion. To solve these problems, an image fusion algorithm based on twin axial-attention and dual-discriminator generating adversarial network is proposed. The generator uses a double-dense convolutional network as a multi-scale feature extractor and introduces spatially enhanced branch and twin axial attention to capture local information and long-range dependencies. The adversarial game between the dual discriminator and the generator is constructed, and the retention degree of differential features is balanced by restricting the similarity between the two source images and the fusion image. The perceptual loss function based on pre-trained VGG19 can overcome the problem of losing high-level features such as semantic-level features. The experimental results on the TNO dataset show that the proposed method achieves prominent fusion results with clear textures and has significant improvements in both subjective and objective evaluation metrics compared to other classical algorithms, demonstrating its advancement.
    Reference | Related Articles | Metrics | Comments0
    An Empirical Study on the Drift Diffusion of Task Interruption in User Identification of Phishing
    WANG Le, WANG Zhiying
    Computer and Modernization    2025, 0 (02): 1-12.   DOI: 10.3969/j.issn.1006-2475.2025.02.001
    Abstract231)      PDF(pc) (2802KB)(150)       Save
    Whether users can correctly identify phishing is the last line of defense against their attacks, and frequent and unavoidable interruptions pose a serious challenge for users to quickly process large amounts of emails and identify phishing. Task interruption has been proven to have both positive and negative effects on the main task, and research conclusions are inconsistent. Therefore, the role of task interruption in identifying phishing needs further explorations. This article is based on the drift diffusion model and constructs a research model for interrupt state, behavior selection, behavior selection, and interrupt state. Bayesian estimation of model parameters is used to analyze the drift rate, boundary height, starting point deviation, and non-decision time of user identification phishing in the presence or absence of interruptions. The analysis of online experimental data found that task interruption has a double-edged sword effect on correctly identifying phishing emails for users. When there is interruption, the user’s reaction time becomes shorter and there is no significant difference in accuracy, but the drift rate decreases, leading to higher boundary heights. In addition, relevant analysis on differences in accuracy, gender, susceptibility, and knowledge and experience found that when there is no interruption, the individual is male, the higher the accuracy, the lower the susceptibility, and the lower the knowledge and experience, the drift rate is faster, and the non-decision-making time for coding is shorter. However, when there is interruption, the individual is female, the lower the accuracy, and the lower the knowledge and experience, the drift rate is faster. This study extends the perspective of third-party task interruptions beyond the subject and object to explore the environmental factors that affect user identification of phishing, provides practical guidance for improving users’ ability to recognize phishing attacks.
    Reference | Related Articles | Metrics | Comments0
    Zero-shot Learning Based on Semantic Extension and Embedding
    GUO Chenguang, MAO Jian, WANG Yunyun
    Computer and Modernization    2025, 0 (02): 19-27.   DOI: 10.3969/j.issn.1006-2475.2025.02.003
    Abstract228)      PDF(pc) (2953KB)(378)       Save
     In zero-shot image classification, semantic embedding technology (i.e., using semantic attributes to describe class labels) provides the means to generate visual features for unknown objects by transferring knowledge from known objects. Current research often utilizes class semantic attributes as auxiliary information for describing class visual features. However, class semantic attributes are typically obtained through external paradigms such as manual annotation, resulting in weak consistency with visual features. Moreover, a single class semantic attribute is insufficient to capture the diversity of visual features. To enhance the diversity of class semantic attributes and their capacity to describe visual features, this paper introduces a Semantic Extension and Embedding for Zero-Shot Learning (SeeZSL) based on semantic extension and embedding. SeeZSL expands semantic information by constructing a latent semantic space for each class, enabling the generation of visual features for unknown classes using this semantic space. Additionally, to address the issues of weak consistency and the lack of discriminative ability between the original feature space and class semantic attributes, a semantic extension-based generation model is integrated with an contrastive-embedding model. The effectiveness of the proposed SeeZSL method was experimentally validated on four benchmark datasets.
    Reference | Related Articles | Metrics | Comments0
    Multi-scale Feature Image Defogging Algorithm Based on Content-guided Attention Fusion
    PU Yaya, WANG Yanbo, SU Yongdong, XU Zhongcheng
    Computer and Modernization    2025, 0 (03): 78-85.   DOI: 10.3969/j.issn.1006-2475.2025.03.012
    Abstract219)      PDF(pc) (2746KB)(288)       Save
     Aiming at the problems of color distortion and detail blur in current defogging methods, a multi-scale feature image defogging algorithm based on content-guided attention fusion is proposed with encoder-decoder network architecture. Firstly, multi-scale feature extraction module is used to encode, and three parallel expanded convolutions with different scales and squeeze and excitation attention are designed to enlarge the receptor field, extract features of different scales, and improve feature utilization. Secondly, in the decoder, the content-guided attention fusion module is designed to dynamically improve different weights for the deep and the shallow features to retain more effective feature information. Finally, pyramid scene parsing network is introduced to improve the ability of global information acquisition. The experimental results show that compared with other algorithms, the proposed algorithm improves 26.13% and 6.39% on the peak signal-to-noise ratio and structural similarity of SOTS datasets, respectively. The entropy and average gradient of the real fog datasets are increased by 3.27% and 21.09% respectively. The proposed algorithm improves the problem of defog incompleteness and detail blur.
    Reference | Related Articles | Metrics | Comments0
    Improved Traffic Sign Detection Algorithm of YOLOv7
    ZHAO Yin, YIN Siqing, ZHANG Yonglai
    Computer and Modernization    2025, 0 (02): 94-99.   DOI: 10.3969/j.issn.1006-2475.2025.02.013
    Abstract215)      PDF(pc) (1110KB)(197)       Save
    In view of the problems such as error detection and missing detection in the small pixel proportion of traffic signs, a traffic sign detection algorithm based on improved YOLOv7 is proposed. In YOLOv7, we introduce the small target detection layer and delete the large target detection layer to better meet the detection needs of small targets. In the backbone network, we introduce the EMA attention mechanism to improve the feature extraction capability of the model for multi-scale targets with reduced computational overhead. At the same time, ELAN-RPC module is constructed to replace the original ELAN, reduce the network calculation and improve the network reasoning speed. In addition, RFE module is introduced in the feature fusion layer to make better use of the details of the shallow feature map and improve the ability of subsequent top-down feature fusion. Experimental results show that the mAP of the improved YOLOv7 on TT100K dataset reaches 89.6%, which is 5.8 percentage points higher than that of the original algorithm, while the number of parameters is reduced by 37%, achieving the detection effect of fewer parameters and higher precision.
    Reference | Related Articles | Metrics | Comments0
    Gaze Estimation Model Based on Hybrid Transformer 
    CHENG Zhang, LIU Dan, WANG Yanxia
    Computer and Modernization    2025, 0 (04): 1-5.   DOI: 10.3969/j.issn.1006-2475.2025.04.001
    Abstract214)      PDF(pc) (1655KB)(175)       Save
     Combined CNN and Transformer, Transformer can gain the advantage of global feature information and improve the awareness of model context information, which can lead to improve model accuracy. A novel gaze estimation model RN-SA(ResNet-MHSA) based on a hybrid Transformer is proposed. In this model, part of the 3×3 spatial convolution layers in ResNet18 are replaced with a block composed of a 1×1 spatial convolution layer and MHSA(Multi-Head Self-Attention) layer, and the DropBlock mechanism is added to the model structure to increase the robustness of the model. Experimental results show that RN-SA model can improve the accuracy of the model while reducing the number of parameters compared with the current better model GazeTR-Hybrid, RN-SA model can improve the accuracy by 4.1% and 3.7% on EyeDiap and Gaze360 datasets, respectively, while the number of parameters is reduced by 15.8%.  Therefore, the combination of CNN and Transformer can be effectively applied to gaze estimation tasks.
    Reference | Related Articles | Metrics | Comments0
    Anomaly Detection Algorithm Based on Bidirectional Multi-scale Knowledge Distillation
    LIU Chongyi, LI Hua, REN Dejun, LIU Yaokai, WANG Yulong
    Computer and Modernization    2025, 0 (02): 58-63.   DOI: 10.3969/j.issn.1006-2475.2025.02.008
    Abstract209)      PDF(pc) (3688KB)(179)       Save

    Aiming at the problem of low anomaly detection and localization accuracy in current knowledge distillation-based anomaly detection algorithms due to the low difference in abnormal feature representation between teacher and student models, an anomaly detection algorithm based on bidirectional multi-scale knowledge distillation is proposed. An asymmetric teacher-student network structure composed of a teacher model, a student model and a reverse distillation student model is employed to suppress the student’s generalization to abnormal features. A feature fusion residual module is introduced between the bidirectional distillation student models to integrate multi-scale features and reduce abnormal disturbances. An attention module is introduced within the forward distillation student model to enhance the learning ability of important features. During the testing phase, anomaly assessment is performed through multi-scale anomaly map fusion. Experimental results on the public dataset MVTec AD show that the proposed algorithm, using ResNet18 as the backbone, achieves high scores of 97.7% at the pixel level and 98.8% at the image level on the area under the receiver operating characteristic curve evaluation metric, effectively improving the current knowledge distillation algorithms.

    Reference | Related Articles | Metrics | Comments0
    Resampling of Imbalanced Data for Optimizing Downstream Tasks 
    GUO Hua
    Computer and Modernization    2025, 0 (02): 28-32.   DOI: 10.3969/j.issn.1006-2475.2025.02.004
    Abstract203)      PDF(pc) (2178KB)(185)       Save
     Data resampling is a key method for correcting imbalanced dataset. Traditional methods construct balanced samples by minimizing geometric errors in the sample space, but they perform poorly in high-dimensional space with complex distribution patterns. Moreover, relying on statistical features lacks specificity for downstream tasks. To address this issue, this paper presents Sampling for Optimizing Downstream Neural Network (SOD-NN), a neural network for data sampling. This approach utilizes the ability of neural networks for nonlinear processing to identify the distribution characteristics of high-dimensional samples. It combines with downstream tasks to create a two-stage network, enabling overall optimization, thereby enhancing the model’s capability to meet the requirements of downstream tasks effectively. Specifically, the dataset is first divided spatially during sampling. Residual processing of sample subsets is then applied to prevent data degradation. Subsequently, a self-attention mechanism is utilized to construct global feature, ensuring consistency with the original sample distribution. Experimental results indicate that the model proposed in this paper significantly improves the recognition performance of minority class samples in downstream classification tasks, enhancing the robustness of processing these tasks.
    Reference | Related Articles | Metrics | Comments0
    Optimization for Camera Self-calibration Based on Horizon Detection in Road Scenes
    HE Guotao1, ZHAO Chunhui2, LIU Zhenyu1, WANG Long1
    Computer and Modernization    2025, 0 (02): 77-85.   DOI: 10.3969/j.issn.1006-2475.2025.02.011
    Abstract199)      PDF(pc) (3553KB)(143)       Save
    The current camera calibration in traffic scenes relies mainly on the key information of the road scene and relies on redundant information such as road dashed lines, parallel lines, etc. to optimize the camera calibration parameters. However, due to the limited information of the scene, the range of vanishing points cannot be fixed, and at the same time, due to the existence of the camera spin angle, the results of the camera calibration have a certain degree of error. Starting from the horizon detection, a deep learning key point detection-based horizon detection algorithm improves the accuracy to 82.46%. Subsequently, the camera self-calibration parameters are optimized by correcting the camera spin angle based on horizon detection and providing stricter constraints by using the horizon. The experimental results show that after correcting the camera spin angle and providing stronger constraints by using the horizon, the camera self-calibration parameters obtain faster convergence and a minimum of 1.79% error.
    Reference | Related Articles | Metrics | Comments0
    Review of Large Language Model Question Answering Systems for International Event Analysis
    LEI Jiyue, SU Peng, NIE Yun, LIN Chuan
    Computer and Modernization    2025, 0 (03): 29-37.   DOI: 10.3969/j.issn.1006-2475.2025.03.005
    Abstract197)      PDF(pc) (984KB)(132)       Save
     As the core focus of current artificial intelligence research, large language models have demonstrated strong cross-domain understanding and generation capabilities. They are widely used in many fields including event analysis, and promote the innovation and development of intelligent question answering system with its excellent performance. Although large language models show strong processing ability in general Q&A scenarios, they still face challenges in dealing with international events with deep professional backgrounds and high dynamics that affect international relations. In recent years, many scholars have focused on domain-specific large language models and quantitative analysis system for international relations, but there are few literature reviews on the cross-field of the combination of the two. In order to provide a comprehensive framework for developers and researchers of large language model question answering systems for international event analysis. Firstly, starting from the early international event analysis system, combined with the actual needs of international event analysis, the applicability of various general large language models in this task is analyzed. Secondly, by referring to the successful cases in the fields of finance, education, medicine, law, etc., the strategy of constructing the domain-specific large-model question-and-answer system for international event analysis is extracted. In addition, the open data set resources closely related to the task are systematically combed. Finally, the current bottleneck and the future development direction are deeply analyzed.
    Reference | Related Articles | Metrics | Comments0
    Twin Feature Fusion Network for Scene Text Image Super Resolution
    FENG Xinjie, WANG Wei
    Computer and Modernization    2025, 0 (02): 86-93.   DOI: 10.3969/j.issn.1006-2475.2025.02.012
    Abstract189)      PDF(pc) (2309KB)(161)       Save
    The aim of the scene text image super-resolution (STISR) method is to enhance the resolution and legibility of text images, thereby improving the performance of downstream text recognition tasks. Previous studies have shown that the introduction of text-prior information can better guide the super-resolution. However, these methods have not effectively utilized text-prior information and have not fully integrated it with image features, limiting super-resolution task performance. In this paper, we propose a Twin Feature Fusion Network (TFFN) to address this problem. The method aims to maximize the utilization of text-prior information from pre-trained text recognizers, with a focus on the recovery of text area content. Firstly, text-prior information is extracted using a text recognition network. Next, a twin feature fusion module is constructed, which employs a twin attention mechanism to facilitate bidirectional interaction between image features and text-prior information. The fusion module further integrates context-enhanced image features and text-prior information. Finally, sequence features are extracted to reconstruct the text image. Experiments on the benchmark TextZoom dataset show that the proposed TFFN improves the recognition accuracy of the ASTER, MORAN, and CRNN text recognition networks by 0.22~0.5, 0.6~1.1 and 0.33~1.1 percentage points, respectively.
    Reference | Related Articles | Metrics | Comments0
    Survey of Application of Knowledge Graph in Field of Intelligent Manufacturing
    JIANG Sulun1, 2, 3, YUAN Decheng1, GUO Qingda2, 3, LIU Jian3, YU Guangping2, 3
    Computer and Modernization    2025, 0 (05): 48-59.   DOI: 10.3969/j.issn.1006-2475.2025.05.007
    Abstract188)      PDF(pc) (3678KB)(1088)       Save
     With the rapid promotion of new-goneration artifica  intelligence, computing, and  other technologies, the manufacturing field urgently needs to undergo intelligence and digitalization and upgrading. Through literature review and applied case studies, it is found that the construction of knowledge graph can promote the development of industrial intelligence, so the field of intelligent manufacturing has begun to apply knowledge graph to manage and optimize intelligent manufacturing equipment data and processes. At present, knowledge graph technology has been maturely applied in the direction of intelligent question answering, personalized recommendation, etc., in order to explore the greater application potential of knowledge graph technology in the field of intelligent manufacturing, the current literature and application status is studied in detail and conduded. This paper firstly starts with the popular technologies such as knowledge acquisition, knowledge fusion, and knowledge reasoning involved in knowledge graphs, and then focuses on the research and analysis of several popular application directions such as industrial fault diagnosis, digital twins, human-computer collaborative interaction, and risk management based on knowledge graphs, and summarizes the general architecture, and discusses the future development trends and difficulties such as the combination with AIGC technology, and puts forward future prospects to provide reference for promoting intelligent manufacturing knowledge graphs.
    Reference | Related Articles | Metrics | Comments0
    Method for Predicting Agricultural Product Prices Based on Improved TimesNet
    WANG Yinbing1, 2, WANG Xingfen1, 3, LI Libo1, 3
    Computer and Modernization    2025, 0 (10): 89-95.   DOI: 10.3969/j.issn.1006-2475.2025.10.014
    Abstract185)      PDF(pc) (4010KB)(47)       Save

    Abstract: Predicting agricultural product prices plays a key role in stabilizing the agricultural market. However, due to the influence of various factors, agricultural product prices exhibit characteristics such as non-linearity and periodicity, making it difficult to accurately predict. To solve this problem, a new agricultural product price prediction model, EMD-ConvNeXtV2-TimesNet, is proposed. The model introduces two innovations based on the TimesNet model: first, it innovatively incorporates an Empirical Mode Decomposition (EMD) module to decompose the original price series, thereby better capturing the intrinsic structural information of the price series; second, it improves the image feature extraction module of TimesNet to a ConvNeXtV2 Block to more effectively capture the cyclical information of prices. Comparative experiments were conducted on the collected datasets of corn, eggs, soybeans, and peanuts. The experimental results show that compared with the best prediction results of comparison models such as DLinear, Informer, Transformer, FiLM, FEDformer, the Mean Absolute Percentage Error (MAPE) and Mean Absolute Error (MAE) are reduced by 38.902%/38.562%, 33.183%/33.108%, 39.471%/35.178%, and 48.525%/47.806% respectively. The new model has achieved significant accuracy improvements. Ablation experiments further confirmed the complementary role of the EMD module and ConvNeXtV2 Block in the model, which more effectively reduces the price prediction error compared to the original TimesNet.

    Reference | Related Articles | Metrics | Comments0
    Video Rain Removal Algorithm Based on Deep Learning
    YAN Qiang1, SHEN Shouting2, BAI Junqing2, CHENG Guojian2
    Computer and Modernization    2025, 0 (02): 100-107.   DOI: 10.3969/j.issn.1006-2475.2025.02.014
    Abstract181)      PDF(pc) (6182KB)(145)       Save
    In view of the fact that most traditional video rain removal algorithms only focus on removing rain marks and are trained only on synthetic data, ignoring more complex degradation factors such as rain accumulation, occlusion, and prior knowledge in real data. In this paper, we propose a two-stage video deraining algorithm that combines synthetic and real videos. The first stage algorithm performs a reverse recovery process under the guidance of the proposed rain removal model Initial-DerainNet. Continuous rain frames containing degradation factors are input into the network and physical prior knowledge is integrated to obtain an initial estimated rain-free frame. The second stage uses adversarial learning to refine the results, that is, to restore the overall color, illumination distribution, etc. of the initially estimated rain-free frame to obtain a more accurate rain-free frame. Experimental results show that the PSNR value of this algorithm reaches 35.22 dB and the SSIM value reaches 0.9596 on the synthetic rain removal data set RainSyntheticDataset100, which is better than benchmark rain removal algorithms such as JORDER, DetailNet, SpacNN, SE, J4Rnet and FastDeRain. On the real rain video test set, the algorithm in this paper can achieve PNSR values of more than 30 dB on rain videos of different dimensions, which is better than other rain removal algorithms in terms of subjective visual effect and data metrics, and can effectively improve the quality of rainy day videos.
    Reference | Related Articles | Metrics | Comments0
    Improved Classroom Behavior Detection Algorithm for YOLOv8
    SU Yansen, MOU Li
    Computer and Modernization    2025, 0 (08): 76-81.   DOI: 10.3969/j.issn.1006-2475.2025.08.011
    Abstract181)      PDF(pc) (1863KB)(97)       Save

    Abstract: Aiming at the problems of low detection accuracy of student classroom behavior under monitoring and difficulty in deploying models, an improved YOLOv8 algorithm is proposed for detecting student behavior. Firstly, the YOLOv8 backbone network is improved by introducing the Swin Transformer network as the backbone feature extraction network to reduce information loss and improve the effectiveness of feature extraction. Secondly, to enhance the model’s attention to the features of distant targets, a flexible dual channel attention mechanism EMA is introduced, which makes the model focus more on targets with fewer pixels at long distances and improves detection accuracy. Finally, in the Neck section, the Slim Neck design paradigm containing GSConv is used to make lightweight improvements to the model. The experimental results on the SCB-Dataset3 dataset show that the improved model has a parameter count of 3.3 M and a computational load of 11.1 GFLOPs, respectively, with a detection accuracy of 88.75%. Compared with the original model, the parameter count is reduced by 40.7%, the computational load is reduced by 15.9%, and the detection accuracy is improved by 7.7 percentage points. This achieves good detection accuracy while achieving model lightweighting.

    Reference | Related Articles | Metrics | Comments0
    Real-time Semantic Segmentation Based on Gate-controlled Fusion
    FENG Zuyan, WEI Yan, CHEN Jiakun, YU Xin
    Computer and Modernization    2025, 0 (02): 121-126.   DOI: 10.3969/j.issn.1006-2475.2025.02.017
    Abstract176)      PDF(pc) (2056KB)(111)       Save
    Feature fusion in real-time semantic segmentation needs to pay attention to both shallow and deep information, while the current feature fusion methods require a huge amount of computation and parameter count, which is difficult to meet the requirements of real-time semantic segmentation in terms of accuracy and speed. To address this problem, a real-time semantic segmentation method based on gated fusion is proposed from the comprehensive consideration of both real-time and performance of the network. The method contains an encoder, a gated feature fusion module, a pixel-level feature extraction module, and a gated aggregation segmentation head. Firstly, the image to be segmented is feature extracted by the encoder. Secondly, the important feature information is accurately extracted by the pixel-level feature extraction module, then the deep semantic information and the shallow location information are feature fused by the gated feature fusion module. Finally the semantic segmentation is completed by the gated aggregation segmentation head. On the dataset CamVid, the mean intersection over union of the model segmentation is 87.31%, and the frame rate of segmentation is 75.3 fps. On the dataset Cityscapes, the mean intersection over union of the model segmentation is 79.19%, and the frame rate of segmentation is 44.1 fps. Experimental results show that the proposed segmentation method performs well in both accuracy and real-time, and it can be effectively applied to real-time semantic segmentation tasks.
    Reference | Related Articles | Metrics | Comments0
    Remote Sensing Image Classification Based on Multi-scale Feature Extraction
    LUO Hao, LI Xianfeng
    Computer and Modernization    2025, 0 (03): 86-92.   DOI: 10.3969/j.issn.1006-2475.2025.03.013
    Abstract176)      PDF(pc) (1491KB)(212)       Save
    In order to improve the accuracy of convolutional neural network in the classification task of remote sensing image, a method based on multi-scale feature extraction for classification is proposed. Aiming at the problems that the scale difference of target objects in remote sensing images is large and ordinary convolutional extraction tends to produce redundant features, a multi-scale hybrid convolution module with ordinary convolution and atrous convolution is proposed, which can effectively enhance the feature extraction ability of the model. In terms of feature fusion, a feature cross fusion module is proposed, which can effectively fuse the semantic information of each branch and make full use of the information between various scale features for deep fusion and interaction. Aiming at the problem of complex land cover information in remote sensing images, a parallel attention module is proposed to apply attention mechanisms to each branch, which makes it pay more attention to the key parts of the image and ignore the redundant information. Compared with the existing methods, the classification performance of the proposed method is significantly improved. On the data sets WHU-RS19、RSSCN7 and SIRI-WHU,the overall accuracy of the proposed method reaches 98.19%、94.18% and 96.37%, respectively.
    Reference | Related Articles | Metrics | Comments0
    Blockchain-based Distributed Contract Electricity Transfer 
    ZHANG Zihao1, 3, YE Meng2, PAN Shixian1, MA Li2, BAO Tao1, YU Qi3
    Computer and Modernization    2025, 0 (02): 13-18.   DOI: 10.3969/j.issn.1006-2475.2025.02.002
    Abstract175)      PDF(pc) (1268KB)(125)       Save
    This paper introduces a distributed contract-based electricity transfer system leveraging blockchain technology to adapt to the diversity of the electricity market and the volatility in electricity demand. The necessary contracts for electricity transfer, namely, original contracts, electricity mutual assurance agreements, and transfer transaction contracts, are categorized into three distinct types. Specific protocols are designed for the generation of each type of contract. Verification of contract authenticity is performed using ordered multi-signature schemes, ensuring the validity of the contracts. Furthermore, broadcast encryption is employed to safeguard the privacy and accuracy of contract content. All contracts are stored on the blockchain, guaranteeing their immutability. The results of experiments demonstrate that this system significantly enhances the efficiency of contract generation and verification, facilitating swift and secure electricity transfers.
    Reference | Related Articles | Metrics | Comments0
    Improved YOLOv8s Algorithm Based on GiraffeDet for Transmission Line Icing Detection
    TANG Rui1, WU Jianchao1, CHEN Jianbo1, CHAI Jiang1, WANG Qian1, HE Yuchen2
    Computer and Modernization    2025, 0 (03): 6-11.   DOI: 10.3969/j.issn.1006-2475.2025.03.002
    Abstract173)      PDF(pc) (4197KB)(231)       Save
    The icing of transmission lines can greatly impact the safety and stability of the power grid system. Due to the distribution of transmission lines in mountainous areas, forest areas, and unmanned open areas, workers cannot obtain on-site information in the event of damage such as rain, snow, and freezing. To accurately identify the icing situation of transmission lines in complex environments such as mountainous and uninhabited areas, this paper proposes an improved YOLOv8s-based detection method. Firstly, SIoU is adopted as the loss function to improve the training speed and accuracy of the model.  Secondly, by replacing some ordinary convolutions with dual convolutions, the information exchange between different channels is enhanced, effectively improving the efficiency of feature extraction, thereby further accelerating the convergence speed of the model. Finally, the GiraffeDet network structure is introduced to replace the original network structure and utilizes multi-scale information and the global context of feature map to make the model perform better in detecting small targets and complex scenes, improving the accuracy and robustness of detection. The experimental results show that compared with YOLOv8s, the improved method meets certain requirements for accuracy, reduces the model size by 7.3 MB, and significantly improves speed.
    Reference | Related Articles | Metrics | Comments0
    Box Office Prediction Model Based on SA-EW-LSTM
    LANG Kun, NIU Chunhui, LI Chenqiong, ZENG Suyu
    Computer and Modernization    2025, 0 (05): 1-9.   DOI: 10.3969/j.issn.1006-2475.2025.05.001
    Abstract173)      PDF(pc) (2701KB)(157)       Save
     Film box office is usually affected by many factors. However, the word-of-mouth, as a key factor, is often neglected by traditional prediction models. In order to improve the prediction accuracy, a novel box office prediction model based on the sentiment analysis, the entropy weight method and the LSTM neural network is presented in this paper. Firstly, the input index system is constructed by selecting eight influence factors, namely, word-of-mouth, box office of the previous day, box office of the same day last week, ticket price, service fee, screening rate, holiday or not, and search index. Secondly, the method of sentiment analysis is used to analyze the text of film review, and the sentiment index is used to quantify the word-of-mouth factor. Then, in order to quantify the impact of different factors on the box office, the entropy weight method is used to assign weights to different factors. Finally, the Long Short-Term Memory neural network is applied to predict the box office. Simulation results indicate that the prediction accuracy of the presented SA-EW-LSTM model reaches 94.9% and 94.8% on two different data sets, respectively, which is obviously superior to the other five classical models, and the effectiveness of the presented model is verified.
    Reference | Related Articles | Metrics | Comments0
    Power Load Forecasting Based on TCN and Lightweight Autoformer
    LI Ming, SHI Chaoshan, WEN Guihao, LUO Yonghang, TAN Yunfei
    Computer and Modernization    2025, 0 (04): 6-11.   DOI: 10.3969/j.issn.1006-2475.2025.04.002
    Abstract162)      PDF(pc) (3080KB)(118)       Save
    The accuracy of power load forecasting is crucial for energy conservation and emission reduction, and higher accuracy can enable power companies to make more reasonable plans and improve economic benefits. Although Autoformer, based on the improved Transformer architecture, has achieved good results in sequence prediction tasks, it did not fully consider the causal relationship of time when extracting temporal features, and there is too much redundant information in the attention layer, which leads to a decrease in model accuracy and memory consumption. To address these issues, this paper proposes a power load forecasting method that combines Time Convolutional Network (TCN) and an improved lightweight Autoformer model. Firstly, a time convolutional network is introduced into the Autoformer encoder to provide a larger receptive field and fully consider the causal relationship of the samples. Then, a distillation mechanism is added between the autocorrelation attention layers to reduce the number of model parameters. Finally, the results of experiment on five public datasets showed that the lightweight Autoformer combined with TCN reduced MSE and MAE by 8.95% to 32.40% and 4.91% to 15.51% respectively compared to the original model, and the prediction performance is significantly better than the other four mainstream methods, demonstrating its excellent performance.
    Reference | Related Articles | Metrics | Comments0
    One-stage Semi-supervised Object Detection by Reusing Unreliable Pseudo-labels
    SHAO Yeqin1, WANG Haiquan2, ZHOU Kunyang3, GUO Yudi2, SHI Quan1
    Computer and Modernization    2025, 0 (03): 52-59.   DOI: 10.3969/j.issn.1006-2475.2025.03.008
    Abstract161)      PDF(pc) (1927KB)(123)       Save
    The key to semi-supervised object detection methods is to assign pseudo labels to the targets of unlabeled data. To guarantee the quality of pseudo-labels, the semi-supervised object detection methods usually use a confidence threshold to filter low-quality pseudo-labels, which will cause most pseudo-labels to be removed due to their low confidence. Contrastive learning is used to reuse most of low-confidence unreliable pseudo labels for boosting the performance of semi-supervised object detection method. Specifically, the pseudo-labels are divided into reliable and unreliable ones according to the prediction confidence. Besides the reliable pseudo-labels, the unreliable pseudo-labels are exploited as negative samples for model training of contrast learning. To balance the number of unreliable pseudo-labels between different classes, a memory module is designed to store the unreliable pseudo-labels of different batches in the training process. The experimental results show that the mAP of the improved semi-supervised method on COCO data set is 13.6%, 23.0%, and 27.5% with the labeling ratio of 1%, 5%, and 10%, which is better than the existing semi-supervised learning methods. On the COCO-additional data set, the mAP of the improved semi-supervised method reaches 44.7%, which is 4.5 percentage points higher than supervised learning.
    Reference | Related Articles | Metrics | Comments0
    UAV Path Planning Based on YOLO and PPO
    ZHANG Huiyu1, LIU Lei1, YAN Dongmei2, LIANG Chengqing3
    Computer and Modernization    2025, 0 (04): 50-55.   DOI: 10.3969/j.issn.1006-2475.2025.04.008
    Abstract161)      PDF(pc) (6883KB)(210)       Save
     This paper proposes an unmanned aerial vehicle path planning method based on deep reinforcement learning for complex and ever-changing three-dimensional unknown environments. This method optimizes strategies within a limited observation space to address the challenges posed by high complexity and uncertainty. Firstly, within a limited perceptual range, the YOLO network is used to extract obstacle information from the image information. Secondly, this paper designs hazard levels to address the issue of varying amounts of obstacle information at different times, and combines the extracted information from hazard levels with radar information as input to the intelligent agent. Finally, based on the proximal strategy optimization algorithm, an action selection strategy under state decomposition is designed to improve the effectiveness of drone actions. Through simulation evaluation in Gazebo, the experimental results show that compared to the proximal strategy optimization algorithm, the average reward per round has increased by 15.6 percentage points, and the average success rate has increased by 2.6 percentage points.  
    Related Articles | Metrics | Comments0
    A Triple Joint Extraction Model for Talent Resume Information
    SHEN Xinke1, 2, LI Yong1, 2, WEN Ming2, REN Yuanyuan2
    Computer and Modernization    2025, 0 (02): 52-57.   DOI: 10.3969/j.issn.1006-2475.2025.02.007
    Abstract159)      PDF(pc) (1346KB)(98)       Save
    The field of talent title evaluation contains a large amount of talent resume information, but resume information often exists in the form of natural language, which experts find difficult to use as a basis for talent title evaluation. To address this issue, this article combines entity extraction and relationship extraction for joint modeling, and constructs a triplet joint extraction model (RLAC) for talent resume information. Firstly, the Chinese pre-trained language model RoBERT-wwm is used to encode the underlying talent resume information. Secondly, the introduction of LSTM network and attention mechanism improves the problem of difficult recognition of head entities in talent resume information, and enhances the ability to extract semantic features in coding context. Thirdly, input the encoded information into the header entity annotator to obtain the header entity. Finally, concatenate the head entity and talent resume information and input them into the tail entity relationship annotator to alleviate the problem of relationship overlap, thus obtaining a triplet. Compared with the baseline model, the experimental results on the talent resume dataset of the proposed model has improved accuracy, recall, and F1 value, indicating that the model has good triplet extraction ability.
    Reference | Related Articles | Metrics | Comments0
    Tongue Constitution Classification Method Based on Deep Learning
    XIE Haiqing, LING Jiaqi, YI Xinbo
    Computer and Modernization    2025, 0 (03): 99-105.   DOI: 10.3969/j.issn.1006-2475.2025.03.015
    Abstract158)      PDF(pc) (5379KB)(665)       Save
    In response to the minimal inter-class differences in tongue images and the insufficient feature extraction by traditional networks, this paper constructs datasets for tongue image semantic segmentation and classification and conducts data preprocessing. Based on RepVGG network algorithm design and optimization, a multi-feature fusion tongue constitution classification network MTSNet based on convolutional neural network is proposed. MTSNet employs a multi-scale feature pyramid and combine high-level and low-level semantic information learned by the network to enhance the network’s representational capabilities. The addition of squeeze-excitation convolutional layers in the RepBlock module enables the network to focus more on information-rich features. The experimental results show that MTSNet significantly enhances classification performance across nine types of tongue constitutions, and its accuracy is 32.11 percentage points higher than that of AlexNet, 22.37 percentage points higher than that of SVM, and 17.68 percentage points higher than that of Resnet-18. Compared with the unoptimized RepVGG network, MTSNet achieves improvements of 9.90 percentage points in accuracy, 14.01 percentage points in macro-averaging, 9.90 percentage points in micro-averaging, and 11.09 percentage points in weighted-averaging. This tongue constitution, classification method provides scientific basis for users’ health management and has good reference application for traditional Chinese medicine’s adjunctive treatment and scientific research.
    Reference | Related Articles | Metrics | Comments0
    DDoS Attack Detection Method Based on Transformer Architecture
    CHI Biwei1, SUN Rui2
    Computer and Modernization    2025, 0 (05): 36-40.   DOI: 10.3969/j.issn.1006-2475.2025.05.005
    Abstract157)      PDF(pc) (1604KB)(205)       Save
    With the rapid development of the Internet, DDoS attacks have become a major challenge in the field of network security. DDoS attacks disrupt the normal operation of target servers by controlling a large number of distributed computers to send massive amounts of malicious requests, seriously affecting the stability and security of network services. Traditional DDoS attack detection methods, such as rule-based detection, statistical methods, and machine learning approaches, often face issues such as high false positive rates and low detection efficiency when dealing with complex and dynamically changing network traffic. To address these issues, this paper proposes a Transformer-based DDoS attack detection system. The system utilizes the powerful self-attention mechanism of the Transformer model to capture long-term dependencies in network traffic, enabling more accurate identification of abnormal traffic patterns. Additionally, by incorporating positional encoding, the system can better handle temporal information and enhance the model’s ability to perceive global network traffic. Experimental results on datasets show that the Transformer-based DDoS detection model significantly outperforms comparison methods in terms of detection accuracy and recall rate, demonstrating the effectiveness of the proposed approach. 
    Reference | Related Articles | Metrics | Comments0
    Network Intrusion Detection Method Based on Convolutional Neural Networks with convLSTM
    ZHANG Yue, GUO Zixin, HUANG Yibin, YAN Tao
    Computer and Modernization    2025, 0 (03): 119-126.   DOI: 10.3969/j.issn.1006-2475.2025.03.018
    Abstract156)      PDF(pc) (1102KB)(171)       Save
     In the field of network intrusion detection, machine learning methods that manually extract features in feature engineering are generally used, but the manual feature extraction method is prone to losing important feature information; In addition, different types of attack traffic play different roles in detection, and existing algorithms generally suffer from important information loss and low accuracy in identifying attack types. A hybrid algorithm based on Convolutional Long-Short Term Memory (convLSTM) and Convolutional Neural Networks (CNN) is proposed for anomaly traffic detection in response to the aforementioned issues, Which directly use the payload of network traffic as data samples without manual extraction of complex traffic features, fully explores the structural features of traffic, extracts temporal and spatial features, and generates accurate intrusion detection feature vectors. The experimental results show that on the CIC-ISDS2017 dataset, the accuracy of the hybrid algorithm convLSTM-CNN in network intrusion detection reaches 99.39%. Compared with the simple DNN, SVM, LSTM, GRU-CNN and other models, it has a higher accuracy and lower false alarm rate, indicating that the algorithm improves the efficiency of abnormal traffic detection.
    Reference | Related Articles | Metrics | Comments0
    Ultrasonic Image Segmentation of Thyroid Nodules Based on HA-UNet++
    ZHU Yongtian, TIAN Fei, DONG Baoliang
    Computer and Modernization    2025, 0 (03): 93-98.   DOI: 10.3969/j.issn.1006-2475.2025.03.014
    Abstract156)      PDF(pc) (2690KB)(148)       Save
    Thyroid disease is one of the most frequently diagnosed nodular lesions in adult population, and it’s incidence is increasing year by year. With the development of artificial intelligence technology, the automatic diagnosis of thyroid ultrasound images using computer vision technology can significantly improve the accuracy and efficiency of diagnosis. However, most image segmentation methods based on deep learning, limited by the size of receptive field, cannot focus on the important features of the image in time and extract them effectively, resulting in low segmentation accuracy. In order to solve the above problems, a new deep learning network model HA-UNet ++ (Hybrid Dilated Convolution-Attention-UNet++) is adopted in this paper to segment ultrasonic images of thyroid nodules. HA-UNet ++ improves backbone network structure at each stage of encoding path. At the same time, hybrid dilated convolution is added to the convolution blocks with three layers of convolution in the network, and attention mechanism is added to each convolution block, so that it can quickly predict the enhanced thyroid nodule data set. On this basis, thyroid nodules are labeled and segmented.
    Reference | Related Articles | Metrics | Comments0
    sORF-BERT: A Method on Identifying Coding sORFs Based on Pre-trained Models
    BIAN Xinye1, XIE Dongmei1, WANG Ziling1, QU Zhijian1, YU Jiafeng2
    Computer and Modernization    2025, 0 (03): 71-77.   DOI: 10.3969/j.issn.1006-2475.2025.03.011
    Abstract154)      PDF(pc) (3041KB)(121)       Save
     Small open reading frames (sORFs), which are open reading frames in the genome that do not exceed 300 bases in length, are identified as crucial for maintaining cellular metabolic balance and fundamental physiological functions of organisms. To excavate the deep characteristics of sORF sequences and to enhance the accuracy of cross-species prediction of coding and non-coding sORFs, a sORF-BERT neural network model is proposed. This model integrates DNABERT pre-training with a data blending encoding strategy and introduces the CAL module to learn multi-scale features of sORFs. Analyses are conducted on prokaryotic genomes, as well as human, mouse, arabidopsis, and escherichia coli datasets. After pre-training and fine-tuning, the sORF-BERT model can effectively capture the rich biological features of sORF sequences and utilize the CAL to better learn sORF features across different scales. Cross-species experimental comparisons of sORF-BERT with six published advanced methods, including CPPred, DeepCPP, CNCI, CPPred-sORF, MiPiped, and PsORFs, reveal that sORF-BERT improves performance across five independent test datasets. Compared to the second-ranked PsORFs, sORF-BERT shows increases of 0.42 ~ 18.72 percentage points in ACC and 1.08 ~ 11.75 percentage points in MCC, thereby demonstrating the superiority of this method in predicting coding sORFs and its potential to advance basic biological research.
    Related Articles | Metrics | Comments0
    Entity-integrated Summarization Model Based on Improved Graph2Seq 
    TAO Yuan, QIAN Huimin
    Computer and Modernization    2025, 0 (06): 1-8.   DOI: 10.3969/j.issn.1006-2475.2025.06.001
    Abstract154)      PDF(pc) (762KB)(140)       Save
    Abstract: In order to address the issues caused by high computational resource consumption and limited attention on key named entities, a novel summarization model named Entity-Sparse-Attention Graph-to-Sequence (ESG2S), based on the Graph2Seq model, is proposed in this paper. Firstly, a graph data is created from a syntactic dependency graph enhanced by the extracted entity nodes from the original text. Secondly, this graph data is then input into an encoder to learn the textual structure. Finally, the encoded graph data is fed into an LSTM decoder integrated with Symmetric Divergence-Enhanced Sparse Attention to generate multiple summaries. Experiments on the CNN/DM dataset show that this model outperforms several recent mainstream methods and is effective in preserving entity information, resulting in summaries with better readability and comprehensiveness.
    Reference | Related Articles | Metrics | Comments0
    Survey on Bundle Recommendation Algorithms
    LI Xiongqing1, 2, PENG Mingtian1, 2, LI Yong1, 2, WANG Junfei1, 2, LIU Dezhi1, 3, BIAN Yuxuan1, 3, CHAI Yuelin1, 3, LIU Yuntao1, 3
    Computer and Modernization    2025, 0 (09): 1-13.   DOI: 10.3969/j.issn.1006-2475.2025.09.001
    Abstract154)      PDF(pc) (975KB)(117)       Save

    Abstract: Bundle recommendation refers to optimizing and recommending the best solution by combining multiple related goods, services, or content, which can meet the various needs of users. With the rapid development of sectors like e-commerce and travel retail, bundle recommendation has become an important approach to improve user experience and business benefits. This paper reviews the research progress and application status of bundle recommendation algorithms. Firstly, the task definition, task characteristics, task challenges, and commonly used evaluation metrics are clarified. The task challenges include the integrity of bundled packages, diversity of bundled packages, data sparsity, cold start problems, and bundle generation problems. Secondly, the existing algorithms are classified into three major categories, data mining-based algorithms, traditional machine learning-based algorithms, deep learning-based algorithms, and further sorted out into seven subcategories. The characteristics of each category are thoroughly analyzed. Thirdly, commonly used datasets for the bundle recommendation task are summarized. Finally, the future development trends of bundle recommendation are discussed.

    Reference | Related Articles | Metrics | Comments0