Latest Articles

Articles in press have been peer-reviewed and accepted, which are not yet assigned to volumes/issues, but are citable by Digital Object Identifier (DOI).
Display Method:
A Space–Time Joint Waveform for Frequency Diverse Array Radar with Spatial Linear Frequency Modulation Weighting
LAN Yu, ZHOU Jianxiong
Available online  , doi: 10.11999/JEIT250561
Abstract:
  Objective  Frequency Diverse Array (FDA) radar exhibits a fast time-varying beampattern and a space–time coupled steering vector, offering potential advantages for multi-target tracking, wide-area surveillance, and mainlobe interference suppression. However, the beampattern of conventional coherent FDA radar is narrow, resulting in a shorter beam dwell time than that of phased arrays. This limitation prevents the ambiguity function of conventional coherent FDA from achieving both high range resolution and low sidelobe level simultaneously. When the baseband signal is modulated with a Linear Frequency Modulation (LFM) waveform, the ambiguity function presents low range resolution and low sidelobe level. Conversely, when the baseband signal is modulated with a phase-coded waveform, it achieves high range resolution but exhibits high sidelobe levels with strip-like high-gain sidelobes. The degradation in range resolution or sidelobe performance significantly constrains detection capability. To address this problem, this study proposes a novel space–time joint FDA waveform with spatial LFM weighting, which simultaneously achieves high range resolution, low sidelobe level, and reduced Doppler sensitivity.  Methods  The spatial-domain modulation scheme and the time-domain baseband waveform are two interdependent factors that determine the ambiguity function performance of FDA radar. Selecting a time-domain baseband waveform with a thumbtack-shaped ambiguity function enables the range resolution to remain independent of space–time coupling. By modulating the spatial weighting phase, the beampattern shape of the FDA can be adjusted to extend beam dwell time, suppress strip-like high-gain sidelobes, and smooth sidelobe energy distribution. The proposed space–time joint waveform thus achieves both high range resolution and low sidelobe level. Doppler tolerance is another key metric for evaluating ambiguity function performance. A space–time joint waveform with spatial phase-coded weighting exhibits high Doppler sensitivity, leading to significantly elevated sidelobe levels and sharp reductions in transmit beamforming gain. In contrast, the spatial LFM weighting method proposed in this study enhances Doppler tolerance while maintaining desirable range and sidelobe characteristics.  Results and Discussions  By combining the spatial LFM weighting method with a time-domain baseband waveform exhibiting a thumbtack-shaped ambiguity function (e.g., a phase-coded waveform), this study addresses the limitation of conventional coherent FDA waveforms, which cannot simultaneously achieve high range resolution and low sidelobe level. The proposed waveform demonstrates robust pulse compression performance, even under target motion. Simulation experiments were conducted to analyze the ambiguity functions under both stationary and motion conditions, and the results are summarized as follows: (1) The average sidelobe levels near the target peak for the space–time joint FDA waveform with spatial LFM weighting and spatial phase-coded weighting are both approximately –30 dB (Fig.3(a)(b)). In comparison, the average sidelobe level near the target peak for the spatial phase-coded weighting FDA using a time-domain LFM baseband waveform is about –10 dB (Fig.3(c)), while that of the coherent FDA with a time-domain phase-coded waveform is about –12 dB (Fig.3(d)). Thus, the two space–time joint FDA waveforms achieve the lowest average sidelobe levels. (2) The imaging results of both space–time joint FDA waveforms show no strip-like high-gain sidelobes (Fig.4(a)(b)). By contrast, the spatial phase-coded weighting FDA and the coherent FDA with a time-domain phase-coded waveform both display prominent high-gain sidelobes (Fig.4(c)(d)). These sidelobes from high Signal-to-Noise Ratio (SNR) targets can obscure nearby low-SNR targets. (3) All four FDA waveforms achieve a range resolution of 0.75 m (Fig. 5), corresponding to a bandwidth of 200 MHz. (4) Under motion conditions, the space–time joint FDA waveform with spatial phase-coded weighting exhibits a notable increase in peak sidelobe level compared with stationary conditions (Fig. 6(a)). In contrast, the space–time joint FDA waveform with spatial LFM weighting maintains the lowest peak sidelobe level among all four FDA configurations (Fig. 6(b)).  Conclusions  This study proposes a space–time joint FDA waveform with spatial LFM weighting. The proposed waveform effectively resolves the issue of degraded range resolution in conventional coherent FDA systems, ensuring that range resolution depends solely on bandwidth. It also eliminates the strip-like high-gain sidelobes commonly observed in conventional FDA waveforms. Under simulation conditions, the average sidelobe level near the target peak is reduced by approximately 10 dB and 18 dB compared with those of the spatial phase-coded weighting FDA and the coherent FDA with a time-domain phase-coded waveform, respectively. This reduction substantially mitigates the masking of low–SNR targets by sidelobes from high-SNR targets and demonstrates strong Doppler tolerance. However, under relative motion conditions, the proposed waveform exhibits Doppler–angle coupling, which will be addressed in future research through the development of coupling mitigation strategies.
Adaptive Cache Deployment Based on Congestion Awareness and Content Value in LEO Satellite Networks
LIU Zhongyu, XIE Yaqin, ZHANG Yu, ZHU Jianyue
Available online  , doi: 10.11999/JEIT250670
Abstract:
  Objective  Low Earth Orbit (LEO) satellite networks are central to future space–air–ground integrated systems, offering global coverage and low-latency communication. However, their high-speed mobility leads to rapidly changing topologies, and strict onboard cache constraints hinder efficient content delivery. Existing caching strategies often overlook real-time network congestion and content attributes (e.g., freshness), which leads to inefficient resource use and degraded Quality of Service (QoS). To address these limitations, we propose an adaptive cache placement strategy based on congestion awareness. The strategy dynamically couples real-time network conditions, including link congestion and latency, with a content value assessment model that incorporates both popularity and freshness.This integrated approach enhances cache hit rates, reduces backhaul load, and improves user QoS in highly dynamic LEO satellite environments, enabling efficient content delivery even under fluctuating traffic demands and resource constraints.  Methods  The proposed strategy combines a dual-threshold congestion detection mechanism with a multi-dimensional content valuation model. It proceeds in three steps. First, satellite nodes monitor link congestion in real time using dual latency thresholds and relay congestion status to downstream nodes through data packets. Second, a two-dimensional content value model is constructed that integrates popularity and freshness. Popularity is updated dynamically using an Exponential Weighted Moving Average (EWMA), which balances historical and recent request patterns to capture temporal variations in demand. Freshness is evaluated according to the remaining data lifetime, ensuring that expired or near-expired content is deprioritized to maintain cache efficiency and relevance. Third, caching thresholds are adaptively adjusted according to congestion level, and a hop count control factor is introduced to guide caching decisions. This coordinated mechanism enables the system to prioritize high-value content while mitigating congestion, thereby improving overall responsiveness and user QoS.  Results and Discussions  Simulations conducted on ndnSIM demonstrate the superiority of the proposed strategy over PaCC (Popularity-Aware Closeness-based Caching), LCE (Leave Copy Everywhere), LCD (Leave Copy Down), and Prob (probability-based caching with probability = 0.5). The key findings are as follows. (1) Cache hit rate. The proposed strategy consistently outperforms conventional methods. As shown in Fig. 8, the cache hit rate rises markedly with increasing cache capacity and Zipf parameter, exceeding those of LCE, LCD, and Prob. Specifically, the proposed strategy achieves improvements of 43.7% over LCE, 25.3% over LCD, 17.6% over Prob, and 9.5% over PaCC. Under high content concentration (i.e., larger Zipf parameters), the improvement reaches 29.1% compared with LCE, highlighting the strong capability of the strategy in promoting high-value content distribution. (2) Average routing hop ratio. The proposed strategy also reduces routing hops compared with the baselines. As shown in Fig. 9, the average hop ratio decreases as cache capacity and Zipf parameter increase. Relative to PaCC, the proposed strategy lowers the average hop ratio by 2.24%, indicating that content is cached closer to users, thereby shortening request paths and improving routing efficiency. (3) Average request latency. The proposed strategy achieves consistently lower latency than all baseline methods. As summarized in Table 2 and Fig. 10, the reduction is more pronounced under larger cache capacities and higher Zipf parameters. For instance, with a cache capacity of 100 MB, latency decreases by approximately 2.9%, 5.8%, 9.0%, and 10.3% compared with PaCC, Prob, LCD, and LCE, respectively. When the Zipf parameter is 1.0, latency reductions reach 2.7%, 5.7%, 7.2%, and 8.8% relative to PaCC, Prob, LCD, and LCE, respectively. Concretely, under a cache capacity of 100 MB and Zipf parameter of 1.0, the average request latency of the proposed strategy is 212.37 ms, compared with 236.67 ms (LCE), 233.45 ms (LCD), 225.42 ms (Prob), and 218.62 ms (PaCC).  Conclusions  This paper presents a congestion-aware adaptive caching placement strategy for LEO satellite networks. By combining real-time congestion monitoring with multi-dimensional content valuation that considers both dynamic popularity and freshness, the strategy achieves balanced improvements in caching efficiency and network stability. Simulation results show that the proposed method markedly enhances cache hit rates, reduces average routing hops, and lowers request latency compared with existing schemes such as PaCC, Prob, LCD, and LCE. These benefits hold across different cache sizes and request distributions, particularly under resource-constrained or highly dynamic conditions, confirming the strategy’s adaptability to LEO environments. The main innovations include a closed-loop feedback mechanism for congestion status, dynamic adjustment of caching thresholds, and hop-aware content placement, which together improve resource utilization and user QoS. This work provides a lightweight and robust foundation for high-performance content delivery in satellite–terrestrial integrated networks. Future extensions will incorporate service-type differentiation (e.g., delay-sensitive vs. bandwidth-intensive services), and orbital prediction to proactively optimize cache migration and updates, further enhancing efficiency and adaptability in 6G-enabled LEO networks.
Design and Optimization for Orbital Angular Momentum–based wireless-powered Noma Communication System
CHEN Ruirui, CHEN Yu, RAN Jiale, SUN Yanjing, LI Song
Available online  , doi: 10.11999/JEIT250634
Abstract:
  Objective  The Internet of Things (IoT) requires not only interconnection among devices but also seamless connectivity among users, information, and things. Ensuring stable operation and extending the lifespan of IoT Devices (IDs) through continuous power supply have become urgent challenges in IoT-driven Sixth-Generation (6G) communications. Radio Frequency (RF) signals can simultaneously transmit information and energy, forming the basis for Simultaneous Wireless Information and Power Transfer (SWIPT). Non-Orthogonal Multiple Access (NOMA), a key technology in Fifth-Generation (5G) communications, enables multiple users to share the same time and frequency resources. Efficient wireless-powered NOMA communication requires a Line-of-Sight (LoS) channel. However, the strong correlation in LoS channels severely limits the degree of freedom, making it difficult for conventional spatial multiplexing to achieve capacity gains. To address this limitation, this study designs an Orbital Angular Momentum (OAM)-based wireless-powered NOMA communication system. By exploiting OAM mode multiplexing, multiple data streams can be transmitted independently through orthogonal OAM modes, thereby significantly enhancing communication capacity in LoS channels.  Methods  The OAM-based wireless-powered NOMA communication system is designed to enable simultaneous energy transfer and multi-channel information transmission for IDs under LoS conditions. Under the constraints of the communication capacity threshold and the harvested energy threshold, this study formulates a sum-capacity maximization problem by converting harvested energy into the achievable uplink information capacity. The optimization problem is decomposed into two subproblems. A closed-form expression for the optimal Power-Splitting (PS) factor is derived, and the optimal power allocation is obtained using the subgradient method. The transmitting Uniform Circular Array (UCA) employs the Movable Antenna (MA) technique to adjust both position and array angle. To maintain system performance under typical parallel misalignment conditions, a beam-steering method is investigated.  Results and Discussions  Simulation results demonstrate that the proposed OAM-based wireless-powered NOMA communication system effectively enhances capacity performance compared with conventional wireless communication systems. As the OAM mode increases, the sum capacity of the ID decreases. This occurs because higher OAM modes exhibit stronger hollow divergence characteristics, resulting in greater energy attenuation of the received OAM signals (Fig. 3). The sum capacity of the ID increases with the PS factor (Fig. 4). However, as the harvested energy threshold increases, the system’s sum capacity decreases (Fig. 5). When the communication capacity threshold increases, the sum capacity first rises and then gradually declines (Fig. 6). In power allocation optimization, allocating more power to the ID with the best channel condition further improves the total system capacity.  Conclusions  To enhance communication capacity under LoS conditions, this study designs an OAM-based wireless-powered NOMA communication system that employs mode multiplexing to achieve independent multi-channel information transmission. On this basis, a sum-capacity maximization problem is formulated under communication capacity and harvested energy threshold constraints by transforming harvested energy into achievable uplink information capacity. The optimization problem is decomposed into two subproblems. A closed-form expression for the optimal PS factor is derived, and the optimal power allocation is obtained using the subgradient method. In future work, the MA technique will be integrated into the proposed OAM-based wireless-powered NOMA system to further optimize sum-capacity performance based on the three-dimensional spatial configuration and adjustable array angle.
Modeling, Detection, and Defense Theories and Methods for Cyber-Physical Fusion Attacks in Smart Grid
WANG Wenting, TIAN Boyan, WU Fazong, HE Yunpeng, WANG Xin, YANG Ming, FENG Dongqin
Available online  , doi: 10.11999/JEIT250659
Abstract:
  Significance   Smart Grid (SG), the core of modern power systems, enables efficient energy management and dynamic regulation through cyber–physical integration. However, its high interconnectivity makes it a prime target for cyberattacks, including False Data Injection Attacks (FDIAs) and Denial-of-Service (DoS) attacks. These threats jeopardize the stability of power grids and may trigger severe consequences such as large-scale blackouts. Therefore, advancing research on the modeling, detection, and defense of cyber–physical attacks is essential to ensure the safe and reliable operation of SGs.  Progress   Significant progress has been achieved in cyber–physical security research for SGs. In attack modeling, discrete linear time-invariant system models effectively capture diverse attack patterns. Detection technologies are advancing rapidly, with physical-based methods (e.g., physical watermarking and moving target defense) complementing intelligent algorithms (e.g., deep learning and reinforcement learning). Defense systems are also being strengthened: lightweight encryption and blockchain technologies are applied to prevention, security-optimized Phasor Measurement Unit (PMU) deployment enhances equipment protection, and response mechanisms are being continuously refined.  Conclusions  Current research still requires improvement in attack modeling accuracy and real-time detection algorithms. Future work should focus on developing collaborative protection mechanisms between the cyber and physical layers, designing solutions that balance security with cost-effectiveness, and validating defense effectiveness through high-fidelity simulation platforms. This study establishes a systematic theoretical framework and technical roadmap for SG security, providing essential insights for safeguarding critical infrastructure.  Prospects   Future research should advance in several directions: (1) deepening synergistic defense mechanisms between the information and physical layers; (2) prioritizing the development of cost-effective security solutions; (3) constructing high-fidelity information–physical simulation platforms to support research; and (4) exploring the application of emerging technologies such as digital twins and interpretable Artificial Intelligence (AI).
A Learning-Based Security Control Method for Cyber-Physical Systems Based on False Data Detection
MIAO Jinzhao, LIU Jinliang, SUN Le, ZHA Lijuan, TIAN Engang
Available online  , doi: 10.11999/JEIT250537
Abstract:
  Objective  Cyber-Physical Systems (CPS) constitute the backbone of critical infrastructures and industrial applications, but the tight coupling of cyber and physical components renders them highly susceptible to cyberattacks. False data injection attacks are particularly dangerous because they compromise sensor integrity, mislead controllers, and can trigger severe system failures. Existing control strategies often assume reliable sensor data and lack resilience under adversarial conditions. Furthermore, most conventional approaches decouple attack detection from control adaptation, leading to delayed or ineffective responses to dynamic threats. To overcome these limitations, this study develops a unified secure learning control framework that integrates real-time attack detection with adaptive control policy learning. By enabling the dynamic identification and mitigation of false data injection attacks, the proposed method enhances both stability and performance of CPS under uncertain and adversarial environments.  Methods  To address false data injection attacks in CPS, this study proposes an integrated secure control framework that combines attack detection, state estimation, and adaptive control strategy learning. A sensor grouping-based security assessment index is first developed to detect anomalous sensor data in real time without requiring prior knowledge of attacks. Next, a multi-source sensor fusion estimation method is introduced to reconstruct the system’s true state, thereby improving accuracy and robustness under adversarial disturbances. Finally, an adaptive learning control algorithm is designed, in which dynamic weight updating via gradient descent approximates the optimal control policy online. This unified framework enhances both steady-state performance and resilience of CPS against sophisticated attack scenarios. Its effectiveness and security performance are validated through simulation studies under diverse false data injection attack settings.  Results and Discussions  Simulation results confirm the effectiveness of the proposed secure adaptive learning control framework under multiple false data injection attacks in CPS. As shown in Fig. 1, system states rapidly converge to steady values and maintain stability despite sensor attacks. Fig. 2 demonstrates that the fused state estimator tracks the true system state with greater accuracy than individual local estimators. In Fig. 3, the compensated observation outputs align closely with the original, uncorrupted measurements, indicating precise attack estimation. Fig. 4 shows that detection indicators for sensor groups 2–5 increase sharply during attack intervals, while unaffected sensors remain near zero, verifying timely and accurate detection. Fig. 5 further confirms that the estimated attack signals closely match the true injected values. Finally, Fig. 6 compares different control strategies, showing that the proposed method achieves faster stabilization and smaller state deviations. Together, these results demonstrate robust control, accurate state estimation, and real-time detection under unknown attack conditions.  Conclusions  This study addresses secure perception and control in CPS under false data injection attacks by developing an integrated adaptive learning control framework that unifies detection, estimation, and control. A sensor-level anomaly detection mechanism is introduced to identify and localize malicious data, substantially enhancing attack detection capability. The fusion-based state estimation method further improves reconstruction accuracy of true system states, even when observations are compromised. At the control level, an adaptive learning controller with online weight adjustment enables real-time approximation of the optimal control policy without requiring prior knowledge of the attack model. Future research will extend the proposed framework to broader application scenarios and evaluate its resilience under diverse attack environments.
Research on Skill-Aware Task Assignment Algorithm under Local Differential Privacy
FANG Xianjin, ZHEN Yaru, ZHANG Pengfei, HUANG Shanshan
Available online  , doi: 10.11999/JEIT250425
Abstract:
  Objective  With the proliferation of mobile smart devices and wireless networks, Spatial Crowdsourcing (SC) has emerged as a new paradigm for collaborative task execution. By leveraging workers’ real-time locations, SC platforms dynamically assign tasks to distributed participants. However, continuous exposure of location data creates privacy risks, including trajectory inference and identity disclosure, which reduce worker participation and threaten system sustainability. Existing privacy-preserving methods either rely on trusted third parties or apply traditional differential privacy mechanisms. The former incurs high costs and security vulnerabilities, whereas the latter struggles to balance the trade-off between noise magnitude and data utility, often reducing task matching accuracy. To address these challenges, this study proposes a skill-aware task assignment algorithm under Local Differential Privacy (LDP) that simultaneously enhances location privacy protection and task assignment performance. The algorithm is particularly effective in settings characterized by uneven skill distributions and complex task requirements.  Methods  To protect workers’ location privacy, a Clip–Laplace (CLP) mechanism is applied to perturb real-time location data under Local Differential Privacy (LDP), ensuring bounded noise while maintaining data utility. To mitigate mismatches between heterogeneous task requirements and imbalanced worker skills, an entropy-based metric is used to evaluate skill diversity. When entropy falls below a predefined threshold, a secondary screening strategy rebalances the distribution by suppressing common skills and prioritizing rare ones. A skill-aware Pruning Greedy task assignment algorithm (PUGR) is then developed. PUGR iteratively selects the worker–task pair with the highest marginal contribution to maximize skill coverage under spatiotemporal and budget constraints. To improve computational efficiency, three pruning strategies are integrated: time–distance pruning, high-reward pruning, and budget-infeasibility pruning. Finally, comparative and ablation experiments on three real-world datasets assess the method using multiple metrics, including Loss of Quality of Service (LQS), Average Remaining Budget Rate (ARBR), and Task Completion Rate (TCR).  Results and Discussions  Experimental results show that the CLP mechanism consistently achieves lower LQS than the traditional Laplace mechanism (LP) across different privacy budgets, effectively reducing errors introduced by noise (Fig. 2). For skill diversity, the entropy-based metric combined with secondary screening nearly doubles the average entropy of candidate workers on the TKY and NYC datasets, demonstrating its effectiveness in balancing skill distribution. In task assignment, the proposed PUGR algorithm completes most worker–task matchings within four iterations, thereby reducing redundant computation and accelerating convergence (Fig. 3). Regarding budget utilization, the ARBR under CLP remains close to the No Privacy (NoPriv) baseline, indicating efficient resource allocation (Fig. 4, Table 2). For task completion, the method achieves a TCR of up to 90% in noise-free settings and consistently outperforms Greedy, OE-ELA, and TsPY under CLP (Fig. 5). Ablation studies further validate the contributions of secondary screening and pruning strategies to overall performance improvement.  Conclusions  This study addresses two central challenges in spatial crowdsourcing: protecting workers’ location privacy and improving skill-aware task assignment. A task assignment framework is proposed that integrates the CLP mechanism with a skill-aware strategy under the LDP model. The CLP mechanism provides strong privacy guarantees while preserving data utility by limiting noise magnitude. An entropy-based metric combined with secondary screening ensures balanced skill distribution, substantially enhancing skill coverage and task execution success in multi-skill scenarios. The PUGR algorithm incorporates skill contribution evaluation with multiple pruning constraints, thereby reducing the search space and improving computational efficiency. Experiments on real-world datasets demonstrate the method’s superiority in terms of LQS, ARBR, and TCR, confirming its robustness, scalability, and effectiveness in balancing privacy protection with assignment performance. Future work will explore dynamic pricing mechanisms based on skill scarcity and personalized, adaptive incentives to foster fairness, long-term worker engagement, and the sustainable development of spatial crowdsourcing platforms.
Multi-modal Joint Automatic Modulation Recognition Method Towards Low SNR Sequences
WANG Zhen, LIU Wei, LU Wanjie, NIU Chaoyang, LI Runsheng
Available online  , doi: 10.11999/JEIT250594
Abstract:
  Objective  The rapid evolution of data-driven intelligent algorithms and the rise of multi-modal data indicate that the future of Automatic Modulation Recognition (AMR) lies in joint approaches that integrate multiple domains, use multiple frameworks, and connect multiple scales. However, the embedding spaces of different modalities are heterogeneous, and existing models lack cross-modal adaptive representation, limiting their ability to achieve collaborative interpretation. To address this challenge, this study proposes a performance-interpretable two-stage deep learning–based AMR (DL-AMR) method that jointly models the signal in the time and transform domains. The approach explicitly and implicitly represents signals from multiple perspectives, including temporal, spatial, frequency, and intensity dimensions. This design provides theoretical support for multi-modal AMR and offers an intelligent solution for modeling low Signal-to-Noise Ratio (SNR) time sequences in open environments.  Methods  The proposed AMR network begins with a preprocessing stage, where the input signal is represented as an in-phase and quadrature (I–Q) sequence. After wavelet thresholding denoising, the signal is converted into a dual-channel representation, with one channel undergoing Short-Time Fourier transform (STFT). This preprocessing yields a dual-stream representation comprising both time-domain and transform-domain signals. The signal is then tokenized through time-domain and transform-domain encoders. In the first stage, explicit modal alignment is performed. The token sequences from the time and transform domains are input in parallel into a contrastive learning module, which explicitly captures and strengthens correlations between the two modalities in dimensions such as temporal structure and amplitude. The learned features are then passed into the feature fusion module. Bidirectional Long Short-Term Memory (BiLSTM) and local representation layers are employed to capture temporally sparse features, enabling subsequent feature decomposition and reconstruction. To refine feature extraction, a subspace attention mechanism is applied to the high-dimensional sparse feature space, allowing efficient capture of discriminative information contained in both high-frequency and low-frequency components. Finally, Convolutional Neural Network – Kolmogorov-Arnold Network (CNN-KAN) layers replace traditional multilayer perceptrons as classifiers, thereby enhancing classification performance under low SNR conditions.  Results and Discussions  The proposed method is experimentally validated on three datasets: RML2016.10a, RML2016.10b, and HisarMod2019.1. Under high SNR conditions (SNR > 0 dB), classification accuracies of 93.36%, 93.13%, and 93.37% are achieved on the three datasets, respectively. Under low SNR conditions, where signals are severely corrupted or blurred by noise, recognition performance decreases but remains robust. When the SNR ranges from –6 dB to 0 dB, overall accuracies of 78.36%, 80.72%, and 85.43% are maintained, respectively. Even at SNR levels below –6 dB, accuracies of 17.10%, 21.30%, and 29.85% are obtained. At particularly challenging low-SNR levels, the model still achieves 43.45%, 44.54%, and 60.02%. Compared with traditional approaches, and while maintaining a low parameter count (0.33–0.41 M), the proposed method improves average recognition accuracy by 2.12–7.89%, 0.45–4.64%, and 6.18–9.53% on the three datasets. The improvements under low SNR conditions are especially significant, reaching 4.89–12.70% (RML2016.10a), 2.62–8.72% (RML2016.10b), and 4.96–11.63% (HisarMod2019.1). The results indicate that explicit modeling of time–transform domain correlations through contrastive learning, combined with the hybrid architecture consisting of LSTM for temporal sequence modeling, CNN for local feature extraction, and KAN for nonlinear approximation, substantially enhances the noise robustness of the model.  Conclusions  This study proposes a two-stage AMR method based on time–transform domain multimodal fusion. Explicit multimodal alignment is achieved through contrastive learning, while temporal and local features are extracted using a combination of LSTM and CNN. The KAN is used to enhance nonlinear modeling, enabling implicit feature-level multimodal fusion. Experiments conducted on three benchmark datasets demonstrate that, compared with classical methods, the proposed approach improves recognition accuracy by 2.62–11.63% within the SNR range of –20 to 0 dB, while maintaining a similar number of parameters. The performance gains are particularly significant under low-SNR conditions, confirming the effectiveness of multimodal joint modeling for robust AMR.
UAV-Assisted Intelligent Data Collection and Computation Offloading for Railway Wireless Sensor Networks
YAN Li, WANG Junkai, FANG Xuming, LIN Wei, LIANG Yiqun
Available online  , doi: 10.11999/JEIT250340
Abstract:
  Objective  Ensuring the safety and stability of train operations is essential in the advancement of railway intelligence. The growing maturity of Wireless Sensor Network (WSN) technology offers an efficient, reliable, low-cost, and easily deployable approach to monitoring railway operating conditions. However, in complex and dynamic maintenance environments, WSNs encounter several challenges, including weak signal coverage at monitoring sites, limited accessibility for tasks such as sensor node battery replacement, and the generation of large volumes of monitoring data. To address these issues, this study proposes a multi-Unmanned Aerial Vehicle (UAV)-assisted method for data collection and computation offloading in railway WSNs. This approach enhances overall system energy efficiency and data freshness, offering a more effective and robust solution for railway safety monitoring.  Methods  An intelligent data collection and computation offloading system is constructed for multi-UAV-assisted railway WSNs. UAV flight constraints within railway safety protection zones are considered, and wireless sensing services are prioritized to ensure preferential transmission for safety-critical tasks. To balance energy consumption and data freshness, the system optimization objective is defined as the weighted sum of UAV energy consumption, WSN energy consumption, and the Age of Information (AoI). A joint optimization algorithm based on Multi-Agent Soft Actor-Critic (MASAC) is proposed, which balances exploration and exploitation through entropy regularization and adaptive temperature parameters. This approach enables efficient joint optimization of UAV trajectories and computation offloading strategies.  Results and Discussions  (1) Compared with the Multi-Agent Deep Deterministic Policy Gradient (MADDPG), MASAC-Greedy, and MASAC-AOU algorithms, the MASAC-based scheme converges more rapidly and demonstrates greater stability (Fig. 4), ultimately achieving the highest reward. In contrast, MADDPG exhibits slower learning and less stable performance. (2) The comparison of multi-UAV flight trajectories under different algorithms shows that the proposed MASAC algorithm enables effective collaboration among UAVs, with each responsible for monitoring distinct regions while strictly adhering to railway safety protection zone constraints (Fig. 5). (3) The MASAC algorithm yields the best objective function value across all evaluated algorithms (Fig. 6). (4) As the number of sensors and the AoI weight increase, UAV energy consumption rises for all algorithms; however, the MASAC algorithm consistently maintains the lowest energy consumption (Fig. 7). (5) In terms of sensor node energy consumption, MADDPG achieves the lowest value, but at the expense of information freshness (Fig. 8). (6) Regarding average AoI performance, the MASAC algorithm performs best across a range of sensor densities and AoI weight settings, with the greatest improvements observed under higher AoI weight conditions (Fig. 9). (7) The AoI performance comparison by sensor type (Table 2) confirms that the system effectively supports priority-based data collection services.  Conclusions  This study proposes a MASAC-based intelligent data collection and computation offloading scheme for railway WSNs supported by multiple UAVs, addressing critical challenges such as limited WSN battery life and the high real-time computational demands of complex railway environments. The proposed algorithm jointly optimizes UAV flight trajectories and computation offloading strategies by integrating considerations of UAV and WSN energy consumption, data freshness, sensing service priorities, and railway safety protection zone constraints. The optimization objective is to minimize the weighted sum of average UAV energy consumption, average WSN energy consumption, and average WSN AoI. Simulation results demonstrate that the proposed scheme outperforms baseline algorithms across multiple performance metrics. Specifically, it achieves faster convergence, efficient multi-UAV collaboration that avoids resource redundancy and spatial overlap, and superior results in UAV energy consumption, sensor node energy consumption, and average AoI.
Research and Design of a Ballistocardiogram-Based Heart Rate Variability (HRV) Monitoring Device Integrated into Pilot Helmets
ZHAO Yanpeng, LI Falin, LI Xuan, YU Haibo, CAO Zhengtao, ZHANG Yi
Available online  , doi: 10.11999/JEIT250342
Abstract:
  Objective  Conventional Heart Rate Variability (HRV) monitoring in aviation is limited by bulky wearable devices that require direct skin contact, are prone to electromagnetic interference during flight, and suffer from electrode displacement during high-G maneuvers. These constraints hinder continuous physiological monitoring, which is critical for flight safety. This study presents a non-contact monitoring approach integrated into pilot helmets, utilizing BallistoCardioGram (BCG) technology to detect cardiac mechanical activity via helmet-mounted inertial sensors. The objective is to establish a novel physiological monitoring paradigm that eliminates the need for skin–electrode interfaces while achieving measurement accuracy suitable for aviation operational standards.  Methods  Hardware ConfigurationA patented BCG sensing module is embedded within the occipital stabilization system of flight protective helmets. Miniaturized, high-sensitivity inertial sensors interface with proprietary signal conditioning circuits that execute a three-stage physiological signal refinement process. First, primary analog amplification scales microvolt-level inputs to measurable voltage ranges. Second, a fourth-order Butterworth bandpass filter (0.5–20 Hz) isolates cardiac mechanical signatures. Third, analog-to-digital conversion quantizes the signals at a 250 Hz sampling rate. Physical integration complies with military equipment standards for helmet structural integrity and ergonomic performance, ensuring full compatibility with existing flight gear without compromising protection or pilot comfort during extended missions.Computational Framework A multi-layer signal processing architecture is implemented to extract physiological features. Raw BCG signals undergo five-level discrete wavelet transformation using Daubechies-4 basis functions, effectively separating cardiac components from respiratory modulation and motion-induced artifacts. J-wave identification is achieved through dual-threshold detection: morphological amplitudes exceeding three times the local baseline standard deviation and temporal positioning within 200 ms sliding analysis windows. Extracted J–J intervals are treated as functional analogs of ElectroCardioGram (ECG)-derived R–R intervals. Time-domain HRV metrics are computed as follows: (1) Standard Deviation of NN intervals (SDNN), representing overall autonomic modulation; (2) Root Mean Square of Successive Differences (RMSSD), indicating parasympathetic activity; (3) Percentage of adjacent intervals differing by more than 50 ms (pNN50). Frequency-domain analysis applies Fourier transformation to quantify Low-Frequency (LF: 0.04–0.15 Hz) and High-Frequency (HF: 0.15–0.4 Hz) spectral powers. The LF/HF ratio is used to assess sympathetic–parasympathetic balance. The entire processing pipeline is optimized for real-time execution under in-flight operational conditions.  Results and Discussions  System validation is conducted under simulated flight conditions to evaluate physiological monitoring performance. Signal acquisition is found to be reliable across static, turbulent, and high-G scenarios, with consistent capture of BCG waveforms. Quantitative comparisons with synchronized ECG recordings show strong agreement between measurement modalities: (1) SDNN: 95.80%; (2) RMSSD: 94.08%; (3) LF/HF ratio: 92.86%. These results demonstrate that the system achieves physiological measurement equivalence to established clinical standards. Artifact suppression is effectively performed by the wavelet-based signal processing framework, which maintains waveform integrity under conditions of aircraft vibration and rapid gravitational transition—conditions where conventional ECG monitoring often fails. Among tested sensor placements, the occipital position exhibits the highest signal-to-noise ratio. Operational stability is maintained during continuous 6-hour monitoring sessions, with no observed signal degradation. This long-duration robustness indicates suitability for extended flight operations.Validation results indicate that the BCG-based approach addresses three primary limitations associated with ECG systems in aviation. The removal of electrode–skin contact mitigates the risk of contact dermatitis during prolonged wear. Non-contact sensing eliminates susceptibility to electromagnetic interference generated by radar and communication systems. Furthermore, mechanical coupling ensures signal continuity during abrupt gravitational changes, which typically displace ECG electrodes and cause signal dropout. The wavelet decomposition method is particularly effective in attenuating rotorcraft harmonic vibrations and turbulence-induced high-frequency noise. Autonomic nervous system modulation is reliably captured through pulse transit time variability, which aligns with neurocardiac regulation indices derived from ECG. Two operational considerations are identified. First, respiratory coupling under hyperventilation may introduce artifacts that require additional filtering. Second, extreme cervical flexion exceeding 45 degrees may degrade signal quality, indicating the potential benefit of redundant sensor configurations under such conditions.  Conclusions  This study establishes a functional, helmet-integrated BCG monitoring system capable of delivering medical-grade HRV metrics without compromising flight safety protocols. The technology represents a shift from contact-based to non-contact physiological monitoring in aviation settings. Future system development will incorporate: (1) Infrared eye-tracking modules to assess blink interval variability for objective fatigue evaluation; (2) Dry-contact electroencephalography sensors to quantify prefrontal cortex activity and assess cognitive workload; (3) Multimodal data fusion algorithms to generate unified indices of physiological strain. The integrated framework aims to enable real-time pilot state awareness during critical operations such as aerial combat maneuvers, hypoxia exposure, and emergency responses. Further technology maturation will prioritize operational validation across diverse aircraft platforms and environmental conditions. System implementation remains fully compliant with military equipment specifications and is positioned for future translation to commercial aviation and human factors research. Broader applications include astronaut physiological monitoring during spaceflight missions and enhanced safety systems in high-performance motorsports.
Mutualistic Backscatter NOMA Method for Coordinated Direct and Relay Transmission System
XU Yao, HU Rongfei, JIA Shaobo, LI Bo, WANG Gang, ZHANG Zhizhong
Available online  , doi: 10.11999/JEIT250405
Abstract:
  Objective  The exponential growth in data traffic necessitates that cellular Internet of Things (IoT) systems achieve both ultra-high spectral efficiency and wide-area coverage to meet the stringent service requirements of vertical applications such as industrial automation and smart cities. Non-Orthogonal Multiple Access-based Coordinated Direct and Relay Transmission (NOMA-CDRT) method can enhance both spectral efficiency and coverage by leveraging power-domain multiplexing and cooperative relaying, making it a promising approach to address these challenges. However, existing NOMA-CDRT frameworks are primarily designed for cellular communications and do not effectively support spectrum sharing or the deep integration of cellular and IoT transmissions. To overcome these limitations, this study proposes a Mutualistic Backscatter NOMA-CDRT (MB-NOMA-CDRT) method. This approach facilitates spectrum sharing and mutualistic coexistence between cellular users and IoT devices, while improving the system’s Ergodic Sum Rate (ESR).  Methods  The proposed MB-NOMA-CDRT method integrates backscatter modulation and power-domain superposition coding to develop a bidirectional communication strategy that unifies information transmission and cooperative assistance, enabling spectrum sharing and mutualistic coexistence between cellular users and IoT devices. Specifically, the base station uses downlink NOMA to serve the cellular center user directly and the cellular edge user via a relaying user. Simultaneously, IoT devices utilize cellular radio frequency signals and backscatter modulation to transmit their data to the base station, thereby achieving spectrum sharing. The backscattered IoT signals act as multipath gains, contributing to improved cellular communication quality. To rigorously characterize the system performance, the squared generalized-K distribution and Meijer-G functions are adopted to derive closed-form expressions for the ESR under both perfect and imperfect Successive Interference Cancellation (SIC). Building on this analytical foundation, a power allocation optimization scheme is developed using an enhanced Particle Swarm Optimization (PSO) algorithm to maximize system ESR. Finally, extensive Monte Carlo simulations are conducted to verify the ESR gains of the proposed method, confirm the theoretical analysis, and demonstrate the efficacy of the optimization scheme.  Results and Discussions  The performance advantage of the proposed MB-NOMA-CDRT method is demonstrated through comparisons of ESR with conventional NOMA-CDRT and Orthogonal Multiple Access (OMA) schemes (Fig. 2 and Fig. 3). The theoretical ESR results closely match the simulation data, confirming the validity of the analytical derivations. Under both perfect and imperfect SIC, the proposed method consistently achieves the highest ESR. This improvement arises from spectrum sharing between cellular users and IoT devices, where the IoT link contributes multipath gain to the cellular link, thereby enhancing overall system performance. To investigate the influence of power allocation, simulation results illustrate the three-dimensional relationship between ESR and power allocation coefficients (Fig. 4). A maximum ESR is observed under specific coefficient combinations, indicating that optimized power allocation can significantly improve system throughput. Furthermore, the proposed optimization scheme demonstrates rapid convergence, with ESR values stabilizing within a few iterations (Fig. 5), supporting its computational efficiency. Finally, ESR performance is compared among the proposed optimization scheme, exhaustive search, and fixed power allocation strategies (Fig. 6). The proposed scheme consistently yields higher ESR across both perfect and imperfect SIC scenarios, demonstrating its superiority in enhancing spectral efficiency while maintaining low computational complexity.  Conclusions  This study proposes a MB-NOMA-CDRT method that enables spectrum sharing between IoT devices and cellular users while improving cellular communication quality through the backscatter-assisted reflection link. To evaluate system performance, closed-form expressions for the ESR are derived under both perfect and imperfect SIC. Building on this analytical foundation, a power allocation optimization scheme based on PSO is developed to maximize the system ESR. Simulation results demonstrate that the proposed method consistently outperforms conventional NOMA-CDRT and OMA schemes in terms of ESR, under both perfect and imperfect SIC conditions. The optimization scheme also exhibits favorable convergence behavior and effectively improves system performance. Given its advantages in spectral efficiency and computational efficiency, the proposed MB-NOMA-CDRT method is well suited to cellular IoT scenarios. Future work will focus on exploring the mathematical conditions necessary to fully characterize and exploit the mutualistic transmission mechanism.
Electromagnetic Finite-Difference Time-Domain Scattering Analysis of Multilayered/Porous Materials in Specific Geometric Meshing
ZHANG Yuxian, YANG Zijiang, HUANG Zhixiang, FENG Xiaoli, FENG Naixing, YANG Lixia
Available online  , doi: 10.11999/JEIT250348
Abstract:
The Finite-Difference Time-Domain (FDTD) method is a widely used tool for analyzing the electromagnetic properties of dielectric media, but its application is often constrained by model complexity and mesh discretization. To enhance the efficiency of electromagnetic scattering simulations in multilayered/porous materials, we proposes an accelerated FDTD scheme in this paper. Computational geometry algorithms can be employed with the proposed method to rapidly generate Yee’s grids, utilizing a three-dimensional voxel array to define material distributions and field components. By exploiting the voxel characteristics, parallel algorithms are employed to efficiently compute Radar Cross Sections (RCS) for non-analytical geometries. In contrast to conventional volumetric mesh generation, which relies on analytic formulas, this work integrates ray-intersection techniques with Signed Distance Functions (SDFs). Calculations of tangent planes and intersection points minimize invalid traversals and reduce computational complexity, thus expediting grid-based electromagnetic parameter assignment for porous and irregular structures. The approach is applied to the RCS calculations of multilayered/porous models, demonstrating excellent consistency with results from popular commercial solvers (FEKO, CST, HFSS) while offering substantially higher efficiency. Numerical experiments confirm significant reductions in computation time and computer memory without compromising accuracy. Overall, the proposed acceleration scheme enhances the FDTD method’s ability to handle complex dielectric structures, providing an effective balance between computational speed and accuracy, and offering innovative solutions for rapid mesh generation and processing of complex internal geometries.  Objective   The FDTD method, a reliable approach for computing the electromagnetic properties of dielectric media, faces constraints in computational efficiency and accuracy due to model structure and mesh discretization. A major challenge in the field is achieving efficient electromagnetic scattering analysis with minimal computational resources while maintaining sufficient wavelength sampling resolution. To address this difficulty, we propose an FDTD-based electromagnetic analysis acceleration scheme that enhances simulation efficiency by significantly improving mesh generation and optimizing grid partitioning for complex multilayered/porous models.  Methods   In this study, those Yee’s grids for complex materials are efficiently generated using computational geometry algorithms and a 3D voxel array to define material distribution and field components. A parallel algorithm leverages voxel data to accelerate RCS calculations for non-analytical geometries. Unlike conventional volumetric meshing methods that rely on analytic formulas, this approach integrates ray-intersection techniques with SDFs. Calculations of tangent planes and intersection points further reduce invalid traversals and geometric complexity, facilitating faster grid-based assignment of electromagnetic parameters. Numerical experiments validate that the method effectively supports porous and multilayered non-analytical structures, demonstrating both high efficiency and accuracy.  Results and Discussions   The accelerated volumetric meshing algorithm is validated using a Boeing 737 model, showing more than a 67.5% reduction in computation time across different resolutions. Efficiency decreases at very fine meshes because of heavier computational loads and suboptimal valid-grid ratios. The method is further evaluated on three multilayered/porous structures, achieving 85.55% faster computation and 9.8% lower memory usage compared with conventional FDTD. In comparison with commercial solvers (FEKO, CST, HFSS), equivalent accuracy is maintained while runtimes are reduced by 87.58% and memory consumption by 81.6%. In all tested cases, errors remain below 6% relative to high-resolution FDTD, confirming that the proposed acceleration scheme provides both high efficiency and reliable accuracy.  Conclusions   In this study, we optimize volumetric mesh generation in FDTD through computational geometry algorithms. By combining ray-intersection techniques with reliable SDFs, the proposed approach efficiently manages internal cavities, while tangent-plane calculations minimize traversal operations and complexity, thereby accelerating scattering analysis. The scheme extends the applicability of FDTD to a broader range of dielectric structures and materials, delivering substantial savings in computation time and memory without compromising accuracy. Designed to support universal geometric model files, the framework shows strong potential for stealth optimization of multi-material structures and the development of electromagnetic scattering systems. It represents an important step toward integrating computational geometry with computational electromagnetics.
The effects of ELF-MF on Aβ42 deposition in AD mice and SWM-related neural oscillations
GENG Duyan, LIU Aoge, YAN Yuxin, ZHENG Weiran
Available online  , doi: 10.11999/JEIT241106
Abstract:
  Objective  Extremely Low-Frequency Magnetic Fields (ELF-MF) have shown beneficial effects in various diseases; however, their influence on Alzheimer’s Disease (AD) remains insufficiently understood. With global population aging, AD has become one of the most prevalent neurodegenerative disorders. Its complex pathogenesis is characterized by neuronal loss, extracellular Amyloid-β (Aβ) deposition, and intracellular neurofibrillary tangles. Cognitive decline, particularly Spatial Working Memory (SWM) impairment, is among its main clinical manifestations. As a crucial cognitive function for encoding and retaining spatial location information, SWM underpins the execution of complex cognitive tasks. Impairment of SWM not only affects daily functioning but also serves as a key indicator of AD progression. Although previous studies have suggested potential cognitive benefits of ELF-MF exposure, systematic investigations integrating pathological, behavioral, and electrophysiological analyses remain limited. This study aims to investigate whether 40 Hz ELF-MF exposure mitigates AD pathology by assessing Aβ42 deposition, SWM performance, and neural oscillatory activity in the hippocampal CA1 region, and to elucidate the relationships between electrophysiological modulation and behavioral improvement.  Methods  An integrated multidisciplinary approach combining immunofluorescence detection, behavioral assessment, and electrophysiological recording is employed. Transgenic AD model mice and Wild-Type (WT) controls are used and assigned to three groups: WT control (Con), AD model group (AD), and AD model group exposed to ELF-MF stimulation (ES). The ES group receives 40 Hz, 10 mT continuous pulse stimulation twice daily for 0.5 h per session over 14 consecutive days, whereas the AD and Con groups undergo sham stimulation during identical time periods. SWM is evaluated using the Object Location Task (OLT). Behavioral performance is quantitatively determined by calculating the Cognitive Index (CI), which reflects the animal’s capacity to recognize spatial novelty. During behavioral testing, Local Field Potential (LFP) signals are synchronously recorded from the hippocampal CA1 region via chronically implanted microelectrodes. Advanced signal processing techniques, including time-frequency distribution analysis and phase-amplitude coupling computation, are applied to characterize neural oscillations within the theta (4~13 Hz) and gamma (30~80 Hz) frequency bands. After completion of the experiments, brain tissues are collected for quantitative measurement of Aβ42 plaque deposition in hippocampal sections through immunofluorescence staining, using standardized imaging and quantification protocols. Statistical analyses are performed to evaluate correlations between behavioral indices and electrophysiological parameters, with the objective of identifying mechanistic relationships underlying the effects of ELF-MF exposure.  Results and Discussions  Exposure to 40 Hz ELF-MF produced significant therapeutic effects across all examined parameters. Pathological analysis revealed markedly reduced Aβ42 deposition in the hippocampal region of treated AD mice compared with untreated controls, supporting the amyloid cascade hypothesis, which identifies Aβ oligomers as critical triggers of neurodegeneration. This reduction suggests that ELF-MF may influence Aβ metabolic pathways, potentially through the regulation of mitochondrial dynamics, as reported in previous studies. Behavioral assessment indicated a pronounced improvement in SWM following ELF-MF exposure, reflected by significantly elevated CI scores in the OLT. Electrophysiological recordings revealed notable alterations in neural oscillatory activity, with treated animals exhibiting increased power spectral density in both theta (4~13 Hz) and gamma (30~80 Hz) bands during memory task performance. The temporal dynamics of theta oscillations also differed among groups: in Con and ES mice, peak theta power occurred approximately 0.5~1 seconds before the behavioral reference point, indicating anticipatory processing, whereas in AD mice, peaks appeared after the reference point, reflecting delayed cognitive responses. Cross-frequency coupling analysis further demonstrated enhanced theta-gamma phase-amplitude coupling strength in the hippocampal CA1 region of ELF-MF-exposed mice, with coupling peaks primarily observed in the lower theta and higher gamma frequencies. Correlation analyses revealed statistically significant positive relationships between behavioral cognitive indices and electrophysiological measures, particularly for theta power and theta-gamma coupling strength. These convergent findings across pathological, behavioral, and electrophysiological domains indicate that ELF-MF exposure may restore impaired neural synchronization mechanisms. Enhanced theta-gamma coupling is particularly relevant, as this neurophysiological mechanism is known to facilitate temporal coordination among neuronal assemblies during memory processing. Although the present study demonstrates clear benefits of ELF-MF stimulation, heterogeneity in previously reported results warrants consideration. The efficacy of ELF-MF appears highly dependent on key stimulation parameters such as frequency, intensity, duration, and exposure intervals. Previous studies have reported divergent effects, ranging from negligible or adverse outcomes to substantial cognitive enhancement under different experimental conditions. This parameter dependency presents challenges for clinical translation and highlights the need for systematic optimization in higher-order animal models.  Conclusions  This study demonstrates that exposure to a 40 Hz ELF-MF effectively reduces Aβ42 deposition in the hippocampal region of AD mice, alleviates SWM deficits, and normalizes neural oscillatory activity in the hippocampal CA1 region. The observed cognitive improvements are closely linked to enhanced oscillations in the theta and gamma frequency bands and to strengthened theta-gamma cross-frequency coupling, indicating that neuromodulatory regulation of neural synchronization underlies behavioral recovery. These findings provide strong evidence supporting the potential of ELF-MF as a noninvasive therapeutic approach for AD, targeting both pathological markers and functional impairments. The study establishes a foundation for future work aimed at optimizing stimulation parameters and advancing translational applications, while highlighting the central role of neural oscillatory restoration as a therapeutic mechanism in neurodegenerative disorders. Further investigations should focus on refining exposure protocols and developing personalized stimulation strategies to accommodate individual variability in treatment responsiveness.
A Cross-Dimensional Collaborative Framework for Header-Metadata-Driven Encrypted Traffic Identification
WANG Menghan, ZHOU Zhengchun, JI Qingbing
Available online  , doi: 10.11999/JEIT250434
Abstract:
  Objective  With the widespread adoption of network communication encryption technologies, encrypted traffic identification has become a critical problem in network security. Traditional identification methods based on payload content face the risk of feature invalidation due to the continuous evolution of encryption algorithms, leading to detection blind spots in dynamic network environments. Meanwhile, the structured information embedded in packet headers, an essential carrier for protocol interaction, remains underutilized. Furthermore, as encryption protocols evolve, existing encrypted traffic identification approaches encounter limitations such as poor feature interpretability and weak model robustness against adversarial attacks. To address these challenges, this paper proposes a cross-dimensional collaborative identification framework for encrypted traffic, driven by header metadata features. The framework systematically reveals and demonstrates the dominant role of header features in encrypted traffic identification, overcoming the constraints of single-perspective analyses and reducing dependence on payload data. It further enables the assessment of deep model performance boundaries and decision credibility. Through effective feature screening and pruning, redundant attributes are eliminated, enhancing the framework’s anti-interference capability in encrypted scenarios. This approach reduces model complexity while improving interpretability and robustness, facilitating the design of lighter and more reliable encrypted traffic identification models.  Methods  This study performs a three-dimensional analysis including (1) network traffic feature selection and identification performance, (2) quantitative evaluation of feature importance in classification, and (3) assessment of model robustness under adversarial perturbations. First, the characteristics, differences, and effects on identification performance are compared among three forms of encrypted traffic packets using a One-Dimensional Convolutional Neural Network (1D-CNN). This comparison verifies the dominant role of header features in encrypted traffic identification. Second, two explainable algorithms, Layer-wise Relevance Propagation (LRP) and Deep Taylor Decomposition (DTD), are employed to further confirm the essential contribution of header features to network traffic classification. The relative importance of header and payload features is quantified from two perspectives: (i) the relevance of backpropagation and (ii) the contribution coefficients derived from Taylor series expansion, thereby enhancing feature interpretability. Finally, adversarial attack experiments are conducted using Projected Gradient Descent (PGD) and random perturbations. By injecting carefully constructed adversarial perturbation data into the initial and terminal parts of the payload, or by adding randomly generated noise to produce adversarial traffic, the study examines the effect of these perturbations on model decision-making. This analysis evaluates the stability and anti-interference capabilities of the encrypted traffic identification model under adversarial conditions.  Results and Discussions  Comparative experiments conducted on the ISCXVPN2016 and ISCXTor2016 datasets yield three key findings. (1) Recognition performance. The model based solely on header features achieves an F1 score up to 6% higher than that of the model using complete traffic, and up to 61% higher than that of the model using only payload features. These results verify that header features possess irreplaceable significance in encrypted traffic identification. The structural information embedded in headers plays a dominant role in enabling the model to accurately classify traffic types. Even without payload data, high identification accuracy can be achieved using header information alone (Figure 2, Table 4). (2) Interpretability evaluation. The LRP and DTD methods are used to quantify the contribution of header features to model classification. The correlation between header features and classification performance is markedly higher than that of payload features, with the average proportion of the correlation score up to 89.8% greater (Figures 3~4, Table 5). This result is highly consistent with the classification behavior of the One-Dimensional Convolutional Neural Network (1D-CNN), further confirming the critical importance and dominant influence of header features in encrypted traffic identification. (3) Anti-interference robustness. The combined Header–Payload model exhibits strong robustness under adversarial attacks. Particularly under low-bandwidth conditions, the model incorporating header features shows a markedly higher maximum performance retention rate under equivalent bandwidth perturbation than the pure payload model, with the maximum difference reaching 98.46%. This finding confirms the essential role of header features in enhancing model robustness (Figures 5~6). Header-based models maintain stable recognition performance, whereas payload information is more susceptible to interference, leading to sharp performance degradation. In addition, the identification performance, contribution quantification, and anti-attack effectiveness of header features are influenced by data type and distribution characteristics. In certain cases, payload features provide auxiliary support, suggesting a complementary relationship between the two feature domains.  Conclusions  This study addresses core challenges in encrypted traffic identification, including feature degradation, limited interpretability, and weak adversarial robustness in traditional payload-dependent methods. A cross-dimensional collaborative identification framework driven by header features is proposed. Through systematic theoretical analysis and experimental validation from three perspectives, the framework demonstrates the irreplaceable value of header features in network traffic identification and overcomes the limitations of conventional single-perspective approaches. It provides a theoretical foundation for improving the efficiency, interpretability, and robustness of encrypted traffic identification models. Future work will focus on enhancing dynamic adaptability, integrating multi-modal features, implementing lightweight architectures, and strengthening adversarial defense mechanisms. These directions are expected to advance encrypted traffic identification technology toward higher intelligence, adaptability, and resilience.
A Waveform Design for Integrated Radar and Jamming Based on Intra-Pulse and Inter-Pulse Multiple-Phase Modulation
ZHANG Shiyuan, LU Xingyu, YAN Huabin, YANG Jianchao, TAN Ke, GU Hong
Available online  , doi: 10.11999/JEIT250600
Abstract:
  Objective  An integrated radar-jamming waveform employing multiple-phase modulation both within pulses (intra-pulse) and between pulses (inter-pulse) is proposed. The design increases the degrees of freedom in waveform synthesis compared with existing integrated signals, thereby improving joint performance in detection and jamming. In detection, phase compensation and complementary synthesis of received echoes are used to reconstruct a Linear Frequency Modulation (LFM) waveform, preserving the range resolution and ambiguity characteristics of LFM. In jamming, multi-parameter control of phase both across and within pulses allows flexible adjustment of the jamming energy distribution in the adversary’s range-Doppler map, enabling targeted energy allocation and concealment strategies. Simulation and experimental results show that the proposed waveform enhances overall detection and jamming performance relative to conventional integrated designs.  Methods  An integrated waveform that combines intra-pulse and inter-pulse multi-phase modulation is proposed. Carefully designed inter-pulse phase perturbations are introduced to prevent jamming energy from concentrating at zero Doppler and to allow precise control of the Doppler distribution of the jamming signal. During echo processing, the inter-pulse perturbations are removed by phase compensation so that inter-pulse complementarity reconstructs a continuous LFM waveform, thereby preserving detection performance. Each pulse is encoded with a binary phase-coded sequence, and additional phase modulation is applied between pulses. The resulting waveform has multiple tunable parameters and increased degrees of freedom, achieves low-sidelobe detection comparable to LFM, and permits flexible allocation of jamming energy across the range-Doppler plane.  Results and Discussions  The proposed integrated waveform is evaluated through simulations and practical experiments. Detection performance is significantly enhanced, with the Signal-to-Clutter-Noise Ratio (SCNR) for moving-target detection reaching 63.46 dB, representing a 25.25 dB improvement over conventional integrated waveforms and only 3.57 dB lower than that of a reference LFM signal (67.03 dB). These findings demonstrate that phase compensation and inter-pulse complementarity effectively enhance target detectability. Jamming performance is governed by the range of inter-pulse random phase perturbations. When the perturbation range is 0°, jamming energy is concentrated in the zero-Doppler main lobe, resulting in limited target masking. Expanding the range to ±90° flattens the Doppler spectrum and substantially weakens the target signature. Further extending the range to ±180° eliminates the zero-frequency main peak and achieves near-uniform diffusion of jamming energy across the Doppler domain. Therefore, by varying the inter-pulse phase range, continuous adjustment between concentrated and distributed jamming energy allocation is achieved. Overall, the waveform maintains detection performance comparable to that of optimal LFM signals while enabling flexible, parameterized control of jamming energy distribution. This design provides an adaptable solution for integrated radar-jamming systems that achieves a balance between efficient detection and adaptive jamming capability.  Conclusions  This study is based on a previously proposed integrated radar-jamming waveform and focuses on solving the problem of uneven jamming energy distribution in the unoptimized design. An integrated radar-jamming waveform based on combined intra-pulse and inter-pulse multiple-phase modulation is proposed by introducing random phase modulation between pulses. The proposed waveform achieves detection performance comparable to that of LFM signals and provides flexible control of jamming effects through multiple adjustable parameters, offering high design freedom. Theoretical analysis shows that intra-pulse modulation alone is insufficiently adaptable. The addition of random inter-pulse phases with variable distribution ranges enables more precise regulation of jamming energy diffusion. Simulation results indicate that increasing the range of inter-pulse phase perturbation leads to progressively wider diffusion of jamming energy, while detection performance remains similar to that of LFM. Therefore, by adjusting the distribution range of inter-pulse phases, the jamming energy pattern can be flexibly shaped, providing greater degrees of freedom in waveform design. Experimental results verify that the proposed waveform exhibits good overall performance in both detection and jamming. However, its practical application remains limited by specific operational conditions, which will be addressed in future studies.
A Spatial-semantic Combine Perception for Infrared UAV Target Tracking
YU Guodong, JIANG Yichun, LIU Yunqing, WANG Yijun, ZHAN Weida, WANG Chunyang, FENG Jianghai, HAN Yueyi
Available online  , doi: 10.11999/JEIT250613
Abstract:
  Objective  In recent years, infrared image-based UAV target tracking technology has attracted widespread attention. In real-world scenarios, infrared UAV target tracking still faces significant challenges due to factors such as complex backgrounds, UAV target deformation, and camera movement. Siamese network-based tracking methods have made breakthroughs in balancing tracking accuracy and efficiency. However, existing approaches rely solely on high-level feature outputs from deep networks to predict target positions, neglecting the effective use of low-level features. This leads to the loss of spatial detail features of infrared UAV targets, severely affecting tracking performance. To efficiently utilize low-level features, some methods have incorporated Feature Pyramid Networks (FPN) into the tracking framework, progressively fusing cross-layer feature maps in a top-down manner, thereby effectively enhancing tracking performance for multi-scale targets. Nevertheless, these methods directly adopt traditional FPN channel reduction operations, which result in significant loss of spatial contextual information and channel semantic information. To address the above issues, a novel infrared UAV target tracking method based on spatial-semantic combine perception is proposed. By capturing spatial multi-scale features and channel semantic information, the proposed approach enhances the model’s capability to track infrared UAV targets in complex backgrounds.  Methods  The proposed method comprises four main components: a backbone network, multi-scale feature fusion, template-search feature interaction, and a detection head. Initially, template and search images containing infrared UAV targets are input into a weight-sharing backbone network to extract features. Subsequently, an FPN is constructed, within which a Spatial-semantic Combine Attention Module (SCAM) is integrated to efficiently fuse multi-scale features. Finally, a Dual-branch global Feature interaction Module (DFM) is employed to facilitate feature interaction between the template and search branches, and the final tracking results are obtained through the detection head. The proposed SCAM enhances the network’s focus on spatial and semantic information by jointly leveraging spatial and channel attention mechanisms, thereby mitigating the loss of spatial and semantic information in low-level features caused by channel dimensionality reduction in traditional FPN. SCAM primarily consists of two components: the Spatial Multi-scale Attention module (SMA) and the Global-local Channel Semantic Attention module (GCSA). The SMA captures long-range multi-scale dependencies efficiently through axial positional embedding and multi-branch grouped feature extraction, thereby improving the network’s perception of global contextual information. GCSA adopts a dual-branch design to effectively integrate global and local information across feature channels, suppress irrelevant background noise, and enable more rational channel-wise feature weighting. The proposed DFM treats the template branch features as the query source for the search branch and applies global cross-attention to capture more comprehensive features of infrared UAV targets. This enhances the tracking network’s ability to attend to the spatial location and boundary details of infrared UAV targets.  Results and Discussions  The proposed method has been validated on the infrared UAV benchmark dataset (Anti-UAV). Quantitative analysis (Table 1) demonstrates that, compared to 10 state-of-the-art methods, the proposed approach achieves the highest average normalized precision score of 76.9%, surpassing the second-best method, LGTrack, by 4.4%. In terms of success rate and localization precision (Table 1), the proposed method also outperforms LGTrack by 4.7% and 2.1%, respectively, evidencing its superiority in infrared UAV target tracking. Qualitative analysis (Figs. 68) further confirms that the proposed method exhibits strong adaptability and robustness when addressing various typical challenges in infrared UAV tracking, such as occlusion, distracting objects, complex backgrounds, scale variations, and rapid deformations. The collaborative design of the individual modules significantly enhances the model’s ability to perceive and represent small targets and dynamic scenes. In addition, qualitative experiments (Fig. 9) conducted on a self-constructed infrared UAV tracking dataset demonstrate the effectiveness and generalization capability of the proposed method in real-world tracking scenarios. Ablation studies (Tables 24) reveal that integrating any individual proposed module consistently improves tracking performance. Compared with the baseline tracker, the integration of all sub-modules leads to improvements of 14.3% in average normalized precision, 12.5% in success rate, and 14.0% in localization precision, thereby verifying the effectiveness of the proposed components.  Conclusions  This paper conducts a systematic theoretical analysis and experimental validation addressing the issue of spatial and semantic information loss in infrared UAV target tracking. Focusing on the limitations of existing FPN-based infrared UAV tracking methods, particularly the drawbacks associated with channel reduction in multi-scale low-level features, a novel infrared UAV target tracking method based on spatial-semantic combine perception is proposed which fully leverages the complementary advantages of spatial and channel attention mechanisms. This method enhances the network’s focus on spatial context and critical semantic information, thereby improving overall tracking performance. The following main conclusions are obtained: (1) The proposed SCAM combining SMA and GCSA, where SMA captures spatial long-range feature dependencies through position coordinate embedding and one-dimensional convolution operations, ensuring the acquisition of multi-scale contextual information, while GCSA achieves more comprehensive semantic feature attention by interacting local and global channel features. (2) The designed DFM, which realizes feature interaction between search branch features and template branch features through global cross-attention, enabling the dual-branch features to complement each other and enhancing network tracking performance. (3) Extensive experimental results demonstrate that the proposed algorithm outperforms existing advanced methods in both quantitative evaluation and qualitative analysis, with an average state accuracy of 0.769, success rate of 0.743, and precision of 0.935, achieving more accurate tracking of infrared UAV targets. Although the algorithm in this paper has been optimized in terms of computing resource utilization efficiency, further research is needed on efficient deployment strategies for embedded and mobile devices to improve real-time performance and computing adaptability.
VCodePPA: A Large-scale Verilog Dataset with PPA Annotations
CHEN Xiyuan, JIANG Yuxuan, XIA Yingjie, HU Ji, ZHOU Yizhao
Available online  , doi: 10.11999/JEIT250449
Abstract:
  Objective  As a predominant hardware description language, the quality of Verilog code directly affects the Power, Performance, and Area (PPA) metrics of the resulting circuits. Current Large Language Model (LLM)-based approaches for generating hardware description languages face a central challenge: incorporating a design feedback mechanism informed by PPA metrics to guide model optimization, rather than relying solely on syntactic and functional correctness. The field faces three major limitations: the absence of PPA metric annotations in training data, which prevents models from learning the effects of code modifications on physical characteristics; evaluation frameworks that remain disconnected from downstream engineering needs; and the lack of systematic data augmentation methods to generate functionally equivalent code with differentiated PPA characteristics. To address these gaps, we present VCodePPA, a large-scale dataset that establishes precise correlations between Verilog code structures and PPA metrics. The dataset comprises 17 342 entries and provides a foundation for data-driven optimization paradigms in hardware design.  Methods  The dataset construction is initiated by collecting representative Verilog code samples from GitHub repositories, OpenCores projects, and standard textbooks. After careful selection, a seed dataset of 3 500 samples covering 20 functional categories is established. These samples are preprocessed through functional coverage optimization, syntax verification with Yosys, format standardization, deduplication, and complexity filtering. An automated PPA extraction pipeline is implemented in Vivado to evaluate performance characteristics, with metrics including LookUp Table (LUT) count, register usage, maximum operating frequency, and power consumption. To enhance dataset diversity while preserving functional equivalence, a multi-dimensional code transformation framework is applied, consisting of nine methods across three dimensions: architecture layer (finite state machine encoding, interface protocol reconstruction, arithmetic unit replacement), logic layer (control flow reorganization, operator rewriting, logic hierarchy restructuring), and timing layer (critical path cutting, register retiming, pipeline insertion or deletion). Efficient exploration of the transformation space is achieved through a Heterogeneous Verilog Mutation Search (HVMS) algorithm based on Monte Carlo Tree Search, which generates 5~10 PPA-differentiated variants for each seed code. A dual-task LLM training strategy with PPA-guided adaptive loss functions is subsequently employed, incorporating contrastive learning mechanisms to capture the relationship between code structure and physical implementation.  Results and Discussions  The VCodePPA dataset achieves broad coverage of digital hardware design scenarios, representing approximately 85%~90% of common design contexts. The multi-dimensional transformation framework generates functionally equivalent yet structurally diverse code variants, with PPA differences exceeding 20%, thereby exposing optimization trade-offs inherent in hardware design. Experimental evaluation demonstrates that models trained with VCodePPA show marked improvements in PPA optimization across multiple Verilog functional categories, including arithmetic, memory, control, and hybrid modules. In testing scenarios, VCodePPA-trained models produced implementations with superior PPA metrics compared with baseline models. The PPA-oriented adaptive loss function effectively overcame the traditional limitation of language model training, which typically lacks sensitivity to hardware implementation efficiency. By integrating contrastive learning and variant comparison loss mechanisms, the model achieved an average improvement of 17.7% across PPA metrics on the test set, influencing 32.4% of token-level predictions in code generation tasks. Notably, VCodePPA-trained models reduced on-chip resource usage by 10%\begin{document}$ \sim $\end{document}15%, decreased power consumption by 8%\begin{document}$ \sim $\end{document}12%, and shortened critical path delay by 5%\begin{document}$ \sim $\end{document}8% relative to baseline models.  Conclusions  This paper introduces VCodePPA, a large-scale Verilog dataset with precise PPA annotations, addressing the gap between code generation and physical implementation optimization. The main contributions are as follows: (1)construction of a seed dataset spanning 20 functional categories with 3 500 samples, expanded through systematic multi-dimensional code transformation to 17 000 entries with comprehensive PPA metrics; (2)development of an MCTS-based homogeneous code augmentation scheme employing nine transformers across architectural, logical, and timing layers to generate functionally equivalent code variants with significant PPA differences; and (3)design of a dual-task training framework with PPA-oriented adaptive loss functions, enabling models to learn PPA trade-off principles directly from data rather than relying on manual heuristics or single-objective constraints. Experimental results demonstrate that models trained on VCodePPA effectively capture PPA balancing principles and generate optimized hardware description code. Future work will extend the dataset to more complex design scenarios and explore advanced optimization strategies for specialized application domains.
Bayesian Optimization-driven Design Space Exploration Method for Coarse-Grained Reconfigurable Cipher Logic Array
JIANG Danping, DAI Zibin, LIU Yanjiang, ZHOU Zhaoxu, SONG Xiaoyu
Available online  , doi: 10.11999/JEIT250624
Abstract:
  Objective  Coarse-Grained Reconfigurable Cipher logic Arrays (CGRCAs) are widely employed in information security systems owing to their high flexibility, strong performance, and inherent security. Design Space Exploration (DSE) plays a critical role in evaluating and optimizing the performance of cryptographic algorithms deployed on CGRCAs. However, conventional DSE approaches require extensive computation time to locate optimal solutions in multi-objective optimization problems and often yield suboptimal performance. To overcome these limitations, this study proposes a Bayesian optimization-based DSE framework, termed Multi-Objective Bayesian optimization-based Exploration (MOBE), which enhances search efficiency and solution quality while effectively satisfying the complex design requirements of CGRCA architectures.  Methods  The high-dimensional characteristics and multi-objective optimization features of the CGRCA are analyzed, and its design space is systematically modeled. A DSE method based on Bayesian optimization is then proposed, comprising initial sampling design, rapid evaluation model construction, surrogate model development, and acquisition function optimization. A knowledge-aware unsupervised learning sampling strategy is introduced to integrate domain-specific knowledge with clustering algorithms, thereby improving the representativeness and diversity of the initial samples. A rapid evaluation model is established to estimate throughput, area overhead, and Function Unit (FU) utilization for each sample, effectively reducing the computational cost of performance evaluation. To enhance both search efficiency and generalizability, a greedy-based hybrid surrogate model is constructed by combining Gaussian Process with Deep Kernel Learning (DKL-GP), random forest, and neural network models. Moreover, an adaptive multi-acquisition function is designed by integrating Expected Hyper Volume Improvement (EHVI) and quasi-Monte Carlo Upper Confidence Bound (qUCB) to identify the most promising samples and maintain a balanced trade-off between exploration and exploitation. The weighting ratio between EHVI and qUCB is dynamically adjusted to accommodate the varying optimization requirements across different search phases.  Results and Discussions  The DSE method based on Bayesian optimization (Algorithm 2) includes initial sampling design, rapid evaluation model construction, surrogate model development, and acquisition function optimization to enhance solution quality and search efficiency. Simulation results show that the knowledge-aware unsupervised learning sampling strategy reduces the Average Distance from Reference Set (ADRS) by up to 28.2% and increases hypervolume by 15.1% compared with existing sampling approaches (Table 3). This improvement primarily arises from the integration of domain knowledge with clustering algorithms. Compared with single surrogate model-based DSE methods, the greedy-based hybrid surrogate model leverages the complementary advantages of multiple surrogate models across different optimization stages, prioritizing samples that contribute most to hypervolume expansion. The hybrid surrogate model achieves a reduction in ADRS of up to 31.7% and an improvement in hypervolume of 20.0% (Table 4). Furthermore, the proposed MOBE framework achieves a 34.9% reduction in ADRS and increases hypervolume by 28.7% relative to state-of-the-art DSE methods (Table 5). Regarding the average performance metrics of Pareto-front samples, MOBE enhances throughput by up to 29.9%, reduces area overhead by 6.0%, and improves FU utilization by 11.6% (Fig. 6), confirming its superiority in overall solution quality. Moreover, the MOBE method exhibits excellent cross-algorithm stability in both hypervolume and Normalized Overall Execution Time (NOET) (Table 6 and Fig. 7).  Conclusions  This study presents a multi-objective DSE method based on Bayesian optimization that enhances both solution quality and search efficiency for CGRCA. The proposed approach employs a knowledge-aware unsupervised learning sampling strategy to generate an initial sample set with high representativeness and diversity. A rapid evaluation model is subsequently developed to reduce the computational cost of performance assessments. Additionally, the integration of adaptive multi-acquisition functions with a greedy-based hybrid surrogate model further improves the efficiency and generalization capability of the DSE framework. Comparative experiments demonstrate the effectiveness of the proposed MOBE method: (1) the sampling strategy reduces the ADRS by up to 28.2% and increases hypervolume by 15.1% compared with existing methods; (2) the greedy-based hybrid surrogate model achieves up to a 31.7% reduction in ADRS and a 20.0% improvement in hypervolume relative to single surrogate model-based approaches; (3) the overall MOBE framework achieves a 34.9% reduction in ADRS and a 28.7% increase in hypervolume compared with state-of-the-art DSE techniques; (4) MOBE improves throughput by up to 29.9%, reduces area overhead by 6.0%, and increases FU utilization by 11.6% relative to existing methods; and (5) MOBE exhibits excellent cross-algorithm stability in hypervolume and NOET. MOBE is applicable to medium-and-high-performance cryptographic application scenarios, including cloud platforms and desktop terminals. Nevertheless, two limitations remain. First, MOBE currently employs only traditional surrogate models, which may constrain feature learning efficiency and modeling accuracy. Second, its validation is confined to a CGRCA architecture previously developed by the research group, lacking verification across existing CGRCA architectures. Future work will address these limitations by incorporating emerging artificial intelligence techniques, such as large models, and conducting extensive experiments on diverse CGRCA architectures to further enhance the generalization and effectiveness of MOBE.
A 3D Underwater Target Tracking Algorithm with Integrated Grubbs-Information Entropy and Improved Particle Filter
CAI Fanglin, WANG Ji, QIU Haowei
Available online  , doi: 10.11999/JEIT250249
Abstract:
  Objective  To address the limited target tracking accuracy of traditional Particle Filter (PF) algorithms in three-dimensional Underwater Wireless Sensor Networks (UWSNs) under abnormal conditions, this study proposes a three-dimensional underwater target tracking algorithm (OGIE-IPF). The algorithm integrates an optimized Grubbs criterion–based information entropy-weighted data fusion with an Improved Particle Filter (IPF). Conventional PF algorithms often suffer from particle degeneracy and impoverishment, which restrict estimation accuracy. Although weight optimization strategies introduced during resampling can enhance particle diversity, existing approaches mainly rely on fixed weighting factors that cannot dynamically adapt to real-time operating conditions. Moreover, current anomaly detection methods for multi-source data fusion fail to effectively address data coupling and heteroscedasticity across dimensions. To overcome these challenges, a dynamic adaptive hierarchical weight optimization strategy is designed for the resampling phase, enabling adaptive particle weighting across hierarchy levels. Additionally, a Mahalanobis distance discrimination mechanism is incorporated into the Grubbs criterion-based anomaly detection method, achieving effective multi-dimensional anomaly detection through covariance-sensitive analysis.  Methods  The proposed OGIE-IPF algorithm enhances target tracking accuracy under underwater abnormal conditions through a distributed data processing framework that integrates multi-source data fusion and adaptive filtering. First, the Unscented Kalman Filter (UKF) is incorporated into the particle filtering framework to construct the importance density function, thereby alleviating particle degeneracy. Simultaneously, a dynamic adaptive hierarchical weight optimization mechanism is proposed during the resampling phase to improve particle diversity. Second, the Mahalanobis distance replaces the conventional standardized residual method in the standard Grubbs criterion for anomaly statistic construction. By incorporating the covariance matrix of multidimensional variables, the method achieves effective anomaly detection for multi-dimensional data. Finally, local target tracking is performed using the IPF combined with the optimized Grubbs criterion for anomaly detection and sensor credibility evaluation, whereas global state estimation is realized through an information entropy-weighted multi-source fusion algorithm.  Results and Discussions  The IPF developed in this study is designed to enhance particle set diversity through optimization of the importance density function and refinement of the resampling strategy. To evaluate algorithm performance, a comparative experimental group with a particle population of 100 is established. Simulation results indicate that the weight distribution variances of the IPF at specific time points and over the entire tracking period are reduced by approximately 98.27% and 97.26%, respectively, compared with the traditional PF (Figs. 2 and 3). These findings suggest that the improved strategy effectively regulates particles with varying weights, resulting in a balanced distribution across hierarchical weight levels. Sensor anomalies are simulated by introducing substantial perturbations in observation noise. The experimental data show that the OGIE-IPF algorithm maintains optimal error metrics throughout the operational period (Figs. 4 and 5), demonstrating superior capability in suppressing abnormal noise interference. To further assess algorithm robustness, two representative scenarios under low-noise and high-noise conditions are constructed for multi-algorithm comparison. The results indicate that OGIE-IPF achieves Root Mean Square Error (RMSE) reductions of 79.78%, 66.78%, and 56.41% compared with the PF, Extended Particle Filter (EPF), and Unscented Particle Filter (UPF) under low-noise conditions, and reductions of 83.41%, 70.38%, and 21.68% under high-noise conditions (Figs. 8 and 11).  Conclusions  The OGIE-IPF algorithm proposed in this study enhances target tracking accuracy in three-dimensional underwater environments through two synergistic mechanisms. First, tracking precision is improved by refining the PF framework to optimize the intrinsic accuracy of the filtering process. Second, data fusion reliability is strengthened via an anomaly detection framework that mitigates interference from erroneous sensor measurements. Simulation results confirm that the OGIE-IPF algorithm produces state estimations more consistent with ground truth trajectories than conventional PF, EPF, and UPF algorithms, achieving lower RMSE and maintaining stable tracking performance under limited particle populations and abnormal noise conditions. Future work will extend the model to incorporate dynamic marine environmental factors and address the effects of malicious node interference within underwater network security systems.
A Hybrid Beamforming Algorithm Based on Riemannian Manifold Optimization with Non-Monotonic Line Search
YAN Junrong, SHI Weitao, LI Pei
Available online  , doi: 10.11999/JEIT250396
Abstract:
  Objective  Fully digital beamforming architectures provide high spectral efficiency but demand one Radio-Frequency (RF) chain per antenna element, resulting in substantial cost, power consumption, and hardware complexity. These limitations hinder their practical deployment in large-scale antenna systems. Hybrid beamforming offers a feasible alternative by reducing hardware requirements while retaining much of the performance. In such systems, analog beamforming modules follow a reduced number of RF chains to control massive antenna arrays. Analog phase shifters are energy-efficient and cost-effective but restricted to constant modulus constraints, which are essential for hardware implementation. In contrast, digital phase shifters offer flexible control over amplitude and phase. The central challenge is to approximate the spectral efficiency of fully digital systems while adhering to analog-domain constraints and minimizing energy and hardware demands. To overcome this challenge, this study proposes a novel hybrid beamforming algorithm that integrates Riemannian manifold optimization with a non-monotonic line search strategy (MO-NMLS). This approach achieves improved trade-offs among spectral efficiency, energy consumption, and hardware complexity.  Methods  The proposed methodology proceeds as follows. First, the joint matrix optimization problem for maximizing spectral efficiency in hybrid beamforming is decomposed into separate transmitter and receiver subproblems by formulating an appropriate objective function. This objective is then reformulated using a least squares approach, reducing the dimensionality of the search space from two to one. To accommodate the constant modulus constraints of analog beamforming, the problem is transformed into an unconstrained optimization on Riemannian manifolds. Both the Euclidean and Riemannian gradients of the modified objective function are derived analytically. Step sizes are adaptively determined using a MO-NMLS, which incorporates historical gradient information to compute dynamic step factors. This mechanism guides the search direction while avoiding convergence to suboptimal local minima due to fixed step sizes. Distinct update rules for the step factor are applied depending on whether the iteration count is odd or even. In each iteration, the current objective function value is compared with those from the preceding L iterations to decide whether to accept the new step and iteration point. After updating the step size, tangent vectors are retracted onto the manifold to generate new iterates until convergence criteria are satisfied. Once the analog precoder is fixed based on the optimized search direction, the corresponding digital precoder is derived in closed form. The dynamic step factor is computed using gradient data from the current and preceding L iterations, allowing the objective function to exhibit non-strict monotonicity within bounded ranges. This adaptive strategy results in faster convergence compared with conventional fixed-step methods.  Results and Discussions  The relationship between internal iteration count and Signal-to-Noise Ratio (SNR) for different beamforming algorithms is shown in Fig. 4. The MO-NMLS algorithm requires significantly fewer iterations than the conventional Conjugate Gradient (CG) method under both fully connected and overlapping subarray architectures. This improved efficiency arises from the use of Riemannian manifold optimization, which inherently satisfies the constant modulus constraints without necessitating computationally intensive Hessian matrix evaluations. Runtime performance is benchmarked in Fig. 5. The MO-NMLS algorithm reduces runtime by 75.3% relative to CG in the fully connected structure and by 79.2% in the overlapping subarray structure. Additionally, MO-NMLS achieves a further 21.1% reduction in runtime under the overlapping subarray architecture compared with the fully connected one, owing to simplified hardware requirements. Spectral efficiency as a function of SNR is presented in Fig. 6. In fully connected systems, MO-NMLS achieves a 0.64% improvement in spectral efficiency over CG while maintaining comparable stability in overlapping subarray architectures. This performance gain stems from the algorithm’s ability to avoid local optima, a key limitation of Orthogonal Matching Pursuit (OMP), which selects paths based solely on residual correlation. The scalability of MO-NMLS with respect to the number of antennas and data streams is demonstrated in Fig. 7. In fully connected systems, MO-NMLS outperforms CG by 1.94%, 2.16%, and 2.74% in spectral efficiency at antenna and data stream configurations of (32, 2), (64, 4), and (128, 8), respectively. While spectral efficiency increases across all algorithms as system scale grows, MO-NMLS exhibits the most substantial gains at higher scales. Energy efficiency improvements under the overlapping subarray architecture are shown in Fig. 8. Compared with the fully connected configuration, MO-NMLS yields energy efficiency gains of 1.2%, 10.9%, and 25.9% at subarray offsets of 1, 8, and 16, respectively. These improvements are attributed to the reduced number of required phase shifters and power amplifiers, which decreases total system power consumption as the subarray offset increases.  Conclusions  The proposed MO-NMLS algorithm achieves an effective balance among spectral efficiency, hardware complexity, and energy consumption in hybrid beamforming systems, while substantially reducing computational runtime. Moreover, the overlapping subarray architecture attains spectral efficiency comparable to that of fully connected systems, with significantly lower execution times. These results highlight the practical advantages of the proposed approach for large-scale antenna systems operating under resource constraints.
Weakly Supervised Recognition of Aerial Adversarial Maneuvers via Contrastive Learning
ZHU Longjun, YUAN Weiwei, MEN Xuefeng, TONG Wei, WU Qi
Available online  , doi: 10.11999/JEIT250495
Abstract:
  Objective  Accurate recognition of aerial adversarial maneuvers is essential for situational awareness and tactical decision-making in modern air warfare. Conventional supervised approaches face major challenges: obtaining labeled flight data is costly due to the intensive human effort required for collection and annotation, and these methods are limited in capturing temporal dependencies inherent in sequential flight parameters. Temporal dynamics are crucial for describing the evolution of maneuvers, yet existing models fail to fully exploit this information. To address these challenges, this study proposes a weakly supervised maneuver recognition framework based on contrastive learning. The method leverages a small proportion of labeled data to learn discriminative representations, thereby reducing reliance on extensive manual annotations. The proposed framework enhances recognition accuracy in data-scarce scenarios and provides a robust solution for maneuver analysis in dynamic adversarial aerial environments.  Methods  The proposed framework extends the Simple Framework for Contrastive Learning of visual Representations (SimCLR) into the time-series domain by incorporating five temporal-specific data augmentation strategies: time compression, masking, permutation, scaling, and flipping. These augmentations generate multi-view samples that form positive pairs for contrastive learning, thereby ensuring temporal invariance in the feature space. A customized ResNet-18 encoder is employed to extract hierarchical features from the augmented time-series data, and a Multi-Layer Perceptron (MLP) projection head maps these features into a contrastive space. The Normalized Temperature-scaled cross-entropy (NT-Xent) loss is adopted to maximize similarity between positive pairs and minimize it between negative pairs, which effectively mitigates pseudo-label noise. To further improve recognition performance, a fine-tuning strategy is introduced in which pre-trained features are combined with a task-specific classification head using a limited amount of labeled data to adapt to downstream recognition tasks. This contrastive learning framework enables efficient analysis of time-series flight data, achieves accurate recognition of fighter aircraft maneuvers, and reduces dependence on large-scale labeled datasets.  Results and Discussions  Experiments are conducted on flight simulation data obtained from DCS World. To address the class imbalance issue, hybrid datasets (Table 1) are constructed, and training data ratios ranging from 2% to 30% are employed to evaluate the effectiveness of the weakly supervised framework. The results demonstrate that contrastive learning effectively captures the temporal patterns within flight data. For example, on the D1 dataset, accuracy with the base method increases from 35.83% with 2% labeled data to 89.62% when the fine-tuning ratio reaches 30% (Tables 36, Fig. 2(a)2(c)). To improve recognition of long maneuver sequences, a linear classifier and a voting strategy are introduced. The voting strategy markedly enhances few-shot learning performance. On the D1 dataset, accuracy reaches 54.5% with 2% labeled data and rises to 97.9% at a 30% fine-tuning ratio, representing a substantial improvement over the base method. On the D6 dataset, which simulates multi-source data fusion scenarios in air combat, the accuracy of the voting method increases from 0.476 with 2% labeled data to 0.928 with 30% labeled data (Fig. 2(d)2(f)), with a growth rate in the low-data phase 53% higher than that of the base method. Additionally, on the comprehensive D7 dataset, the accuracy standard deviation of the voting method is only 0.011 (Fig. 2(g), Fig. 3), significantly lower than the 0.015 observed for the base method. The superiority of the proposed framework can be attributed to two factors: the suppression of noise through integration of multiple prediction results using the voting strategy and the extraction of robust features from unlabeled data via contrastive learning pre-training. Together, these techniques enhance generalization and stability in complex scenarios, confirming the effectiveness of the method in leveraging unlabeled data and managing multi-source information.  Conclusions  This study applies the SimCLR framework to maneuver recognition and proposes a weakly supervised approach based on contrastive learning. By incorporating targeted data augmentation strategies and combining self-supervised learning with fine-tuning, the method exploits the latent information in time-series data, yielding substantial improvements in recognition performance under limited labeled data conditions. Experiments on simulated air combat datasets demonstrate that the framework achieves stable recognition across different data categories, offering practical insights for feature learning and model optimization in time-series classification tasks. Future research will focus on three directions: first, integrating real flight data to evaluate the model’s generalization capability in practical scenarios; second, developing dynamically adaptive data augmentation strategies to enhance performance in complex environments; and third, combining reinforcement learning and related techniques to improve autonomous decision-making in dynamic aerial missions, thereby expanding opportunities for intelligent flight operations.
A Review of Clutter Suppression Techniques in Ground Penetrating Radar: Mechanisms, Methods, and Challenges
LEI Wentai, WANG Yiming, ZHONG Jiwei, XU Qiguo, JIANG Yuyin, LI Cheng
Available online  , doi: 10.11999/JEIT250524
Abstract:
  Significance   Ground Penetrating Radar (GPR) is a widely adopted non-destructive subsurface detection technology, extensively applied in urban subsurface exploration, transportation infrastructure monitoring, geophysical surveys, and military operations. It is employed to detect underground pipelines, structural foundations, road voids, and concealed defects in roadbeds, railway tracks, and tunnels, as well as shallow geological formations and military targets such as unexploded ordnance. However, the presence of clutter—unwanted signals including direct coupling waves, ground reflections, and non-target echoes—severely degrades GPR data quality and complicates target detection, localization, imaging, and parameter estimation. Effective clutter suppression is therefore essential to enhance the accuracy and reliability of GPR data interpretation, making it a central research focus in improving GPR performance across diverse application domains.  Progress   Significant progress has been achieved in GPR clutter suppression, largely through two main approaches: signal model-based and neural network-based methods. Signal model-based techniques, such as time–frequency analysis, subspace decomposition, and dictionary learning, rely on physical modeling to distinguish clutter from target signals. These methods provide clear interpretability but are limited in addressing complex and non-linear clutter patterns. Neural network-based methods, employing architectures such as Convolutional Neural Networks, U-Net, and Generative Adversarial Networks, are more effective in capturing non-linear features through data-driven learning. Recent advances, including multi-scale convolutional autoencoders, attention mechanisms, and hybrid models, have further enhanced clutter suppression under challenging conditions. Quantitative metrics such as Mean Squared Error, Peak Signal-to-Noise Ratio, and Structural Similarity Index are commonly used for performance evaluation, often complemented by qualitative visual assessment.  Conclusion  The complexity and diversity of GPR clutter, originating from direct coupling, ground reflections, equipment imperfections, non-uniform media, and non-target scatterers, demand robust suppression strategies. Signal model-based methods provide strong theoretical foundations but are constrained by simplified assumptions, whereas neural network-based approaches offer greater adaptability at the expense of large data requirements and high computational cost. Hybrid approaches that integrate the strengths of both paradigms show considerable potential in addressing complex clutter scenarios. The selection of evaluation metrics plays a pivotal role in algorithm design, with quantitative measures offering objective assessment and qualitative analyses providing intuitive validation. Despite recent advances, significant challenges remain in suppressing non-linear clutter, enabling real-time processing, and reducing reliance on labeled data.  Prospect   Future research in GPR clutter suppression is likely to emphasize integrating the strengths of signal model-based and neural network-based methods to develop robust and adaptive solutions. Real-time processing and online learning will be prioritized to meet the requirements of dynamic applications. Self-supervised and unsupervised learning approaches are expected to reduce dependence on costly labeled datasets and improve model adaptability. Cross-task learning and multi-modal fusion, combining data from multiple sensors or frequencies, are expected to enhance robustness and precision. Furthermore, deeper integration of physical principles, including electromagnetic wave propagation and media properties, into algorithm design is expected to improve suppression accuracy and computational efficiency, advancing the development of more intelligent and effective GPR systems.
Microfabrication Method for Amorphous Wires GMI Magnetic Sensors
ZHANG Bo, WEN Xiaolong, WAN Yadong, ZHANG Chao, LI Jianhua
Available online  , doi: 10.11999/JEIT250338
Abstract:
  Objective  Compared with amorphous ribbons and thin films, amorphous wires exhibit superior Giant MagnetoImpedance (GMI) performance, making them promising materials for GMI magnetic sensors. Their flexible and heterogeneous morphology, however, complicates precise positioning during device fabrication. Additionally, the poor wettability of amorphous wires hinders control of contact resistance during soldering, often resulting in inconsistent device performance. This study proposes a microfabrication method for GMI magnetic sensors based on amorphous wires. Through-glass vias are employed as alignment markers, and auxiliary fixtures are used to accurately position and secure the wires on a glass wafer. Using photolithography and electroplating, bonding pads are fabricated to establish reliable electrical interconnections between the wires and the pads, enabling device-level processing and integration. A winding machine is then applied to wind the signal pickup coil on the device surface, completing fabrication of the GMI magnetic sensor. This approach avoids deformation and stress accumulation caused by direct coil winding on the amorphous wires, thereby improving manufacturability and ensuring stable performance of amorphous wire-based GMI magnetic sensors.  Methods  A glass wafer is employed as the substrate, owing to its high surface flatness and mechanical rigidity, which provide stable support for the flexible amorphous wire structure. To mitigate deformation caused by wire flexibility during winding, a microelectronics process integration scheme based on the glass wafer is implemented. A metal seed layer is first deposited by magnetron sputtering. Ultraviolet lithography and electroplating are then applied to form a high-precision array of electrical interconnection pads on the wafer surface. The ends of the amorphous wire are threaded through through-glass vias fabricated along the wafer edge by laser ablation and subsequently secured, ensuring accurate positioning over the bonding pad area while maintaining the natural straight form of the wire (Fig. 4). The amorphous wire is interconnected with the pads using electroplating. Standardized devices with an amorphous wire–glass substrate–interconnection structure are obtained by wafer dicing. After the microstructure of the amorphous wire and substrate is established, a winding machine is used to wind enameled wire onto the structure to form the signal pickup coil. The number of turns and spacing are precisely controlled according to the design. The sensor structure with the wound pickup coil is mounted on a Printed Circuit Board (PCB) with bonding pads. Finally, flip-chip bonding is performed to achieve secondary interconnection between the sensor structure and the PCB, completing fabrication of the sensor device.  Results and Discussions  The fabricated sensor device based on microelectronics processes is shown in Fig. 6(a). A 40 μm diameter enameled wire is uniformly wound on the substrate surface to form the signal pickup coil, with the number of turns and spacing precisely controlled by programmed parameters of the winding machine. As shown in the magnified view in Fig. 6(b), the bonding pad areas at both ends of the amorphous wire are completely covered by a copper layer. The copper plating defines the electrical connection area of the amorphous wire, while polyimide provides reliable fixation and surface protection on the substrate. The performance of five fabricated amorphous wire GMI magnetic sensors is presented in Fig. 13 and Table 1. The standard deviation of sensor output ranges from 0.0272 to 0.0163, and the sensors exhibit similar sensitivity, indicating good consistency. The output characteristic curves are shown in Fig. 14. Fitting analysis shows that both the Pearson correlation coefficient and the coefficient of determination are close to 1, demonstrating excellent linearity. When a 1 MHz excitation signal is applied to the amorphous wire, the output voltage exhibits a linear relationship with the external magnetic field within the range of –1 Oe to +1 Oe, with a sensitivity of 5.7 V/Oe. The magnetic noise spectrum, measured inside a magnetic shielding barrel, is shown in Fig. 15. The results indicate that the magnetic noise level of the sensor is approximately 55 pT/√Hz.  Conclusions  A fabrication method for amorphous wire-based GMI magnetic sensors is proposed using a glass substrate integration process. The sensor is constructed through microfabrication of a glass substrate–amorphous wire microstructure. The method is characterized by three features: (i) highly reliable interconnections between the amorphous wire and bonding pads are established by electroplating, yielding a 10 mm × 0.6 mm × 0.5 mm microstructure with fixed amorphous wires; (ii) a signal pickup coil is precisely wound on the microstructure surface with a winding machine, ensuring accurate control of coil turns and spacing; and (iii) electrical connection and circuit integration with a PCB are completed by flip-chip bonding. Compared with conventional amorphous wire GMI sensors, this approach provides two technical advantages. The microfabrication interconnection process reduces contact resistance fluctuations, addressing sensor performance dispersion. In addition, the combination of conventional winding and microelectronics techniques ensures device consistency while avoiding the high cost of full-process microfabrication. This method improves process compatibility and manufacturing repeatability, offering a practical route for engineering applications of GMI magnetic sensors.
Three-Dimensional Imaging Method for Concealed Human Targets Based on Array Stitching
QIU Chen, CHEN Jiahui, SHAO Fengzhi, LI Nian, XU Zihan, GUO Shisheng, CUI Guolong
Available online  , doi: 10.11999/JEIT250334
Abstract:
  Objective  Traditional Through-the-Wall Radar (TWR) systems based on planar multiple-input multiple-output arrays often face high hardware complexity, calibration challenges, and increased system cost. To overcome these limitations, we propose a Three-Dimensional (3D) imaging framework based on array stitching. The method uses either time-sequential or simultaneous operation of multiple small-aperture radar sub-arrays to emulate a large aperture. This strategy substantially reduces hardware complexity while maintaining accurate 3D imaging of concealed human targets.  Methods  The proposed framework integrates three core techniques: 3D weighted total variation (3DWTV) reconstruction, Lucy–Richardson (LR) deconvolution, and 3D wavelet transform (3DWT)-based fusion. Radar echoes are first collected from horizontally and vertically distributed sub-arrays that emulate a planar aperture. Each sub-array image is independently reconstructed using 3DWTV, which enforces spatial sparsity to suppress noise while preserving structural details. The horizontal and vertical images are then multiplicatively fused to jointly recover azimuth and elevation information. To reduce diffraction-induced blurring, LR deconvolution models system degradation through the Point Spread Function (PSF) and iteratively refines scene reflectivity, thereby enhancing cross-range resolution. Finally, 3DWT decomposes the images into multi-scale sub-bands (e.g., LLL, LLH, LHL), which are selectively fused using absolute-maximum and fuzzy-logic rules. The inverse wavelet transform is then applied to reconstruct the final 3D image, retaining both global and local features.  Results and Discussions  The proposed method is evaluated through both simulations and real-world experiments using a Stepped-Frequency Continuous-Wave (SFCW) radar operating from 1.6 to 2.2 GHz with a 2Tx–4Rx configuration. In simulations, compared with baseline algorithms such as Back-Projection (BP) and the Fast Iterative Shrinkage-Thresholding Algorithm (FISTA), the proposed method achieves better performance. Image Entropy (IE) decreases from 9.7125 for BP and 9.7065 for FISTA to 8.0711, which reflects improved image quality. Experimental tests conducted in indoor environments further confirm robustness. For both standing and sitting postures, IE is reduced from 9.9982 to 7.0030 and from 9.9947 to 6.2261, respectively.  Conclusions  This study presents a low-cost, high-resolution 3D imaging method for TWR systems based on array stitching. By integrating 3DWTV reconstruction, LR deconvolution, and 3DWT fusion, the method effectively reconstructs concealed human postures using a limited aperture. The approach simplifies hardware design, reduces system complexity, and preserves imaging quality under sparse sampling, thereby providing a practical solution for portable and scalable TWR systems.
Low Complexity Sequential Decoding Algorithm of PAC Code for Short Packet Communication
DAI Jingxin, YIN Hang, WANG Yuhuan, LV Yansong, YANG Zhanxin, LV Rui, XIA Zhiping
Available online  , doi: 10.11999/JEIT250533
Abstract:
  Objective  With the rise of the intelligent Internet of Things (IoT), short packet communication among IoT devices must meet stringent requirements for low latency, high reliability, and very short packet length, posing challenges to the design of channel coding schemes. As an advanced variant of polar codes, Polarization-Adjusted Convolutional (PAC) codes enhance the error-correction performance of polar codes at medium and short code lengths, approaching the dispersion bound in some cases. This makes them promising for short packet communication. However, the high decoding complexity required to achieve near-bound error-correction performance limits their practicality. To address this, we propose two low complexity sequential decoding algorithms: Low Complexity Fano Sequential (LC-FS) and Low Complexity Stack (LC-S). Both algorithms effectively reduce decoding complexity with negligible loss in error-correction performance.  Methods  To reduce the decoding complexity of Fano-based sequential decoding algorithms, we propose the LC-FS algorithm. This method exploits special nodes to terminate decoding at intermediate levels of the decoding tree, thereby reducing the complexity of tree traversal. Special nodes are classified into two types according to decoder structure: low-rate nodes (Type-\begin{document}$ \mathrm{T} $\end{document} node) and high-rate nodes [Rate-1 and Single Parity-Check (SPC) nodes]. This classification minimizes unnecessary hardware overhead by avoiding excessive subdivision of special nodes. For each type, a corresponding LC-FS decoder and node-movement strategy are developed. To reduce the complexity of stack-based decoding algorithms, we propose the LC-S algorithm. While preserving the low backtracking feature of stack-based decoding, this method introduces tailored decoding structures and node-movement strategies for low-rate and high-rate special nodes. Therefore, the LC-S algorithm achieves significant complexity reduction without compromising error-correction performance.  Results and Discussions  The performance of the proposed LC-FS and LC-S decoding algorithms is evaluated through extensive simulations in terms of Frame Error Rate (FER), Average Computational Complexity (ACC), Maximum Computational Complexity (MCC), and memory requirements. Traditional Fano sequential, traditional stack, and Fast Fano Sequential (FFS) decoding algorithms are set as benchmarks. The simulation results show that the LC-FS and LC-S algorithms exhibit negligible error-correction performance loss compared with traditional Fano sequential and stack decoders (Fig. 5). Across different PAC codes, both algorithms effectively reduce decoding complexity. Specifically, as increases, the reductions in ACC and MCC become more pronounced. For ACC, LC-FS decoding algorithm (\begin{document}$T = 4$\end{document}) achieves reductions of 13.77% (\begin{document}$N = 256$\end{document}, \begin{document}$K = 128$\end{document}), 11.42% (\begin{document}$N = 128$\end{document}, \begin{document}$K = 64$\end{document}), and 25.52% (\begin{document}$N = 64$\end{document}, \begin{document}$K = 32$\end{document}) on average compared with FFS (Fig. 6). LC-S decoding algorithm (\begin{document}$T = 4$\end{document}) reduces ACC by 56.48% (\begin{document}$N = 256$\end{document}, \begin{document}$K = 128$\end{document}), 47.63% (\begin{document}$N = 128$\end{document}, \begin{document}$K = 64$\end{document}), and 49.61% (\begin{document}$N = 64$\end{document}, \begin{document}$K = 32$\end{document}) on average compared with the traditional stack algorithm (Fig. 6). For MCC, LC-FS decoding algorithm (\begin{document}$T = 4$\end{document}) achieves reductions of 29.71% (\begin{document}$N = 256$\end{document}, \begin{document}$K = 128$\end{document}), 21.18% (\begin{document}$N = 128$\end{document}, \begin{document}$K = 64$\end{document}), and 23.62% (\begin{document}$N = 64$\end{document}, \begin{document}$K = 32$\end{document}) on average compared with FFS (Fig. 7). LC-S decoding algorithm (\begin{document}$T = 4$\end{document}) reduces MCC by 67.17% (\begin{document}$N = 256$\end{document}, \begin{document}$K = 128$\end{document}), 49.33% (\begin{document}$N = 128$\end{document}, \begin{document}$K = 64$\end{document}), and 51.84% (\begin{document}$N = 64$\end{document}, \begin{document}$K = 32$\end{document}) on average compared with the traditional stack algorithm (Fig. 7). By exploiting low-rate and high-rate special nodes to terminate decoding at intermediate levels of the decoding tree, the LC-FS and LC-S algorithms also reduce memory requirements (Table 2). However, as \begin{document}$T$\end{document} increases, the memory usage of LC-S rises because all extended paths of low-rate special nodes are pushed into the stack. The increase in \begin{document}$T$\end{document} enlarges the number of extended paths, indicating its critical role in balancing decoding complexity and memory occupation (Fig. 8).  Conclusions  To address the high decoding complexity of sequential decoding algorithms for PAC codes, this paper proposes two low complexity approaches: the LC-FS and LC-S algorithms. Both methods classify special nodes into low-rate and high-rate categories and design corresponding decoders and movement strategies. By introducing Type-\begin{document}$ \mathrm{T} $\end{document} nodes, the algorithms further eliminate redundant computations during decoding, thereby reducing complexity. Simulation results demonstrate that the LC-FS and LC-S algorithms substantially decrease decoding complexity while maintaining the error-correction performance of PAC codes at medium and short code lengths.
Construction of Multi-Scroll Conservative Chaotic System and Its Application in Image Encryption
AN Xinlei, LI Zhifu, XUE Rui, XIONG Li, ZHANG Li
Available online  , doi: 10.11999/JEIT250432
Abstract:
  Objective  Existing conservative chaotic systems often suffer from structural simplicity and weak nonlinear characteristics, and research on complex dynamical behaviors such as multi-scroll structures remains limited, constraining their potential in engineering applications. To address security risks in face image transmission and the inefficiency of traditional global encryption methods, this study constructs a conservative chaotic system with multi-scroll characteristics, investigates its complex dynamical behavior, and designs a face-detection-based selective image encryption algorithm targeting sensitive regions. The work explores the practical application of conservative chaotic systems in image encryption.  Methods  A five-dimensional conservative hyperchaotic system is constructed on the basis of the generalized Hamiltonian system, and the controlled generation of multi-scroll chaotic flows is achieved through modulation of the Hamiltonian energy function. The Hessian matrix is used to analyze the stationary points of the Hamiltonian energy function, thereby revealing the relationship between scroll structures and stationary points. The spatial distribution of multi-scroll chaotic flows is further characterized by energy isosurfaces. The complex dynamical behaviors of the proposed system are investigated using Lyapunov exponent spectra and phase diagrams, while the sequence complexity is evaluated with the SE complexity algorithm. On this basis, an image encryption algorithm integrated with face detection technology is designed. The algorithm applies a diffusion–scrambling strategy to selectively encrypt facial regions. The security performance is evaluated through multiple indicators, including key space, pixel correlation, and information entropy.  Results and Discussions  Analysis of stationary points in the Hamiltonian energy function revealed a positive correlation between their number and scroll generation. Extreme points primarily drive scroll formation, whereas saddle points define transition zones, indicating that the scroll structure can be effectively regulated through the Hamiltonian energy function. The Lyapunov exponent spectrum of the multi-scroll conservative chaotic system is distributed symmetrically about the x-axis and exhibits an integer Lyapunov dimension, fully confirming the system’s volume-conserving property. Under different initial conditions, the system demonstrates diverse coexistence behaviors, including phase trajectories of varying types and scales. Complexity evaluation further showed that the multi-scroll conservative chaotic system achieves markedly higher spectral entropy complexity, supporting its potential for image encryption applications. Experimental validation demonstrated that the proposed algorithm can accurately detect faces and selectively encrypt sensitive regions, thereby avoiding the computational inefficiency of indiscriminate global encryption. Moreover, the algorithm exhibited strong performance across multiple security metrics.  Conclusions  A conservative chaotic system is constructed on the basis of the generalized Hamiltonian system, and its complex dynamical behavior and application in image encryption are investigated. The study provides theoretical references for the generation of multi-scroll conservative chaotic flows and offers practical guidance for the application of image encryption technology.
For Electric Power Disaster Early Warning Scenarios: A Large Model and Lightweight Models Joint Deployment Scheme Based on Limited Spectrum Resources
CHEN Lei, HUANG Zaichao, LIU Chuan, ZHANG Weiwei
Available online  , doi: 10.11999/JEIT250321
Abstract:
  Objective  Traditional approaches to electric power disaster early warning rely on dedicated, scenario-specific systems, leading to redundant data collection and high development costs. To enhance accuracy and reduce costs, comprehensive early warning frameworks based on Artificial Intelligence (AI) large models have become an important research direction. However, large models are typically deployed in the cloud, and limited wireless spectrum resources constrain the uploading of complete data streams. Deploying lightweight models at terminal devices through substantial model compression can alleviate spectrum limitations but inevitably compromises model performance.  Methods  To address these limitations, this study proposes a cloud–terminal collaborative joint deployment scheme integrating large and lightweight models. In this framework, a high-precision large model is deployed in the cloud to process complex tasks, whereas lightweight models are deployed at terminal devices to handle simple tasks. Task offloading decisions are governed by a confidence threshold that dynamically determines whether computation occurs locally or in the cloud. A power-domain Non-Orthogonal Multiple Access (NOMA) technique is incorporated to allow multiple terminals to share identical time–frequency resources, thereby improving system detection accuracy by increasing the proportion of tasks processed in the cloud. Additionally, for scenarios considering (i) only uplink shared-channel bandwidth constraints and (ii) both terminal access collision constraints and shared-channel bandwidth constraints, corresponding algorithms are developed to determine the maximum number of terminals supported under a given bandwidth and to identify the optimal confidence threshold that maximizes detection accuracy.  Results and Discussions  (1) As shown in Figures 3(a) and 3(b), when the uplink shared-channel bandwidth \begin{document}$ W $\end{document} increases, the number of supported terminals rises for both the proposed scheme and the orthogonal multiple access (OMA)-based scheme. This occurs because a larger \begin{document}$ W $\end{document} enables more terminals with low-confidence detection results to upload raw data to the cloud for further processing, thereby enhancing detection accuracy and reducing the missed detection rate. (2) In contrast, the number of supported terminals \begin{document}$ M $\end{document} in the pure on-device processing scheme remains constant with varying \begin{document}$ W $\end{document}, as this scheme relies entirely on the lightweight model deployed at the terminal and is therefore independent of bandwidth. (3) Compared with the OMA-based and pure on-device schemes, the proposed approach markedly increases the number of supported terminals, confirming that non-orthogonal reuse of time–frequency resources and cloud–terminal collaborative deployment of large and lightweight models are key to improving system performance. (4) As shown in Table 3, an increase in the number of preambles reduces the probability of terminal access collisions, allowing more terminals to successfully transmit raw data to the cloud for detection. Therefore, the missed detection rate decreases, and overall detection accuracy improves.  Conclusions  For electric power disaster early warning scenarios, this study integrates power-domain NOMA and proposes a cloud–terminal collaborative deployment scheme combining a large model with lightweight models. By dynamically determining whether tasks are processed locally by a lightweight model or in the cloud by a large model, the system achieves optimized detection accuracy and a reduced missed detection rate. Numerical results indicate that, under given uplink shared-channel bandwidth, minimum detection accuracy, and maximum missed detection rate, the introduction of power-domain NOMA effectively increases the number of supported terminals. Furthermore, when both terminal access collision constraints and shared-channel bandwidth constraints are considered, optimizing the confidence threshold to regulate the number of terminals transmitting data to the cloud further enhances detection accuracy and reduces the missed detection rate.
Entropy Quantum Collaborative Planning Method for Emergency Path of Unmanned Aerial Vehicles Driven by Survival Probability
WANG Enliang, ZHANG Zhen, SUN Zhixin
Available online  , doi: 10.11999/JEIT250694
Abstract:
  Objective  Natural disaster emergency rescue places stringent requirements on the timeliness and safety of Unmanned Aerial Vehicle (UAV) path planning. Conventional optimization objectives, such as minimizing total distance, often fail to reflect the critical time-sensitive priority of maximizing the survival probability of trapped victims. Moreover, existing algorithms struggle with the complex constraints of disaster environments, including no-fly zones, caution zones, and dynamic obstacles. To address these challenges, this paper proposes an Entropy-Enhanced Quantum Ripple Synergy Algorithm (E2QRSA). The primary goals are to establish a survival probability maximization model that incorporates time decay characteristics and to design a robust optimization algorithm capable of efficiently handling complex spatiotemporal constraints in dynamic disaster scenarios.  Methods  E2QRSA enhances the Quantum Ripple Optimization framework through four key innovations: (1) information entropy–based quantum state initialization, which guides population generation toward high-entropy regions; (2) multi-ripple collaborative interference, which promotes beneficial feature propagation through constructive superposition; (3) entropy-driven parameter control, which dynamically adjusts ripple propagation according to search entropy rates; and (4) quantum entanglement, which enables information sharing among elite individuals. The model employs a survival probability objective function that accounts for time-sensitive decay, base conditions, and mission success probability, subject to constraints including no-fly zones, warning zones, and dynamic obstacles.  Results and Discussions  Simulation experiments are conducted in medium- and large-scale typhoon disaster scenarios. The proposed E2QRSA achieves the highest survival probabilities of 0.847 and 0.762, respectively (Table 1), exceeding comparison algorithms such as SEWOA and PSO by 4.2–16.0%. Although the paths generated by E2QRSA are not the shortest, they are the most effective in maximizing survival chances. The ablation study (Table 3) confirms the contribution of each component, with the removal of multi-ripple interference causing the largest performance decrease (9.97%). The dynamic coupling between search entropy and ripple parameters (Fig. 2) is validated, demonstrating the effectiveness of the adaptive control mechanism. The entanglement effect (Fig. 4) is shown to maintain population diversity. In terms of constraint satisfaction, E2QRSA-planned paths consume only 85.2% of the total available energy (Table 5), ensuring a safe return, and all static and dynamic obstacles are successfully avoided, as visually verified in the 3D path plots (Figs. 6 and 7).  Conclusions  E2QRSA effectively addresses the challenge of UAV path planning for disaster relief by integrating adaptive entropy control with quantum-inspired mechanisms. The survival probability objective captures the essential requirements of disaster scenarios more accurately than conventional distance minimization. Experimental validation demonstrates that E2QRSA achieves superior solution quality and faster convergence, providing a robust technical basis for strengthening emergency response capabilities.
Multi-modal Joint Distillation Optimization for Source Code Vulnerability Detection
ZHANG Xuejun, ZHANG Yiffan, LIU Cancan, JIA Xiaohong, CHEN Zhuo, ZHANG Lei
Available online  , doi: 10.11999/JEIT250453
Abstract:
  Objective  As software systems increase in scale and complexity, the frequency of security vulnerabilities in source code rises accordingly, threatening system reliability, data integrity, and user privacy. Conventional automated vulnerability detection methods typically depend on a narrow set of shallow features—such as API call sequences, opcode patterns, or syntactic heuristics—rendering them susceptible to learning spurious correlations and unable to capture the rich semantic and structural information essential for accurate detection. Moreover, most existing approaches either rely on single-modal representations or weakly integrate multiple modalities without explicitly addressing distribution mismatches across them. This often results in overfitting to seen datasets and limited generalization to unseen codebases, particularly across different projects or programming languages. Although recent advances in machine learning and deep learning have improved source code analysis, effectively modeling the complex interactions between code semantics and execution structures remains a major challenge. To overcome these limitations, this paper proposes a multi-modal joint Distillation Optimization for Vulnerability Detection (mVulD-DO), a multimodal framework that combines deep feature distillation with dynamic global feature alignment. The proposed method aims to enhance semantic comprehension, structural reasoning, and cross-modal consistency, which are critical for robust vulnerability detection. By enforcing both intra-modality refinement and inter-modality alignment, mVulD-DO addresses the semantic-structural gap that constrains traditional methods.  Methods  The mVulD-DO framework begins by extracting multiple semantic modalities from raw source code—function names, variable names, token_type attributes, and local code slices—using program slicing and syntactic parsing techniques. In parallel, a Program Dependency Graph (PDG) is constructed to capture both control-flow and data-flow relationships, generating a heterogeneous graph that represents explicit and implicit program behaviors. Each semantic modality is embedded using pretrained encoders and subsequently refined via a deep feature distillation module, which integrates multi-head self-attention and multi-scale convolutional layers to emphasize salient patterns and suppress redundant noise. To model the sequential dependencies intrinsic to program execution, a Bidirectional Long Short-Term Memory (BLSTM) network captures long-range contextual interactions. For structural representation, a Graph Attention Network (GAT) processes the PDG-C to produce topology-aware embeddings. To bridge the gap between modalities, adaptive dynamic Sinkhorn regularization is applied to globally align the distributions of semantic and structural embeddings. This approach mitigates modality mismatches while preserving flexibility by avoiding rigid one-to-one correspondences. Finally, the distilled and aligned multimodal features are fused and passed through a lightweight fully connected classifier for binary vulnerability prediction. The model is jointly optimized using both classification and alignment objectives, improving its ability to generalize across unseen codebases.  Results and Discussions  Comprehensive comparisons conducted on the mixed CVEFixes+SARD dataset—covering 25 common CWE vulnerability types with an 8:1:1 train:validation:test split—demonstrate that traditional source code detectors, which directly map code to labels, often rely on superficial patterns and show limited generalization, particularly for out-of-distribution samples. These methods achieve accuracies ranging from 55.41% to 85.84%, with recall typically below 80% (Table 1). In contrast, mVulD-DO leverages multi-layer feature distillation to purify and enhance core representations, while dynamic Sinkhorn alignment mitigates cross-modal inconsistencies. This results in accuracy of 87.11% and recall of 83.59%, representing absolute improvements of 1.27% and 6.26%, respectively, over the strongest baseline method (ReGVD). Although mVulD-DO reports a slightly higher False Positive Rate (FPR) of 3.54%—2.92% above that of ReGVD—it remains lower than that of most traditional detectors. This modest increase is considered acceptable in practice, given that failing to detect a critical vulnerability typically incurs greater cost than issuing additional alerts. Compared with instruction-tuned large language models (e.g., VulLLM), which maintain low FPRs below 10% but suffer from recall below 75%, mVulD-DO offers a more favorable trade-off between false alarms and coverage of true vulnerabilities. Ablation studies (Table 2) further validate the contribution of each component. Removing function name embeddings (unfunc) results in a 1.3% decrease in F1 score; removing variable name embeddings (unvar) causes a 1.3% drop; and omitting token_type attributes (untype) leads to a 3.35% reduction. The most substantial performance degradation—9.11% in F1—occurs when the deep feature distillation module is disabled (undis), highlighting the critical role of multi-scale semantic refinement and noise suppression. Additional evaluations on vulnerability-sensitive subsets—Call, OPS, Array, and PTR—demonstrate consistent benefits from Sinkhorn alignment. F1 score improvements over unaligned variants are observed as follows: 1.45% for Call, 4.22% for OPS, 1.38% for Array, and 0.36% for PTR (Table 3), confirming the generalization advantage across a broad spectrum of vulnerability types.  Conclusions  Experimental results demonstrate that the proposed mVulD-DO framework consistently outperforms existing vulnerability detection methods in recall, F1-score, and accuracy, while maintaining a low FPR The effectiveness of deep feature distillation, multi-scale semantic extraction, and dynamic Sinkhorn alignment is validated through extensive ablation and visualization analyses. Despite these improvements, the model incurs additional computational overhead due to multimodal distillation and Sinkhorn alignment, and shows sensitivity to hyperparameter settings, which may limit its suitability for real-time applications. Moreover, while strong performance is achieved on the mixed dataset, the model's generalization across unseen projects and programming languages remains an open challenge. Future work will focus on developing lightweight training strategies—such as knowledge distillation and model pruning—to reduce computational costs. Additionally, incorporating unsupervised domain adaptation and incremental alignment mechanisms will be critical to support dynamic code evolution and enhance cross-domain robustness. These directions aim to improve the scalability, adaptability, and practical deployment of multimodal vulnerability detection systems in diverse software environments.
Advances in Deep Neural Network Based Image Compression: A Survey
BAI Yuanchao, LIU Wenchang, JIANG Junjun, LIU Xianming
Available online  , doi: 10.11999/JEIT250567
Abstract:
  Significance   With the continuous advancement of information technology, digital images are evolving toward ultra-high-definition formats characterized by increased resolution, dynamic range, color depth, sampling rates, and multi-viewpoint support. In parallel, the rapid development of artificial intelligence is reshaping both the generation and application paradigms of digital imagery. As visual big data converges with AI technologies, the volume and diversity of image data expand exponentially, creating unprecedented challenges for storage and transmission. As a core technology in digital image processing, image compression reduces storage costs and bandwidth requirements by eliminating internal information redundancy, thereby serving as a fundamental enabler for visual big data applications. However, traditional image compression standards increasingly struggle to meet rising industrial demands due to limited modeling capacity, inadequate perceptual adaptability, and poor compatibility with machine vision tasks. Deep Neural Network (DNN)-based image compression methods, leveraging powerful modeling capabilities, end-to-end optimization mechanisms, and compatibility with both human perception and machine understanding, are progressively exceeding conventional coding approaches. These methods demonstrate clear advantages and broad potential across diverse application domains, drawing growing attention from both academia and industry.  Progress   This paper systematically reviews recent advances in DNN-based image compression from three core perspectives: signal fidelity, human visual perception, and machine analysis. First, in signal fidelity–oriented compression, the rate–distortion optimization framework is introduced, with detailed discussion of key components in lossy image compression, including nonlinear transforms, quantization strategies, entropy coding mechanisms, and variable-rate techniques for multi-rate adaptation. The synergistic design of these modules underpins the architecture of modern DNN-based image compression systems. Second, in perceptual quality–driven compression, the principles of joint rate–distortion–perception optimization models are examined, together with a comparative analysis of two major perceptual paradigms: Generative Adversarial Network (GAN)-based models and diffusion model–based approaches. Both strategies employ perceptual loss functions or generative modeling techniques to markedly improve the visual quality of reconstructed images, aligning them more closely with the characteristics of the human visual system. Finally, in machine analysis–oriented compression, a co-optimization framework for rate–distortion–accuracy trade-offs is presented, with semantic fidelity as the primary objective. From the perspective of integrating image compression with downstream machine analysis architectures, this section analyzes how current methods preserve essential semantic information that supports tasks such as object detection and semantic segmentation during the compression process.  Conclusions  DNN-based image compression shows strong potential across signal fidelity, human visual perception, and machine analysis. Through end-to-end jointly optimized neural network architectures, these methods provide comprehensive modeling of the encoding process and outperform traditional approaches in compression efficiency. By leveraging the probabilistic modeling and image generation capabilities of DNNs, they can accurately estimate distributional differences between reconstructed and original images, quantify perceptual losses, and generate high-quality reconstructions that align with human visual perception. Furthermore, their compatibility with mainstream image analysis frameworks enables the extraction of semantic features and the design of collaborative optimization strategies, allowing efficient compression tailored to machine vision tasks.  Prospects   Despite significant progress in compression performance, perceptual quality, and task adaptability, DNN-based image compression still faces critical technical challenges and practical limitations. First, computational complexity remains high. Most high-performance models rely on deep and sophisticated architectures (e.g., attention mechanisms and Transformer models), which enhance modeling capability but also introduce substantial computational overhead and long inference latency. These limitations are particularly problematic for deployment on mobile and embedded devices. Second, robustness and generalization continue to be major concerns. DNN-based compression models are sensitive to input perturbations and vulnerable to adversarial attacks, which can lead to severe reconstruction distortions or even complete failure. Moreover, while they perform well on training data and similar distributions, their performance often degrades markedly under cross-domain scenarios. Third, the evaluation framework for perceptual- and machine vision–oriented compression remains immature. Although new evaluation dimensions have been introduced, no unified and objective benchmark exists. This gap is especially evident in machine analysis–oriented compression, where downstream tasks vary widely and rely on different visual models. Therefore, comparability across methods is limited and consistent evaluation metrics are lacking, constraining both research and practical adoption. Overall, DNN-based image compression is in transition from laboratory research to real-world deployment. Although it demonstrates clear advantages over traditional approaches, further advances are needed in efficiency, robustness, generalization, and standardized evaluation protocols. Future research should strengthen the synergy between theoretical exploration and engineering implementation to accelerate widespread adoption and continued progress in areas such as multimedia communication, edge computing, and intelligent image sensing systems.
A Method for Named Entity Recognition in Military Intelligence Domain Using Large Language Models
LI Yongbin, LIU Lian, ZHENG Jie
Available online  , doi: 10.11999/JEIT250764
Abstract:
  Objective  Named Entity Recognition (NER) is a fundamental task in information extraction within specialized domains, particularly military intelligence. It plays a critical role in situation assessment, threat analysis, and decision support. However, conventional NER models face major challenges. First, the scarcity of high-quality annotated data in the military intelligence domain is a persistent limitation. Due to the sensitivity and confidentiality of military information, acquiring large-scale, accurately labeled datasets is extremely difficult, which severely restricts the training performance and generalization ability of supervised learning–based NER models. Second, military intelligence requires handling complex and diverse information extraction tasks. The entities to be recognized often possess domain-specific meanings, ambiguous boundaries, and complex relationships, making it difficult for traditional models with fixed architectures to adapt flexibly to such complexity or achieve accurate extraction. This study aims to address these limitations by developing a more effective NER method tailored to the military intelligence domain, leveraging Large Language Models (LLMs) to enhance recognition accuracy and efficiency in this specialized field.  Methods  To achieve the above objective, this study focuses on the military intelligence domain and proposes a NER method based on LLMs. The central concept is to harness the strong semantic reasoning capabilities of LLMs, which enable deep contextual understanding of military texts, accurate interpretation of complex domain-specific extraction requirements, and autonomous execution of extraction tasks without heavy reliance on large annotated datasets. To ensure that general-purpose LLMs can rapidly adapt to the specialized needs of military intelligence, two key strategies are employed. First, instruction fine-tuning is applied. Domain-specific instruction datasets are constructed to include diverse entity types, extraction rules, and representative examples relevant to military intelligence. Through fine-tuning with these datasets, the LLMs acquire a more precise understanding of the characteristics and requirements of NER in this field, thereby improving their ability to follow targeted extraction instructions. Second, Retrieval-Augmented Generation (RAG) is incorporated. A domain knowledge base is developed containing expert knowledge such as entity dictionaries, military terminology, and historical extraction cases. During the NER process, the LLM retrieves relevant knowledge from this base in real time to support entity recognition. This strategy compensates for the limited domain-specific knowledge of general LLMs and enhances recognition accuracy, particularly for rare or complex entities.  Results and Discussions  Experimental results indicate that the proposed LLM–based NER method, which integrates instruction fine-tuning and RAG, achieves strong performance in military intelligence NER tasks. Compared with conventional NER models, it demonstrates higher precision, recall, and F1-score, particularly in recognizing complex entities and managing scenarios with limited annotated data. The effectiveness of this method arises from several key factors. The powerful semantic reasoning capability of LLMs enables a deeper understanding of contextual nuances and ambiguous expressions in military texts, thereby reducing missed and false recognitions commonly caused by rigid pattern-matching approaches. Instruction fine-tuning allows the model to better align with domain-specific extraction requirements, ensuring that the recognition results correspond more closely to the practical needs of military intelligence analysis. Furthermore, the incorporation of RAG provides real-time access to domain expert knowledge, markedly enhancing the recognition of entities that are highly specialized or morphologically variable within military contexts. This integration effectively mitigates the limitations of traditional models that lack sufficient domain knowledge.  Conclusions  This study proposes a LLM–based NER method for the military intelligence domain, effectively addressing the challenges of limited annotated data and complex extraction requirements encountered by traditional models. By combining instruction fine-tuning and RAG, general-purpose LLMs can be rapidly adapted to the specialized demands of military intelligence, enabling the construction of an efficient domain-specific expert system at relatively low cost. The proposed method provides an effective and scalable solution for NER tasks in military intelligence scenarios, enhancing both the efficiency and accuracy of information extraction in this field. It offers not only practical value for military intelligence analysis and decision support but also methodological insight for NER research in other specialized domains facing similar data and complexity constraints, such as aerospace and national security. Future research will focus on optimizing instruction fine-tuning strategies, expanding the domain knowledge base, and reducing computational cost to further improve model performance and applicability.
Wave-MambaCT: Low-dose CT Artifact Suppression Method Based on Wavelet Mamba
CUI Xueying, WANG Yuhang, LIU Bin, SHANGGUAN Hong, ZHANG Xiong
Available online  , doi: 10.11999/JEIT250489
Abstract:
  Objective  Low-Dose Computed Tomography (LDCT) reduces patient radiation exposure but introduces substantial noise and artifacts into reconstructed images. Convolutional Neural Network (CNN)-based denoising approaches are limited by local receptive fields, which restrict their abilities to capture long-range dependencies. Transformer-based methods alleviate this limitation but incur quadratic computational complexity relative to image size. In contrast, State Space Model (SSM)–based Mamba frameworks achieve linear complexity for long-range interactions. However, existing Mamba-based methods often suffer from information loss and insufficient noise suppression. To address these limitations, we propose the Wave-MambaCT model.  Methods  The proposed Wave-MambaCT model adopts a multi-scale framework that integrates Discrete Wavelet Transform (DWT) with a Mamba module based on the SSM. First, DWT performs a two-level decomposition of the LDCT image, decoupling noise from Low-Frequency (LF) content. This design directs denoising primarily toward the High-Frequency (HF) components, facilitating noise suppression while preserving structural information. Second, a residual module combined with a Spatial-Channel Mamba (SCM) module extracts both local and global features from LF and HF bands at different scales. The noise-free LF features are then used to correct and enhance the corresponding HF features through an attention-based Cross-Frequency Mamba (CFM) module. Finally, inverse wavelet transform is applied in stages to progressively reconstruct the image. To further improve denoising performance and network stability, multiple loss functions are employed, including L1 loss, wavelet-domain LF loss, and adversarial loss for HF components.  Results and Discussions  Extensive experiments on the simulated Mayo Clinic datasets, the real Piglet datasets, and the hospital clinical dataset DeepLesion show that Wave-MambaCT provides superior denoising performance and generalization. On the Mayo dataset, a PSNR of 31.6528 is achieved, which is higher than that of the suboptimal method DenoMamba (PSNR 31.4219), while MSE is reduced to 0.00074 and SSIM and VIF are improved to 0.8851 and 0.4629, respectively (Table 1). Visual results (Figs. 46) demonstrate that edges and fine details such as abdominal textures and lesion contours are preserved, with minimal blurring or residual artifacts compared with competing methods. Computational efficiency analysis (Table 2) indicates that Wave-MambaCT maintains low FLOPs (17.2135 G) and parameters (5.3913 M). FLOPs are lower than those of all networks except RED-CNN, and the parameter count is higher only than those of RED-CNN and CTformer. During training, 4.12 minutes per epoch are required, longer only than RED-CNN. During testing, 0.1463 seconds are required per image, which is at a medium level among the compared methods. Generalization tests on the Piglet datasets (Figs. 7, 8, Tables 3, 4) and DeepLesion (Fig. 9) further confirm the robustness and generalization capacity of Wave-MambaCT.In the proposed design, HF sub-bands are grouped, and noise-free LF information is used to correct and guide their recovery. This strategy is based on two considerations. First, it reduces network complexity and parameter count. Second, although the sub-bands correspond to HF information in different orientations, they are correlated and complementary as components of the same image. Joint processing enhances the representation of HF content, whereas processing them separately would require a multi-branch architecture, inevitably increasing complexity and parameters. Future work will explore approaches to reduce complexity and parameters when processing HF sub-bands individually, while strengthening their correlations to improve recovery. For structural simplicity, SCM is applied to both HF and LF feature extraction. However, redundancy exists when extracting LF features, and future studies will explore the use of different Mamba modules for HF and LF features to further optimize computational efficiency.  Conclusions  Wave-MambaCT integrates DWT for multi-scale decomposition, a residual module for local feature extraction, and an SCM module for efficient global dependency modeling to address the denoising challenges of LDCT images. By decoupling noise from LF content through DWT, the model enables targeted noise removal in the HF domain, facilitating effective noise suppression. The designed RSCM, composed of residual blocks and SCM modules, captures fine-grained textures and long-range interactions, enhancing the extraction of both local and global information. In parallel, the Cross-band Enhancement Module (CEM) employs noise-free LF features to refine HF components through attention-based CFM, ensuring structural consistency across scales. Ablation studies (Table 5) confirm the essential contributions of both SCM and CEM modules to maintaining high performance. Importantly, the model’s staged denoising strategy achieves a favorable balance between noise reduction and structural preservation, yielding robustness to varying radiation doses and complex noise distributions.
Source Code Vulnerability Detection Method Integrating Code Sequences and Property Graphs
YANG Hongyu, LUO Jingchuan, CHENG Xiang, HU Juncheng
Available online  , doi: 10.11999/JEIT250470
Abstract:
  Objective  Code vulnerabilities create opportunities for hacker intrusions, and if they are not promptly identified and remedied, they pose serious threats to cybersecurity. Deep learning–based vulnerability detection methods leverage large collections of source code to learn secure programming patterns and vulnerability characteristics, enabling the automated identification of potential security risks and enhancing code security. However, most existing deep learning approaches rely on a single network architecture, extracting features from only one perspective, which constrains their ability to comprehensively capture multi-dimensional code characteristics. Some studies have attempted to address this by extracting features from multiple dimensions, yet the adopted feature fusion strategies are relatively simplistic, typically limited to feature concatenation or weighted combination. Such strategies fail to capture interdependencies among feature dimensions, thereby reducing the effectiveness of feature fusion. To address these challenges, this study proposes a source code vulnerability detection method integrating code sequences and property graphs. By optimizing both feature fusion and vulnerability detection processes, the proposed method effectively enhances the accuracy and robustness of vulnerability detection.  Methods  The proposed method consists of four components: feature representation, feature extraction, feature fusion, and vulnerability detection (Fig. 1). First, vector representations of the code sequence and the Code Property Graph (CPG) are obtained. Using word embedding and node embedding techniques, the code sequence and graph nodes are mapped into fixed-dimensional vectors, which serve as inputs for subsequent feature extraction. Next, the pre-trained UniXcoder model is employed to capture contextual information and extract semantic features from the code. In parallel, a Residual Gated Graph Convolution Network (RGGCN) is applied to the CPG to capture complex structural information, thereby extracting graph structural features. To integrate these complementary representations, a Multimodal Attention Fusion Network (MAFN) is designed to model the interactions between semantic and structural features. This network generates informative fused features for the vulnerability detection task. Finally, a Multilayer Perceptron (MLP) performs classification on the semantic features, structural features, and fused features. An interpolated prediction classifier is then applied to optimize the detection process by balancing multiple prediction outcomes. By adaptively adjusting the model’s focus according to the characteristics of different code samples, the classifier enables the detection model to concentrate on the most critical features, thereby improving overall detection accuracy.  Results and Discussions  To validate the effectiveness of the proposed method, comparative experiments were conducted against baseline approaches on the Devign, Reveal, and SVulD datasets. The experimental results are summarized in (Tables 13). On the Devign dataset, the proposed method achieved an accuracy improvement of 1.38% over SCALE and a precision improvement of 5.19% over CodeBERT. On the Reveal dataset, it improved accuracy by 0.08% compared to SCALE, with precision being closest to that of SCALE. On the SVulD dataset, the method achieved an accuracy improvement of 0.13% over SCALE and a precision gain of 8.15% over Vul-LMGNNs. Collectively, these results demonstrate that the proposed method consistently yields higher accuracy and precision. This improvement can be attributed to its effective integration of semantic information extracted by UniXcoder and structural information captured by RGGCN. By contrast, CodeBERT and LineVul effectively learn code semantics but exhibit insufficient understanding of complex structural patterns, resulting in weaker detection performance. Devign and Reveal employ gated graph neural networks to capture structural information from code graphs but lack the ability to model semantic information contained in code sequences, which constrains their performance. Vul-LMGNNs attempt to improve detection performance by jointly learning semantic and structural features; however, their feature fusion strategy relies on simple concatenation. This approach fails to account for correlations between features, severely limiting the expressive power of the fused representation and reducing detection performance. In contrast, the proposed method fully leverages and integrates semantic and structural features through multimodal attention fusion. By modeling feature interactions rather than treating them independently, it achieves superior accuracy and precision, enabling more effective vulnerability detection.  Conclusions  Fully integrating code features across multiple dimensions can significantly enhance vulnerability detection performance. Compared with baseline methods, the proposed approach enables deeper modeling of interactions among code features, allowing the detection model to develop a more comprehensive understanding of code characteristics and thereby achieve superior detection accuracy and precision.
Multimodal Hypergraph Learning Guidance with Global Noise Enhancement for Sentiment Analysis under Missing Modality Information
HUANG Chen, LIU Huijie, ZHANG Yan, YANG Chao, SONG Jianhua
Available online  , doi: 10.11999/JEIT250649
Abstract:
  Objective  Multimodal Sentiment Analysis (MSA) has shown considerable promise in interdisciplinary domains such as Natural Language Processing (NLP) and Affective Computing, particularly by integrating information from ElectroEncephaloGraphy (EEG) signals, visual images, and text to classify sentiment polarity and provide a comprehensive understanding of human emotional states. However, in complex real-world scenarios, challenges including missing modalities, limited high-level semantic correlation learning across modalities, and the lack of mechanisms to guide cross-modal information transfer substantially restrict the generalization ability and accuracy of sentiment recognition models. To address these limitations, this study proposes a Multimodal Hypergraph Learning Guidance method with Global Noise Enhancement (MHLGNE), designed to improve the robustness and performance of MSA under conditions of missing modality information in complex environments.  Methods  The overall architecture of the MHLGNE model is illustrated in Figure 2 and consists of the Adaptive Global Noise Sampling Module, the Multimodal Hypergraph Learning Guiding Module, and the Sentiment Prediction Target Module. A pretrained language model is first applied to encode the multimodal input data. To simulate missing modality conditions, the input data are constructed with incomplete modal information, where a modality \begin{document}$ m\in \{e,v,t\} $\end{document} is randomly absent. The adaptive global noise sampling strategy is then employed to supplement missing modalities from a global perspective, thereby improving adaptability and enhancing both robustness and generalization in complex environments. This design allows the model to handle noisy data and missing modalities more effectively. The Multimodal Hypergraph Learning Guiding Module is further applied to capture high-level semantic correlations across different modalities, overcoming the limitations of conventional methods that rely only on feature alignment and fusion. By guiding cross-modal information transfer, this module enables the model to focus on essential inter-modal semantic dependencies, thereby improving sentiment prediction accuracy. Finally, the performance of MHLGNE is compared with that of State-Of-The-Art (SOTA) MSA models under two conditions: complete modality data and randomly missing modality information.  Results and Discussions  Three publicly available MSA datasets (SEED-IV, SEED-V, and DREAMER) are employed, with features extracted from EEG signals, visual images, and text. To ensure robustness, standard cross-validation is applied, and the training process is conducted with iterative adjustments to the noise sampling strategy, modality fusion method, and hypergraph learning structure to optimize sentiment prediction. Under the complete modality condition, MHLGNE is observed to outperform the second-best M2S model across most evaluation metrics, with accuracy improvements of 3.26%, 2.10%, and 0.58% on SEED-IV, SEED-V, and DREAMER, respectively. Additional metrics also indicate advantages over other SOTA methods. Under the random missing modality condition, MHLGNE maintains superiority over existing MSA approaches, with improvements of 1.03% in accuracy, 0.24% in precision, and 0.08 in Kappa score. The adaptive noise sampling module is further shown to effectively compensate for missing modalities. Unlike conventional models that suffer performance degradation under such conditions, MHLGNE maintains robustness by generating complementary information. In addition, the multimodal hypergraph structure enables the capture of high-level semantic dependencies across modalities, thereby strengthening cross-modal information transfer and offering clear advantages when modalities are absent. Ablation experiments confirm the independent contributions of each module. The removal of either the adaptive noise sampling or the multimodal hypergraph learning guiding module results in notable performance declines, particularly under high-noise or severely missing modality conditions. The exclusion of the cross-modal information transfer mechanism causes a substantial decline in accuracy and robustness, highlighting its essential role in MSA.  Conclusions  The MHLGNE model, equipped with the Adaptive Global Noise Sampling Module and the Multimodal Hypergraph Learning Guiding Module, markedly improves the performance of MSA under conditions of missing modalities and in tasks requiring effective cross-modal information transfer. Experiments on SEED-IV, SEED-V, and DREAMER confirm that MHLGNE exceeds SOTA MSA models across multiple evaluation metrics, including accuracy, precision, Kappa score, and F1 score, thereby demonstrating its robustness and effectiveness. Future work may focus on refining noise sampling strategies and developing more sophisticated hypergraph structures to further strengthen performance under extreme modality-missing scenarios. In addition, this framework has the potential to be extended to broader sentiment analysis tasks across diverse application domains.
Secrecy Rate Maximization Algorithm for IRS Assisted UAV-RSMA Systems
WANG Zhengqiang, KONG Weidong, WAN Xiaoyu, FAN Zifu, DUO Bin
Available online  , doi: 10.11999/JEIT250452
Abstract:
  Objective  Under the stringent requirements of Sixth-Generation(6G) mobile communication networks for spectral efficiency, energy efficiency, low latency, and wide coverage, Unmanned Aerial Vehicle (UAV) communication has emerged as a key solution for 6G and beyond, leveraging its Line-of-Sight propagation advantages and flexible deployment capabilities. Functioning as aerial base stations, UAVs significantly enhance network performance by improving spectral efficiency and connection reliability, demonstrating irreplaceable value in critical scenarios such as emergency communications, remote area coverage, and maritime operations. However, UAV communication systems face dual challenges in high-mobility environments: severe multi-user interference in dense access scenarios that substantially degrades system performance, alongside critical physical-layer security threats resulting from the broadcast nature and spatial openness of wireless channels that enable malicious interception of transmitted signals. Rate-Splitting Multiple Access (RSMA) mitigates these challenges by decomposing user messages into common and private streams, thereby providing a flexible interference management mechanism that balances decoding complexity with spectral efficiency. This makes RSMA especially suitable for high-density user access scenarios. In parallel, Intelligent Reflecting Surfaces (IRS) have emerged as a promising technology to dynamically reconfigure wireless propagation through programmable electromagnetic unit arrays. IRS improves the quality of legitimate links while reducing the capacity of eavesdropping links, thereby enhancing physical-layer security in UAV communications. It is noteworthy that while existing research has predominantly centered on conventional multiple access schemes, the application potential of RSMA technology in IRS-assisted UAV communication systems remains relatively unexplored. Against this background, this paper investigates secure transmission strategies in IRS-assisted UAV-RSMA systems.  Methods  This paper investigates the effect of eavesdroppers on the security performance of UAV communication systems and proposes an IRS-assisted RSMA-based UAV communication model. The system comprises a multi-antenna UAV base station, an IRS mounted on a building, multiple single-antenna legitimate users, and multiple single-antenna eavesdroppers. The optimization problem is formulated to maximize the system secrecy rate by jointly optimizing precoding vectors, common secrecy rate allocation, IRS phase shifts, and UAV positioning. The problem is highly non-convex due to the strong coupling among these variables, rendering direct solutions intractable. To overcome this challenge, a two-layer optimization framework is developed. In the inner layer, with UAV position fixed, an alternating optimization strategy divides the problem into two subproblems: (1) joint optimization of precoding vectors and common secrecy rate allocation and (2) optimization of IRS phase shifts. Non-convex constraints are transformed into convex forms using techniques such as Successive Convex Approximation (SCA), relaxation variables, first-order Taylor expansion, and Semidefinite Relaxation (SDR). In the outer layer, the Particle Swarm Optimization (PSO) algorithm determines the UAV deployment position based on the optimized inner-layer variables.  Results and Discussions  Simulation results show that the proposed algorithm outperforms RSMA without IRS, NOMA with IRS, and NOMA without IRS in terms of secrecy rate. (Fig. 2) illustrates that the secrecy rate increases with the number of iterations and converges under different UAV maximum transmit power levels and antenna configurations. (Fig. 3) demonstrates that increasing UAV transmit power significantly enhances the secrecy rate for both the proposed and benchmark schemes. This improvement arises because higher transmit power strengthens the signal received by legitimate users, increasing their achievable rates and enhancing system secrecy performance. (Fig. 4) indicates that the secrecy rate grows with the number of UAV antennas. This improvement is due to expanded signal coverage and greater spatial degrees of freedom, which amplify effective signal strength in legitimate user channels. (Fig. 5) shows that both the proposed scheme and NOMA with IRS achieve higher secrecy rate as the number of IRS reflecting elements increases. The additional elements provide greater spatial degrees of freedom, improving channel gains for legitimate users and strengthening resistance to eavesdropping. In contrast, benchmark schemes operating without IRS assistance exhibit no performance improvement and maintain constant secrecy rate. This result highlights the critical role of the IRS in enabling secure communications. Finally, (Fig. 6) demonstrates the optimal UAV position when \begin{document}${P_{\max }} = 30{\text{ dBm}}$\end{document}. Deploying the UAV near the center of legitimate users and adjacent to the IRS minimizes the average distance to users, thereby reducing path loss and fully exploiting IRS passive beamforming. This placement strengthens legitimate signals while suppressing the eavesdropping link, leading to enhanced secrecy performance.  Conclusions  This study addresses secure communication scenarios with multiple eavesdroppers by proposing an IRS-assisted secure resource allocation algorithm for UAV-enabled RSMA systems. An optimization problem is formulated to maximize the system secrecy rate under multiple constraints, including UAV transmit power, by jointly optimizing precoding vectors, common rate allocation, IRS configurations, and UAV positioning. Due to the non-convex nature of the problem, a hierarchical optimization framework is developed to decompose it into two subproblems. These are effectively solved using techniques such as SCA, SDR, Gaussian randomization, and PSO. Simulation results confirm that the proposed algorithm achieves substantial secrecy rate gains over three benchmark schemes, thereby validating its effectiveness.
HRIS-Aided Layered Sparse Reconstruction Hybrid Near- and Far-Field Source Localization Algorithm
YANG Qingqing, PU Xuelai, PENG Yi, LI Hui, YANG Qiuping
Available online  , doi: 10.11999/JEIT250429
Abstract:
  Objective  Advances in Reconfigurable Intelligent Surface (RIS) technology have enabled larger arrays and higher frequencies, which expand the near-field region and improve positioning accuracy. The fundamental differences between near- and far-field propagation necessitate hybrid localization algorithms capable of seamlessly integrating both regimes.  Methods  A localization framework for mixed near- and far-field sources is proposed by integrating Fourth-Order Cumulant (FOC) matrices with hierarchical sparse reconstruction. A hybrid RIS architecture incorporating active elements is employed to directly receive pilot signals, thereby reducing parameter-coupling errors that commonly occur in passive RIS over multi-hop channels and enhancing reliability in Non-Line-Of-Sight (NLOS) scenarios. Symmetrically placed active elements are employed to construct three FOC matrices for three-dimensional position estimation. The two-dimensional angle search is decomposed into two sequential one-dimensional searches, where elevation and azimuth are estimated separately to reduce computational complexity. The first FOC matrix (C1), formed from vertically symmetric elements, captures elevation characteristics. The second matrix (C2), constructed from centrally symmetric elements, suppresses nonlinear terms related to distance. The third matrix (C3) applies the previously estimated angles to select active elements, incorporates near-field effects, and enables accurate distance estimation as well as discrimination between near-field and far-field signals. To further improve the efficiency and accuracy of spectral searches, a hierarchical multi-resolution strategy based on sparse reconstruction is introduced. This method partitions the continuous parameter space into discrete intervals, incrementally generates a multi-resolution dictionary, and applies a progressive search procedure for precise position parameter estimation. During the search process, a tuning factor constrains the maximum reconstruction error between the sparse matrix and the projection of the original signal subspace. In addition, the algorithm exploits the orthogonality between the signal and noise subspaces to design a weight matrix, which reduces the effects of noise and position errors on the sparse solution. This hierarchical search enables rapid, coarse-to-fine parameter estimation and substantially improves localization accuracy.  Results and Discussions  The performance of the proposed algorithm is evaluated against Two-Stage Multiple Signal Classification (TSMUSIC), hybrid Orthogonal Matching Pursuit (OMP), and Holographic Multiple-Input Multiple-Output (HMIMO)-based methods with respect to noise resistance, convergence speed, and computational efficiency. Under varying SNR conditions (Fig. 5), traditional subspace methods exhibit degraded performance at low SNR because of reliance on signal–noise subspace orthogonality. In contrast, the proposed algorithm employs the FOC matrix to achieve accurate elevation and azimuth estimation while suppressing Gaussian noise. The hierarchical sparse reconstruction strategy further enhances estimation accuracy, resulting in superior far-field localization performance. Unlike the HMIMO-based algorithm, which depends on dynamic codebook switching, the proposed method retains nonlinear distance-dependent phase terms and constructs the distance codebook from initial angle estimates, thereby improving near-field localization accuracy. In Experiment 2, the effect of varying snapshot numbers on parameter estimation is examined. Owing to the angle-decoupling capability of the FOC matrix, the algorithm achieves rapid reduction in Root Mean Square Error (RMSE) even with a small number of snapshots. As the number of snapshots increases, estimation accuracy improves steadily and approaches convergence, indicating robustness against noise and fast convergence under low-snapshot conditions. Conventional methods typically require predefined near-field and far-field grids. By contrast, the nonlinear phase retention mechanism enables automatic discrimination between near-field and far-field sources without a predetermined distance threshold. While the nonlinear phase term introduces slightly slower convergence during distance decoupling, the proposed method still outperforms TSMUSIC and hybrid OMP. However, angle estimation errors during the decoupling process provide the HMIMO-based approach with a slight advantage in distance estimation accuracy (Fig. 6). Computational complexity is also compared between the hierarchical multi-resolution framework and traditional global search strategies (Fig. 7). Standard hybrid-field localization algorithms, such as TSMUSIC and hybrid OMP, require simultaneous optimization of angle and distance parameters, leading to exponential growth of computational cost. In contrast, the hierarchical strategy applies a phased search in which elevation and azimuth are estimated sequentially, reducing the two-dimensional angle spectrum search to two one-dimensional searches. The combination of progressive grid contraction, layer-by-layer tuning factors, and step-size decay narrows the search range efficiently, enabling rapid convergence through a three-layer dynamic grid structure. The distance dictionary constructed from angle estimates further removes redundant grids, thereby reducing complexity compared with global search methods.  Conclusions  This study presents a 3D localization framework for mixed near- and far-field sources in RIS-assisted systems by combining FOC decoupling with hierarchical sparse reconstruction. The method decouples angle and range estimation and uses a multi-resolution search strategy, achieving reliable performance and rapid convergence even under low SNR conditions and with limited snapshots. Simulation results demonstrate that the proposed approach consistently outperforms TSMUSIC, hybrid OMP, and HMIMO-based techniques, confirming its efficiency and robustness in mixed-field environments.
Optimal Federated Average Fusion of Gaussian Mixture–Probability Hypothesis Density Filters
XUE Yu, XU Lei
Available online  , doi: 10.11999/JEIT250759
Abstract:
  Objective  To realize optimal decentralized fusion tracking of uncertain targets, this study proposes a federated average fusion algorithm for Gaussian Mixture–Probability Hypothesis Density (GM-PHD) filters, designed with a hierarchical structure. Each sensor node operates a local GM-PHD filter to extract multi-target state estimates from sensor measurements. The fusion node performs three key tasks: (1) maintaining a master filter that predicts the fusion result from the previous iteration; (2) associating and merging the GM-PHDs of all filters; and (3) distributing the fused result and several parameters to each filter. The association step decomposes multi-target density fusion into four categories of single-target estimate fusion. We derive the optimal single-target estimate fusion both in the absence and presence of missed detections. Information assignment applies the covariance upper-bounding theory to eliminate correlation among all filters, enabling the proposed algorithm to achieve the accuracy of Bayesian fusion. Simulation results show that the federated fusion algorithm achieves optimal tracking accuracy and consistently outperforms the conventional Arithmetic Average (AA) fusion method. Moreover, the relative reliability of each filter can be flexibly adjusted.  Methods  The multi-sensor multi-target density fusion is decomposed into multiple groups of single-target component merging through the association operation. Federated filtering is employed as the merging strategy, which achieves the Bayesian optimum owing to its inherent decorrelation capability. Section 3 rigorously extends this approach to scenarios with missed detections. To satisfy federated filtering’s requirement for prior estimates, a master filter is designed to compute the predicted multi-target density, thereby establishing a hierarchical architecture for the proposed algorithm. In addition, auxiliary measures are incorporated to compensate for the observed underestimation of cardinality.  Results and Discussions  modified Mahalanobis distance (Fig.3). The precise association and the single-target decorrelation capability together ensure the theoretical optimality of the proposed algorithm, as illustrated in Fig. 2. Compared with conventional density fusion, the Optimal Sub-Pattern Assignment (OSPA) error is reduced by 8.17% (Fig. 4). The advantage of adopting a small average factor for the master filter is demonstrated in Figs. 5 and 6. The effectiveness of the measures for achieving cardinality consensus is also validated (Fig. 7). Another competitive strength of the algorithm lies in the flexibility of adjusting the average factors (Fig. 8). Furthermore, the algorithm consistently outperforms AA fusion across all missed detection probabilities (Fig. 9).  Conclusions  This paper achieves theoretically optimal multi-target density fusion by employing federated filtering as the merging method for single-target components. The proposed algorithm inherits the decorrelation capability and single-target optimality of federated filtering. A hierarchical fusion architecture is designed to satisfy the requirement for prior estimates. Extensive simulations demonstrate that: (1) the algorithm can accurately associate filtered components belonging to the same target, thereby extending single-target optimality to multi-target fusion tracking; (2) the algorithm supports flexible adjustment of average factors, with smaller values for the master filter consistently preferred; and (3) the superiority of the algorithm persists even under sensor malfunctions and high missed detection rates. Nonetheless, this study is limited to GM-PHD filters with overlapping Fields Of View (FOVs). Future work will investigate its applicability to other filter types and spatially non-overlapping FOVs.
Research Progress of Deep Learning Enabled Automatic Modulation Classification Technology
ZHENG Qinghe, LI Binglin, YU Zhiguo, JIANG Weiwei, ZHU Zhengyu, XU Chi, HUANG Chongwen, GUI Guan
Available online  , doi: 10.11999/JEIT250674
Abstract:
  Significance   With the advancement of sixth-generation (6G) wireless communication systems towards the terahertz frequency band and space–air–ground integrated networks, the communication environment is becoming increasingly heterogeneous and densely deployed. This evolution imposes stringent precision requirements at the sub-symbol period level for Automatic Modulation Classification (AMC). Under complex channel conditions, AMC faces several challenges: feature mixing and distortion caused by time-varying multipath channels, substantial degradation in recognition accuracy of traditional methods under low Signal-to-Noise Ratio (SNR) conditions, and elevated complexity in detecting mixed modulation signals introduced by Sparse Code Multiple Access (SCMA) techniques. Addressing these challenges, this paper first analyzes the fundamental constraints on AMC method design from the perspective of signal transmission characteristics in communication models. It then systematically reviews Deep Learning (DL)-based AMC approaches, summarizes the difficulties these methods encounter in different wireless communication scenarios, evaluates the performance of representative DL models, and concludes with a discussion of current limitations in AMC together with promising research directions.  Process   Current research on AMC technology under complex channel conditions mainly focuses on three methodological categories: Likelihood-Based (LB), Feature-Based (FB), and DL, emphasizing both theoretical exploration and algorithmic innovation. Among these, end-to-end DL approaches have demonstrated superior performance in AMC tasks. By stacking multiple layers of nonlinear activation functions, DL models establish strong nonlinear fitting capabilities that allow them to uncover hidden patterns in radio signals. This enables DL to achieve high robustness and accuracy in complex environments. Convolutional Neural Networks (CNNs), leveraging their hierarchical local perception mechanism, can effectively capture amplitude and phase distortion characteristics of modulated signals, showing distinctive advantages in spatial feature extraction. Recurrent Neural Networks (RNNs), through the temporal memory function of gated units, exhibit theoretical superiority in modeling dynamic signal impairments such as inter-symbol interference, carrier frequency offset, carrier phase offset, and timing errors. More recently, Transformer architectures have achieved global feature association modeling through self-attention mechanisms, thereby enhancing the ability to identify key features and markedly improving AMC accuracy under low SNR conditions. The application potential of Transformers in AMC can be further extended by integrating multi-scale feature fusion, optimizing computational efficiency, and improving generalization.  Prospects   With the continuous growth of communication demands and the increasing complexity of application scenarios, the efficient and reliable management and utilization of wireless spectrum resources has become a central research focus. AMC enables mobile communication systems to achieve dynamic channel adaptation and heterogeneous network integration. Driven by the development of space–air–ground integrated networks, the application scope of AMC has expanded beyond traditional terrestrial cellular systems to emerging domains such as satellite communication and vehicular networking. DL-based AMC frameworks can capture dynamic channel responses through joint time–frequency domain representations, enhance transient feature extraction via attention mechanisms, and effectively decouple the coupling effects of multipath fading and Doppler shifts. By applying neural architecture search and model quantization–compression techniques, DL models can achieve low-complexity, real-time inference at the edge, thereby supporting end-to-end latency control in Vehicle-to-Everything (V2X) communication links. Furthermore, advanced DL architectures introduce feature enhancement mechanisms to preserve signal phase integrity, improving resilience against channel distortion. In dynamic optical network monitoring, feature extraction networks tailored to time-varying channels can adaptively capture the evolution of nonlinear phase shifts. Through implicit channel compensation, DL enables collaborative learning of time-domain and frequency-domain features. At present, AMC technology is progressing towards elastic architectures that support dynamic reconstruction of model parameters through online knowledge distillation and meta-learning frameworks, offering adaptive and lightweight solutions for Internet-of-Things (IoT) scenarios.  Conclusions  This paper systematically reviews the current research and challenges of AMC technology in the context of 6G networks. First, the applications of CNNs, RNNs, Transformers, and hybrid DL models in AMC are discussed in detail, with analysis of the technical advantages and limitations of each approach. Next, three representative application scenarios are examined: the mobile communication, the optical communication, and the IoT, highlighting the specific challenges faced by AMC technology. At present, the development of DL-driven AMC has moved beyond model design to include deployment and application challenges in real wireless communication environments. For example, constructing DL architectures with continuous learning capabilities is essential for adapting to dynamic communication conditions, while developing large-scale DL models provides an effective way to improve cross-scenario generalization. Future research should emphasize directions that integrate prior knowledge of the physical layer with DL architectures, strengthen feature fusion strategies, and advance hardware–algorithm co-design frameworks.
A one-dimensional 5G millimeter-wave wide-angle Scanning Array Antenna Using AMC Structure
MA Zhangang, ZHANG Qing, FENG Sirun, ZHAO Luyu
Available online  , doi: 10.11999/JEIT250719
Abstract:
  Objective  With the rapid advancement of 5G millimeter-wave technology, antennas are required to achieve high gain, wide beam coverage, and compact size, particularly in environments characterized by strong propagation loss and blockage. Conventional millimeter-wave arrays often face difficulties in reconciling wide-angle scanning with high gain and broadband operation due to element coupling and narrow beamwidths. To overcome these challenges, this study proposes a one-dimensional linear array antenna incorporating an Artificial Magnetic Conductor (AMC) structure. The AMC’s in-phase reflection is exploited to improve bandwidth and gain while enabling wide-angle scanning of ±80° at 26 GHz. By adopting a 0.4-wavelength element spacing and stacked topology, the design provides an effective solution for 5G millimeter-wave terminals where spatial constraints and performance trade-offs are critical. The findings highlight the potential of AMC-based arrays to advance antenna technology for future high-speed, low-latency 5G applications by combining broadband operation, high directivity, and broad coverage within compact form factors.  Methods  This study develops a high-performance single-polarized one-dimensional linear millimeter-wave array antenna through a multi-layered structural design integrated with AMC technology. The design process begins with theoretical analysis of the pattern multiplication principle and array factor characteristics, which identify 0.4-wavelength element spacing as an optimal balance between wide-angle scanning and directivity. A stacked three-layer antenna unit is then constructed, consisting of square patch radiators on the top layer, a cross-shaped coupling feed structure in the middle layer, and an AMC-loaded substrate at the bottom. The AMC provides in-phase reflection in the 21–30 GHz band, enhancing bandwidth and suppressing surface wave coupling. Full-wave simulations (HFSS) are performed to optimize AMC dimensions, feed networks, and array layout, confirming bandwidth of 23.7–28 GHz, peak gain of 13.9 dBi, and scanning capability of ±80°. A prototype is fabricated using printed circuit board technology and evaluated with a vector network analyzer and anechoic chamber measurements. Experimental results agree closely with simulations, demonstrating an operational bandwidth of 23.3–27.7 GHz, isolation better than −15 dB, and scanning coverage up to ±80°. These results indicate that the synergistic interaction between AMC-modulated radiation fields and the array coupling mechanism enables a favorable balance among wide bandwidth, high gain, and wide-angle scanning.  Results and Discussions  The influence of array factor on directional performance is analyzed, and the maximum array factor is observed when the element spacing is between 0.4λ and 0.46λ (Fig. 2). The in-phase reflection of the AMC structure in the 21–30 GHz range significantly enhances antenna characteristics, broadening the bandwidth by 50% compared with designs without AMC and increasing the gain at 26 GHz by 1.5 dBi (Fig. 10, Fig. 13). The operational bandwidth of 23.3–27.7 GHz is confirmed by measurements (Fig. 17a). When the element spacing is optimized to 4.6 mm (0.4λ) and the coupling radiation mechanisms are adjusted, the H-plane half-power beamwidth (HPBW) of the array elements is extended to 180° (Fig. 8, Fig. 9), with a further gain improvement of 0.6 dBi at the scanning edges (Fig. 11b). The three-layer stacked structure—comprising the radiation, isolation, and AMC layers—achieves isolation better than –15 dB (Fig. 17a). Experimental validation demonstrates wide-angle scanning capability up to ±80°, showing close agreement between simulated and measured results (Fig. 11, Fig. 17b). The proposed antenna is therefore established as a compact, high-performance solution for 5G millimeter-wave terminals, offering wide bandwidth, high gain, and broad scanning coverage.  Conclusions  A one-dimensional linear wide-angle scanning array antenna based on an AMC structure is presented for 5G millimeter-wave applications. Through theoretical analysis, simulation optimization, and experimental validation, balanced improvement in broadband operation, high gain, and wide-angle scanning is achieved. Pattern multiplication theory and array factor analysis are applied to determine 0.4-wavelength element spacing as the optimal compromise between scanning angle and directivity. A stacked three-layer configuration is adopted, and the AMC’s in-phase reflection extends the bandwidth to 23.7–28.5 GHz, representing a 50% increase. Simulation and measurement confirm ±80° scanning at 26 GHz with a peak gain of 13.8 dBi, which is 1.3 dBi higher than that of non-AMC designs. The close consistency between experimental and simulated results verifies the feasibility of the design, providing a compact and high-performance solution for millimeter-wave antennas in mobile communication and vehicular systems. Future research is expected to explore dual-polarization integration and adaptation to complex environments.
Research on Federated Unlearning Approach Based on Adaptive Model Pruning
MA Zhenguo, HE Zixuan, SUN Yanjing, WANG Bowen, LIU Jianchun, XU Hongli
Available online  , doi: 10.11999/JEIT250503
Abstract:
  Objective  The rapid proliferation of Internet of Things (IoT) devices and the enforcement of data privacy regulations, including the General Data Protection Regulation (GDPR) and the Personal Information Protection Act, have positioned Federated Unlearning (FU) as a critical mechanism to safeguard the “right to be forgotten” in Edge Computing (EC). Existing class-level unlearning approaches often adopt uniform model pruning strategies. However, because edge nodes vary substantially in computational capacity, storage, and network bandwidth, these methods suffer from efficiency degradation, leading to imbalanced training delays and decreased resource utilization. This study proposes FU with Adaptive Model Pruning (FunAMP), a framework that minimizes training time while reliably eliminating the influence of target-class data. FunAMP dynamically assigns pruning ratios according to node resources and incorporates a parameter correlation metric to guide pruning decisions. In doing so, it addresses the challenge of resource heterogeneity while preserving compliance with privacy regulations.  Methods  The proposed framework establishes a quantitative relationship among model training time, node resources, and pruning ratios, on the basis of which an optimization problem is formulated to minimize overall training time. To address this problem, a greedy algorithm (Algorithm 2) is designed to adaptively assign appropriate pruning ratios to each node. The algorithm discretizes the pruning ratio space and applies a binary search strategy to balance computation and communication delays across nodes. Additionally, a Term Frequency–Inverse Document Frequency (TF–IDF)-based metric is introduced to evaluate the correlation between model parameters and the target-class data. For each parameter, the TF score reflects its activation contribution to the target class, whereas the IDF score measures its specificity across all classes. Parameters with high TF–IDF scores are iteratively pruned until the assigned pruning ratio is satisfied, thereby ensuring the effective removal of target-class data.  Results and Discussions  Simulation results confirm the effectiveness of FunAMP in balancing training efficiency and unlearning performance under resource heterogeneity. The effect of pruning granularity on model accuracy (Fig. 2): fine granularity (e.g., 0.01) preserves model integrity, whereas coarse settings degrade accuracy due to excessive parameter removal. Under fixed training time, FunAMP consistently achieves higher accuracy than FunUMP and Retrain (Fig. 3), as adaptive pruning ratios reduce inter-node waiting delays. For instance, FunAMP attains 76.48% accuracy on LeNet and 83.60% on AlexNet with FMNIST, outperforming baseline methods by 5.91% and 4.44%, respectively. The TF–IDF-driven pruning mechanism fully removes contributions of target-class data, achieving 0.00% accuracy on the target data while maintaining competitive performance on the remaining data (Table 2). Robustness under varying heterogeneity levels is further verified (Fig. 4). Compared with baselines, FunAMP markedly reduces the training time required to reach predefined accuracy and delivers up to 11.8× speedup across four models. These results demonstrate FunAMP’s capability to harmonize resource utilization, preserve model performance, and ensure unlearning efficacy in heterogeneous edge environments.  Conclusions  To mitigate training inefficiency caused by resource heterogeneity in FU, this study proposes FunAMP, a framework that integrates adaptive pruning with parameter relevance analysis. A system model is constructed to formalize the relationship among node resources, pruning ratios, and training time. A greedy algorithm dynamically assigns pruning ratios to edge nodes, thereby minimizing global training time while balancing computational and communication delays. Furthermore, a TF–IDF-driven metric quantifies the correlation between model parameters and target-class data, enabling the selective removal of critical parameters to erase target-class contributions. Theoretical analysis verifies the stability and reliability of the framework, while empirical results demonstrate that FunAMP achieves complete removal of target-class data and sustains competitive accuracy on the remaining classes. This work is limited to single-class unlearning, and extending the approach to scenarios requiring the simultaneous removal of multiple classes remains an important direction for future research.
Secure Beamforming Design for Multi-User Near-Field ISAC Systems
DENG Zhixiang, ZHANG Zhiwei
Available online  , doi: 10.11999/JEIT250462
Abstract:
  Objective  Integrated Sensing and Communication (ISAC) systems, a key enabling technology for 6G, achieve the joint realization of communication and sensing by sharing spectrum and hardware. However, radar targets may threaten the confidentiality of user communications, necessitating secure transmission against potential eavesdropping. At the same time, large-scale antenna arrays and high-frequency bands are expected to be widely deployed to meet future performance requirements, making near-field wireless transmission increasingly common. This trend creates a mismatch between existing ISAC designs that rely on the far-field assumption and the characteristics of real propagation environments. In this study, we design optimal secrecy beamforming for a multi-user near-field ISAC system to improve the confidentiality of user communications while ensuring radar sensing performance. The results show that distance degrees of freedom inherent in the near-field model, together with radar sensing signals serving as Artificial Noise (AN), provide significant gains in communication secrecy.  Methods  A near-field ISAC system model is established, in which multiple communication users and a single target, regarded as a potential eavesdropper, are located within the near-field region of a transmitter equipped with a Uniform Linear Array (ULA). Based on near-field channel theory, channel models are derived for all links, including the communication channels from the transmitter to the users, the transmitter to the target, and the radar echo-based sensing channel.The secrecy performance of each user is quantified as the difference between the achievable communication rate and the eavesdropping rate at the target, and the sum secrecy rate across all users is adopted as the metric for system-wide confidentiality. The sensing performance of the ISAC system is evaluated using the Cramér–Rao bound (CRB), obtained from the Fisher Information Matrix (FIM) for parameter estimation. To enhance secrecy, a joint optimization problem is formulated for the beamforming vectors of communication and radar sensing signals, with the objective of maximizing the sum secrecy rate under base station transmit power and sensing performance constraints.As the joint optimization problem is inherently non-convex, an algorithm combining Semi-Definite Relaxation (SDR) and Weighted Minimum Mean Square Error (WMMSE) is developed. The equivalence between the MMSE-transformed problem and the original secrecy rate maximization problem is first established to handle non-convexity. The CRB constraint is then expressed in convex form using the Schur complement. Finally, SDR is applied to recast the problem into a convex optimization framework, which allows a globally optimal solution to be derived.  Results and Discussions  Numerical evaluations show that the proposed near-field ISAC secrecy beamforming design achieves clear advantages in communication confidentiality compared with far-field and non-AN schemes. Under the near-field channel model, the designed beams effectively concentrate energy on legitimate users while suppressing information leakage through radar sensing signals (Fig. 3b). Even when communication users and radar targets are angularly aligned, the secure beamforming scheme attains spatial isolation through distance-domain degrees of freedom, thereby maintaining positive secrecy rates (Fig. 3a).Joint optimization of communication beams and radar sensing signals significantly improves multi-user secrecy rates while satisfying the CRB constraint. Compared with conventional AN-assisted methods, the proposed solution exhibits superior trade-off performance between sensing and communication (Fig. 4).The number of antennas is directly correlated with beam focusing performance: increasing the antenna count produces more concentrated beam patterns. In the near-field model, however, the incorporation of the distance dimension amplifies this effect, yielding larger performance gains than those observed in conventional far-field systems (Fig. 5).Raising the transmit power further improves the received signal quality at the users, which proportionally enhances system secrecy. The near-field scheme achieves more substantial gains than far-field baselines under higher transmit power conditions (Fig. 6).This paper also examines the effect of user population on secrecy performance. A larger number of users increases inter-user interference, which degrades overall secrecy (Fig. 7). Nevertheless, owing to the intrinsic interference suppression capability of the near-field scheme and the ability of AN to impair eavesdroppers’ decoding, the proposed method maintains stronger robustness against multi-user interference compared with conventional approaches.  Conclusions  This study investigates multi-user secure communication design in near-field ISAC systems and proposes a beamforming optimization scheme that jointly enhances sensing accuracy and communication secrecy. A non-convex optimization model is established to maximize the multi-user secrecy sum rate under base station transmit power and CRB constraints, where radar sensing signals are exploited as AN to impair potential eavesdroppers. To address the complexity of the problem, a joint optimization algorithm combining SDR and WMMSE is developed, which reformulates the original non-convex problem into a convex form solvable with standard optimization tools.
Detection and Localization of Radio Frequency Interference via Cross-domain Multi-feature from SAR Raw Data
FU Zewen, WEI Tingting, LI Ningning, LI Ning
Available online  , doi: 10.11999/JEIT250701
Abstract:
  Objective  The increasing congestion of the electromagnetic spectrum presents major challenges for Synthetic Aperture Radar (SAR) systems, where Radio Frequency Interference (RFI) can severely degrade imaging quality and compromise interpretation accuracy. Existing detection methods have critical limitations: time-domain approaches are insensitive to weak interference, whereas transform-domain methods perform poorly in characterizing broadband interference. This study develops a cross-domain framework that integrates complementary features from multiple domains, enabling robust RFI detection and accurate localization. The proposed approach addresses the deficiencies of single-domain methods and provides a reliable solution for operational SAR systems.  Methods  This study introduces two methodological innovations. First, a weighted feature fusion framework combines the first-order derivatives of time-domain kurtosis and skewness using Principal Component Analysis (PCA)-optimized weights, thereby capturing both global statistical distributions and local dynamic variations. Second, a differential time-frequency analysis technique applies the Short-Time Fourier Transform (STFT) with logarithmic ratio operations and adaptive thresholding to achieve sub-pulse interference localization. The overall workflow integrates K-means clustering for initial detection, STFT-based feature enhancement, binary region identification, and Inverse STFT (ISTFT) reconstruction. The proposed approach is validated against three state-of-the-art methods using both simulated data and Sentinel-1 datasets.  Results and Discussions  Experimental results demonstrate marked improvements across all evaluation metrics. For simulated data, the proposed method achieves a signal accuracy (SA) of 98.56% and a False Alarm (FA) rate of 0.65% (Table 2), representing a 3.13% gain in SA compared with conventional methods. The Root Mean Square Error (RMSE) reaches 0.1902 (Table 3), corresponding to a 10.9% improvement over existing techniques. Visual analysis further confirms more complete interference detection (Fig. 2) and cleaner suppression results (Figs. 4 and 7), with target features preserved. For measured data, the method maintains robust performance, achieving a gray entropy of 0.7843 (Table 5), and effectively mitigating the severe FAs observed in traditional approaches (Fig. 8).  Conclusions  In complex and dynamic electromagnetic environments, traditional RFI detection methods often show inaccuracies or even fail when processing NarrowBand Interference (NBI) or WideBand Interference (WBI), limiting their operational applicability. To address this challenge, this study proposes an engineering-oriented interference detection method designed for practical SAR operations. By combining time-domain kurtosis with the first derivative of skewness, the approach significantly enhances detection accuracy and adaptability. Furthermore, a localization technique is introduced that precisely identifies interference positions. Using time-frequency domain analysis, the method calculates differential values between the time-frequency representations of echo signals with and without interference, and determines interference locations through threshold-based judgment. Extensive simulations and Sentinel-1 experiments confirm the universality and effectiveness of the proposed method in both detection and localization.
Dual Mode Index Modulation-aided Orthogonal Chirp Division Multiplexing System in High-dynamic Scenes
NING Xiaoyan, TANG Zihan, YIN Qiaoling, WANG Shihan
Available online  , doi: 10.11999/JEIT250475
Abstract:
  Objective  In high-dynamic environments, the Orthogonal Chirp Division Multiplexing (OCDM) system has attracted significant attention due to its inherent advantage of time-frequency two-dimensional expansion gain. The OCDM with Index Modulation (OCDM-IM) system extends the index domain of the traditional OCDM system, selectively activating subcarriers through index modulation. This reduces inter-carrier interference to some extent. However, the OCDM-IM system necessitates that certain subcarriers remain inactive, which, on one hand, diminishes the time-frequency expansion gain of the OCDM system and, on the other hand, leads to more pronounced Doppler interference in high-dynamic environments. Additionally, the inactive subcarriers do not contribute to data transmission, resulting in throughput loss. To overcome these challenges, this study proposes a novel communication system architecture, the Dual Mode Index Modulation-aided OCDM (DM-OCDM-IM). This architecture incorporates a dual-mode index mapping scheme and introduces new modulation dimensions within the OCDM system. The DM-OCDM-IM system preserves the interference immunity associated with the time-frequency two-dimensional expansion of the OCDM system while achieving higher spectral efficiency with low-order constellation modulation, offering enhanced communication performance in high-dynamic scenarios.  Methods  In this study, a DM-OCDM-IM communication system architecture is proposed, consisting of two main components: the dual mode index modulation module and the receiving algorithm. In the dual mode index modulation module, the DM-OCDM-IM system partitions the subcarriers in each subblock into two groups, each transmitting constant-amplitude and mutually distinguishable constellation symbols. This design expands the modulation dimensions and improves spectral efficiency. At the same time, low-order constellation modulation can be applied in a single dimension, thereby strengthening the system’s anti-jamming capability in high-dynamic environments. The constant-amplitude dual mode index mapping scheme also reduces performance fluctuations caused by channel gain variations and offers ease of hardware implementation. For signal reception, the system must contend with substantial Doppler frequency shifts and the computational complexity of demodulation in high-dynamic conditions. To address this, the DM-OCDM-IM employs a receiving algorithm based on feature decomposition of the Discrete Fresnel Transform (DFnT), which reduces complexity. The discrete time-domain transmit signal is reconstructed by applying the Discrete Fourier Transform (DFT) and feature decomposition to the received frequency-domain signal. Finally, the original transmitted bits are recovered through index demodulation and constellation demodulation of the reconstructed time-domain signal using a maximum-likelihood receiver.  Results and Discussions  The performance of the proposed DM-OCDM-IM system is simulated and compared with that of the existed Dual Mode Index Modulation-aided OFDM (DM-OFDM-IM) system and the OCDM-IM system under three channel conditions: AWGN, multipath, and Doppler frequency shift. The results show that, relative to the DM-OFDM-IM system, the proposed DM-OCDM-IM system exploits multipath diversity more effectively and exhibits stronger resistance to fading in all three channels (Fig. 5, Fig. 6). When compared with the OCDM-IM system, the Bit Error Rate (BER) performance of the proposed DM-OCDM-IM system is significantly improved across all three channel conditions, particularly at high spectral efficiency (Fig.7(b), Fig.8(b)). These results confirm that the introduction of the dual mode index modulation technique extends the modulation dimensions within the OCDM framework. Information is transmitted not only through index modulation but also through dual mode modulation, enabling higher spectral efficiency without increasing the modulation order. At the same time, the time-frequency expansion gain characteristic of OCDM is preserved, while receiver complexity is effectively controlled. These combined features make the proposed DM-OCDM-IM system well suited for communication in high-dynamic channel environments.  Conclusions  This paper establishes a novel DM-OCDM-IM system framework. First, by integrating a constant-amplitude dual mode index mapping scheme into the traditional OCDM system, the proposed design expands the modulation dimensions and allows the use of low-order constellation modulation in a single dimension. This improves spectral efficiency while enhancing system reliability in high-dynamic environments. Second, to reduce receiver-side complexity, a receiving algorithm based on feature decomposition of the DFnT is proposed, simplifying the digital signal processing of the DM-OCDM-IM system. Finally, the performance of the system is evaluated under AWGN, multipath, and Doppler frequency shift channels. The results demonstrate that, compared with the existed DM-OFDM-IM system, the proposed DM-OCDM-IM system exhibits stronger resistance to multipath fading and Doppler frequency shifts. In comparison with the OCDM-IM system, the proposed DM-OCDM-IM design preserves the time-frequency expansion gain of OCDM and provides stronger fading resistance at high spectral efficiency. Therefore, the proposed DM-OCDM-IM system offers superior adaptability in high-dynamic scenarios and has the potential to serve as a next-generation physical-layer waveform for mobile communications.
Joint Focus Measure and Context-Guided Filtering for Depth From Focus
JIANG Ying, DENG Huiping, XIANG Sen, WU Jin
Available online  , doi: 10.11999/JEIT250540
Abstract:
  Objective  Depth from Focus (DFF) seeks to determine scene depth by analyzing the focus variation of each pixel in an image. A key challenge in DFF is identifying the best-focused slice within the focal stack. However, focus variation in weakly textured regions is often subtle, making it difficult to detect focused areas, which adversely affects the accuracy of depth maps. To address this issue, this study proposes a depth estimation network that integrates focus measures and contextual information from the focal stack. The network accurately locates the best-focused pixels and generates a reliable depth map. By explicitly incorporating focus cues into a Convolutional Neural Network (CNN) and thoroughly considering spatial correlations within the scene, the approach facilitates comprehensive inference of focus states in weakly textured regions. This enables the network to capture both local focus-related details and global contextual information, thereby enhancing the accuracy and efficiency of depth estimation in challenging regions.  Methods  The proposed network consists of two main components. The first is focus region detection, which extracts focus-related features from the focal stack. A focus measure operator is introduced into the network during learning, yielding the maximum response when an image region is in sharp focus. After identifying the best-focused slices within the stack, the detected focus features are fused with those extracted by a 2D CNN. Because focus variations in weakly textured regions are often subtle, the representation of focus regions is enhanced to improve sensitivity to such changes. The second component comprises a semantic network and a semantic context module. A semantic context network is used to extract semantic cues, and semantic-guided filtering is then applied to the focus volume, integrating target features (focus volume) with guiding features (semantic context features). When local focus cues are indistinguishable, the global semantic context allows reliable inference of the focus state. This framework combines the strengths of deep learning and traditional methods while accounting for the specific characteristics of DFF and CNN architectures. Therefore, it produces robust and accurate depth maps, even in challenging regions.  Results and Discussions  The proposed architecture is evaluated through quantitative and qualitative comparisons on two public datasets. Prediction reliability is assessed using multiple evaluation metrics, including Mean Squared Error (MSE) and squared relative error (Sqr.rel.). Quantitative results (Tables 1 and 2) show that the proposed method consistently outperforms existing approaches on both datasets. The small discrepancy between predicted and ground-truth depths indicates precise depth estimation with reduced prediction errors. In addition, higher accuracy is achieved while computational cost remains within a practical range. Qualitative analysis (Figures 10 and 11) further demonstrates superior depth reconstruction and detail preservation, even when a limited number of focal stack slices is used. The generalization ability of the network is further examined on the unlabeled Mobile Depth dataset (Figure 12). The results confirm that depth can be reliably recovered in diverse unseen scenes, indicating effectiveness for real-world applications. Ablation studies (Table 3) validate the contribution of each proposed module. Optimal performance is obtained when both the Focus Measure (FM) and the Semantic Context-Guided Module (SCGM) are applied. Parameter count comparisons further indicate that the proposed approach achieves a balance between performance and complexity, delivering robust accuracy without excessive computational burden.  Conclusions  This study proposes a CNN–based DFF framework to address the challenge of depth estimation in weakly textured regions. By embedding focus measure operators into the deep learning architecture, the representation of focused regions is enhanced, improving focus detection sensitivity and enabling precise capture of focus variations. In addition, the introduction of semantic context information enables effective integration of local and global focus cues, further increasing estimation accuracy. Experimental results across multiple datasets show that the proposed model achieves competitive performance compared with existing methods. Visual results on the Mobile Depth dataset further demonstrate its generalization ability. Nonetheless, the model shows limitations in extremely distant regions. Future work could incorporate multimodal information or frequency-domain features to further improve depth accuracy in weakly textured scenes.
Dynamic Inversion Algorithm for Rainfall Intensity Based on Dual-Mode Microwave Radar Combined Rain Gauge
ZHANG Qishuo, ZHANG Wenxin, GAO Mengyu, XIONG Fei
Available online  , doi: 10.11999/JEIT250535
Abstract:
  Objective  Microwave meteorological radar has broad application potential in rainfall detection due to its non-contact measurement, high spatiotemporal resolution, and multi-parameter retrieval capability. However, in the context of climate change, increasingly complex rainfall events require monitoring systems to deliver high-precision, multi-dimensional, real-time data to support disaster warning and climate research. Conventional single-mode radars, constrained by fixed functionalities, cannot fully meet these requirements, which has led to the development of multi-mode radar technology. The dual-mode radar examined in this study employs Frequency Modulated Continuous Wave (FMCW) and Continuous Wave (CW) modes. These modes adopt different algorithmic principles for raindrop velocity measurement: FMCW enables spatially stratified detection and strong anti-interference performance, whereas CW provides more accurate measurements of raindrop fall speed, yielding integral rainfall information in the vertical column. Despite these advantages, retrieval accuracy remains limited by the reliance of traditional algorithms on fixed empirical parameters, which restrict adaptability to regional climate variations and dynamic microphysical precipitation processes, and hinder real-time response to variations in rain Drop Size Distribution (DSD). Ground rain gauges, by contrast, provide near-true reference data through direct measurement of rainfall intensity. To address the above challenges, this paper proposes a dynamic inversion algorithm that integrates dual-mode (FMCW–CW) radar with rain gauge data, enhancing adaptability and retrieval accuracy for rainfall monitoring.  Methods  Two models are developed for the two radar modes. For the FMCW mode, which can retrieve DSD parameters, a fusion algorithm based on Attention integrated with a double-layer Long Short-Term Memory (LSTM) network (LSTM–Attention–LSTM) is proposed. The first LSTM extracts features from DSD data and rain gauge–measured rainfall intensity through its hidden state output, with a dropout layer applied to randomly discard neurons and reduce overfitting. The Attention mechanism calculates feature similarity using dot products and converts it into attention weights. The second LSTM then processes the time series and integrates the hidden-layer features, which are passed through a fully connected layer to generate the retrieval results. For the CW mode, which cannot directly retrieve DSD parameters and is constrained to the reflectivity factor–Rainfall rate (Z–R) relationship (Z=aRb), an algorithm based on the Extended Kalman Filter (EKF) is proposed to optimize this relationship. The method dynamically models the Z–R parameters, computes the residual between predicted rainfall intensity and rain gauge observations, and updates the prior estimates accordingly. Physical constraints are applied to parameters a and b during state updates to ensure consistency with physical laws, thereby enabling accurate fitting of the Z–R relationship.  Results and Discussions  Experimental results show that both models enhance the accuracy of rainfall intensity retrieval. For the FMCW mode, the LSTM–Attention–LSTM model applied to the test dataset outperforms traditional physical models, single-layer LSTM, and double-layer LSTM. It effectively captures the temporal variation of rainfall intensity, with the absolute error relative to observed values remaining below 0.25 mm/h (Fig. 5). Compared with the traditional physical model, the LSTM–Attention–LSTM reduces RMSE and MAE by 46% and 38%, achieving values of 0.1623 mm/h and 0.147 mm/h, respectively, and increases R2 by 14.5% to 0.95 (Table 2). For the CW mode, the Z–R relationship optimized by the EKF model provides the best fit for the Z and R distribution in the validation dataset (Fig. 6). Rainfall intensity retrieved with this algorithm on the test set exhibits the smallest deviation from actual observations compared with convective cloud empirical formulas, Beijing plain area empirical formulas, and the dynamic Z–R method. The corresponding RMSE, MAE, and R2 reach 0.1076 mm/h, 0.094 mm/h, and 0.972, respectively (Fig. 7; Table 4).  Conclusions  This study proposes two multi-source data fusion schemes that integrate dual-mode radar with rain gauges for short-term rainfall monitoring. Experimental results confirm that both methods significantly improve the accuracy of rainfall intensity retrieval and demonstrate strong dynamic adaptability and robustness.
Belief Propagation-Ordered Statistics Decoding Algorithm with Parameterized List Structures
LIANG Jifan, WANG Qianfan, SONG Linqi, LI Lvzhou, MA Xiao
Available online  , doi: 10.11999/JEIT250552
Abstract:
  Objective  Traditional Belief Propagation–Ordered Statistics Decoding (BP-OSD) algorithms for quantum error-correcting codes often rely on a single normalization factor (\begin{document}$ \alpha $\end{document}) in the Belief Propagation (BP) stage, which restricts the search space and limits decoding performance. An enhanced BP-OSD algorithm is presented to address this limitation by employing a list of candidate \begin{document}$ \alpha $\end{document} values. The central idea is to perform BP decoding iteratively for multiple \begin{document}$ \alpha $\end{document} values, with the resulting posterior probabilities post-processed by Ordered Statistics Decoding (OSD). To balance performance gains with computational tractability, the multi-\begin{document}$ \alpha $\end{document} BP-OSD process is embedded within a two-stage framework: the more computationally intensive parameter-listed decoding is activated only when an initial BP decoding with a fixed \begin{document}$ {\alpha }_{0} $\end{document} fails. This design broadens the parameter search to improve decoding performance, while conditional activation ensures that computational complexity remains manageable, particularly at low physical error rates.  Methods  The proposed enhanced BP-OSD algorithm (Algorithm 1) introduces a two-stage decoding process. In the first stage, decoding is attempted using standard BP with a single predetermined normalization factor (\begin{document}$ {\alpha }_{0} $\end{document}), providing a computationally efficient baseline. If this attempt fails to produce a valid syndrome match, the second stage is activated. In the second stage, parameter listing is employed: BP decoding is executed independently across a predefined list of \begin{document}$ L $\end{document} distinct normalization factors \begin{document}$ \left\{{\alpha }_{1},{\alpha }_{2}, \cdots,{\alpha }_{L}\right\} $\end{document}. Each run generates a set of posterior probabilities corresponding to a different BP operational point. These posterior probabilities are then individually post-processed by an OSD module, forming a pool of candidate error patterns. The final decoded output is selected from this pool according to the maximum likelihood criterion, or the minimum Hamming weight criterion under a depolarizing channel. Complexity analysis shows that this conditional two-stage design ensures that the average computational cost remains comparable to that of standard BP decoding, particularly at low physical error rates where the first stage frequently succeeds.  Results and Discussions  The effectiveness of the proposed algorithm is evaluated through Monte Carlo simulations on both Surface codes \begin{document}$ \left[\kern-0.15em\left[ {2{d}^{2}-2d+\mathrm{1,1},d} \right]\kern-0.15em\right] $\end{document} and Quantum Low-Density Parity-Check (QLDPC) codes \begin{document}$ \left[\kern-0.15em\left[ {\mathrm{882,24}} \right]\kern-0.15em\right] $\end{document} under a depolarizing channel. For Surface codes, the enhanced BP-OSD algorithm achieves a substantially lower logical error rate compared with both the Minimum-Weight Perfect Matching (MWPM) algorithm and the original BP algorithm (Fig. 4(a)). The error threshold is improved from approximately \begin{document}$ 15.5\% $\end{document} (MWPM) to about \begin{document}$ 18.3\% $\end{document} with the proposed method. The average decoding time comparison in Fig. 4(b) demonstrates that, particularly at low physical error rates, the proposed algorithm maintains a decoding speed comparable to the original BP algorithm. This efficiency results from the two-stage design, in which the more computationally intensive parameter-listed search is activated only when required. For QLDPC codes (Fig. 5(a), the proposed algorithm outperforms both the original BP and BP-OSD algorithms in terms of logical error rate, even when a smaller OSD candidate list per α value is employed. As shown in Table 3, increasing the parameter list size L (e.g., \begin{document}$ L=\mathrm{4,8},16 $\end{document}) improves decoding performance, although the gains diminish as L grows. This observation supports the choice of L = 16 as an effective balance between performance and complexity. Furthermore, the activation probability of the second stage (Table 2) decreases rapidly as the physical error rate declines, confirming the efficiency of the two-stage framework.  Conclusions  An enhanced BP-OSD algorithm for quantum error-correcting codes is presented, featuring a parameter-listing strategy for the normalization factor (\begin{document}$ \alpha $\end{document}) in the BP stage. Unlike conventional approaches that rely on a single \begin{document}$ \alpha $\end{document}, the proposed method explores multiple \begin{document}$ \alpha $\end{document} values, with the resulting posterior probabilities processed by the OSD module to select the most likely output. This systematic expansion of the search space improves decoding performance. To control computational overhead, a two-stage decoding mechanism is employed: the parameter-listed BP-OSD is activated only when an initial BP decoding with a fixed \begin{document}$ {\alpha }_{0} $\end{document} fails. Complexity analysis, supported by numerical simulations, shows that the average computational cost of the proposed algorithm remains comparable to that of standard BP decoding in low physical error rate regimes. Monte Carlo simulations further demonstrate its efficacy. For Surface codes, the enhanced BP-OSD achieves lower logical error rates than the MWPM algorithm and raises the error threshold from approximately 15.5% to 18.3%. For QLDPC codes, it exceeds both the original BP and BP-OSD algorithms in logical error rate performance, even with a reduced OSD candidate list size in the second stage. Overall, the proposed algorithm provides a promising pathway toward high-performance, high-threshold quantum error correction by balancing decoding power with operational efficiency, highlighting its potential for practical applications.
Efficient Storage Method for Real-Time Simulation of Wide-Range Multipath Delay Spread Channels
LI Weishi, ZHOU Hui, JIAO Xun, XU Qiang, TANG Youxi
Available online  , doi: 10.11999/JEIT250525
Abstract:
  Objective  The real-time channel emulator is a critical tool in wireless device research and development, enabling accurate and repeatable experiments in controlled laboratory environments. This capability reduces testing costs by avoiding extensive field trials and accelerates development cycles by allowing rapid iteration and validation of wireless devices under realistic conditions. With the rapid advancement of aerial platforms—including drones, High-Altitude Pseudo-Satellites (HAPS), and Unmanned Aerial Vehicles (UAVs)—for integrated sensing and communication, high-resolution imaging, and environmental reconstruction in complex wireless environments, the challenges of channel modeling have increased considerably. In particular, there is growing demand for real-time simulation of wide-range multipath delay spread channels. Existing simulation methods, although effective in traditional scenarios, face substantial limitations in hardware storage resources when handling such channels. This study addresses these limitations by proposing an efficient storage method for real-time emulation of wide-range multipath channels. The method reduces memory overhead while preserving high fidelity in channel reproduction, thereby offering a practical and optimized solution for next-generation wireless communication research.  Methods  In conventional real-time channel emulation, a combined simulation approach is adopted, employing cascaded common delay and multipath delay spread components. The common delay component is implemented using a single high-capacity memory module, whereas the multipath delay spread component is implemented using a Dense Tapped Delay Line (D-TDL). This design reduces storage resource requirements by multiplexing the common delay component, but the achievable multipath delay spread range remains limited. Moreover, the multipath delay is constrained by the common delay component, reducing flexibility and compromising the ability to emulate complex scenarios. The Sparse Tapped Delay Line (S-TDL) scheme is used in some algorithms to extend the multipath delay emulation range by cascading block memory modules. However, this method introduces inter-tap delay dependencies and cannot adapt to the requirements of wide-range multipath delay spread channels. Alternatively, Time-Division Multiplexing (TDM) is applied in other algorithms to improve the utilization efficiency of block memory modules and decouple multipath delay control. Despite this, TDM is constrained by the read/write bandwidth of memory, making it unsuitable for real-time channel emulation of large-bandwidth signals. To overcome the multi-tap delay coupling issue in the S-TDL algorithm, an Optimized Sparse Tapped Delay Line (OS-TDL) algorithm is proposed. By analyzing delay-dependent relationships among multipath taps, theoretical derivation establishes an analytical relationship between the number of multipaths and the delay spread range achievable under decoupling constraints. Redundant taps are introduced to eliminate inter-tap delay dependencies, enabling flexible configuration of arbitrary multipath delay combinations. The algorithm formulates a joint optimization model that balances hardware memory allocation and multipath delay spread fidelity, supports wide-range multipath scenarios without being limited by memory read/write bandwidth, and allows real-time emulation of large-bandwidth signals. The central innovation lies in dynamically constraining tap activation and sparsity patterns to reduce redundant memory while preserving wide-range multipath delay spread channel characteristics. Compared with conventional approaches, the proposed algorithm significantly enhances storage resource utilization efficiency in wide-range multipath channel emulation. On this basis, a concrete algorithmic procedure is developed, in which an input multipath delay sequence is computationally processed to derive delay configuration parameters and activation sequences for multiple cascaded memory units. Comprehensive validation procedures for the algorithm are presented in later sections.  Results and Discussions  Conventional S-TDL algorithms are constrained by inter-tap delay coupling, which limits their ability to achieve high-fidelity emulation of wide-range multipath delay variations. To overcome this limitation, a comparative simulation of three algorithms—the memory resource exclusive algorithm, the TDM memory resource algorithm, and the OS-TDL algorithm proposed herein—is systematically conducted. A controlled variable approach is employed to evaluate storage resource utilization efficiency across three key dimensions: signal sampling rate, number of emulated multipath components, and multipath delay spread range. Theoretical analysis and simulation results show that the proposed OS-TDL algorithm significantly reduces memory requirements compared with conventional methods, while maintaining emulation fidelity. Its effectiveness is further verified through experimental implementation on AMD’s Virtex UltraScale+ series high-performance Field-Programmable Gate Array (FPGA), using the XCVU13P verification platform. Comparative FPGA resource measurements under identical system specifications confirm the superiority of the proposed algorithm, demonstrating its ability to improve memory efficiency while accurately reproducing wide-range multipath delay spread channels.  Conclusions  This study addresses the challenge of storage resource utilization efficiency in real-time channel emulation for wide-range multipath delay spread by analyzing the inter-tap delay dependency inherent in conventional S-TDL algorithms. An OS-TDL algorithm is proposed to emulate wide-range multipath delay spread channels. Both simulation and hardware verification results demonstrate that the proposed algorithm substantially improves storage efficiency while accurately reproducing multipath wide-range delay spread characteristics. These findings confirm that the algorithm meets the design requirements of real-time channel emulators for increasingly complex verification scenarios.
Highly Dynamic Doppler Space Target Situation Awareness Algorithm for Spaceborne ISAR
ZHOU Yichen, WANG Yong, DING Wenjun
Available online  , doi: 10.11999/JEIT250667
Abstract:
  Objective  With the growing number of operational satellites in orbit, Space Situation Awareness (SSA) has become a critical capability for ensuring the safety of space operations. Traditional ground-based radar and optical systems face inherent limitations in tracking deep-space objects due to atmospheric interference and orbital obscuration. Therefore, spaceborne Inverse Synthetic Aperture Radar (ISAR) has emerged as a pivotal technology for on-orbit target characterization, offering all-weather, long-duration observation. However, higher-order Three-Dimensional (3D) spatial-variant range migration and phase errors, caused by the complex relative motion between a spaceborne ISAR platform and its target, can seriously degrade imaging quality. Meanwhile, conventional Two-Dimensional (2D) Range–Doppler (RD) imaging provides valuable intensity distributions of scattering points but remains a projection of the target’s 3D structure. The absence of geometric information limits accurate attitude estimation and collision risk assessment. To address these challenges and achieve more comprehensive SSA, this paper proposes a joint space target imaging and attitude estimation algorithm.  Methods  This paper proposes a joint space target imaging and attitude estimation algorithm composed of three main components: space target imaging characterization, high-resolution imaging, and attitude estimation. First, the imaging characteristics of satellite targets are analyzed to establish the mapping relationship between the image domain and the Doppler parameters of individual scattering points. Second, adaptive segmentation in the two-dimensional (2D) image domain combined with high-precision regional compensation is applied to obtain high-resolution imaging results. Finally, the spatial distribution characteristics of the Doppler parameters are exploited to derive an explicit expression for the second-order Doppler parameters and to estimate the planar component attitude of the target, such as that of the solar wing.  Results and Discussions  The proposed SSA method achieves high-resolution imaging even in the presence of orbital error and complex 3D spatial-variant Doppler error. Moreover, target attitude estimation can be performed without the need for rectangular component extraction. The effectiveness of the algorithm is verified through three simulation experiments. When the target adopts different attitudes, the method successfully produces both high-resolution imaging results and accurate target attitude estimation (Fig. 7, Fig. 8). To further evaluate performance, comparative simulations are conducted (Fig. 9, Fig. 10). In addition, a method for estimating the long- and short-edge pointing of the satellite solar wing is presented in Section 3.3. The effectiveness of the proposed high-precision imaging algorithm for spinning targets is analyzed in Section 3.4, where the third simulation demonstrates the extended SSA capability of the algorithm (Fig. 11, Fig. 12).  Conclusions  This paper proposes a joint high-resolution imaging and attitude estimation algorithm to address the situational awareness requirements of highly dynamic Doppler space targets. First, the imaging characteristics of satellite targets and the mapping relationship between scattering points and higher-order Doppler parameters are derived. Second, an adaptive region segmentation algorithm is developed to compensate for 3D spatial-variant errors, thereby significantly enhancing imaging resolution. Meanwhile, an explicit correlation between Doppler parameters and satellite attitude is established based on the characteristics of planar components. Simulation results under different imaging conditions confirm the validity and reliability of the algorithm. Compared with conventional approaches, the proposed method achieves joint compensation of orbital and rotational errors. Furthermore, the attitude estimation process does not require rectangular component segmentation and remains effective even when rectangular components are partially obscured.
The Research on Interference Suppression Algorithms for Millimeter-Wave Radar in Multi-Interference Environments
TAN Haonan, DONG Mei, CHEN Boxiao
Available online  , doi: 10.11999/JEIT250617
Abstract:
  Objective  With the widespread application of millimeter-wave radar in intelligent driving, mutual interference among radars has become increasingly prominent. Interference signals appear as sharp pulses in the time domain and elevated background noise in the frequency domain, severely degrading target information acquisition and threatening road traffic safety. To address this challenge, this paper proposes a joint envelope recovery–based signal reconstruction algorithm that exploits the time-domain characteristics of signals to enhance target detection performance in multi-interference environments.  Methods  The proposed algorithm consists of two core steps. Step 1: Interference region detection. A dual-criterion mechanism, combining interference envelope detection with transition point detection within the envelope, is employed. This approach substantially improves the accuracy of detecting both interference regions and useful signal segments in multi-interference environments. Step 2: Signal reconstruction. The detected useful signal segments and interference-free portions are used to reconstruct the interference regions. To ensure continuity and improve reconstruction accuracy, the Hilbert transform is applied to perform normalized envelope amplitude coordination on the reconstructed signal.  Results and Discussions  The algorithm first detects interference regions and useful signal segments with high precision through the dual-criterion mechanism, and then reconstructs the interference regions using the detected segments. Simulation results show that the algorithm achieves an interference detection accuracy of 93.7% and a useful signal segment detection accuracy of 97.2%, exceeding comparative algorithms (Table 3). The reconstructed signal effectively eliminates sharp interference pulses in the time domain, smooths the signal amplitude, and markedly improves the Signal-to-Interference-plus-Noise Ratio (SINR) in the frequency domain (Fig. 11). Compared with other interference suppression algorithms, the proposed method exhibits superior suppression performance (Fig. 12), achieving an SINR improvement of more than 3 dB in the frequency domain and maintaining better suppression effects across different SINR conditions (Fig. 13). In real-road tests, the algorithm successfully detects multiple interference regions and useful signal segments (Fig. 14) and significantly enhances the SINR after reconstruction (Fig. 15).  Conclusions  This paper proposes a joint envelope recovery–based signal reconstruction algorithm to address inaccurate target detection in multi-interference environments for millimeter-wave radar. The algorithm employs a dual-criterion mechanism to accurately detect interference regions and valid signal segments, and reconstructs the interference regions using the detected useful segments. The Hilbert transform is further applied to achieve collaborative normalization of the signal envelope. Experimental results demonstrate that the algorithm effectively identifies interference signals and reconstructs interference regions in multi-interference scenarios, significantly improving the signal-to-noise ratio, suppressing interference, and enabling accurate target information acquisition. These findings provide an effective anti-jamming solution for intelligent driving systems operating in multi-interference environments.
Research on Active Control Strategies for High-speed Train Pantographs Based on Reinforcement Learning-guided Model Predictive Control Algorithms
PENG Yuxiang, HAN Zhiwei, WANG Hui, HONG Weijia, LIU Zhigang
Available online  , doi: 10.11999/JEIT250343
Abstract:
  Objective  The coupled dynamics of the pantograph-catenary system are a critical determinant of current collection stability and the overall operational efficiency of high-speed trains. This study proposes an active control strategy that addresses complex operating conditions to mitigate fluctuations in pantograph-catenary contact force. Conventional approaches face inherent limitations: model-free Reinforcement Learning (RL) suffers from low sample efficiency and a tendency to converge to local optima, while Model Predictive Control (MPC) is constrained by its short optimization horizon. To integrate their complementary advantages, this paper develops a Reinforcement Learning-Guided Model Predictive Control (RL-GMPC) algorithm for active pantograph control. The objective is to design a controller that combines the long-term planning capability of RL with the online optimization and constraint-handling features of MPC. This hybrid framework is intended to overcome the challenges of sample inefficiency, short-sighted planning, and limited adaptability, thereby achieving improved suppression of contact force fluctuations across diverse operating speeds and environmental disturbances.  Methods   A finite element model of the pantograph-catenary system is established, in which a simplified three-mass pantograph model is integrated with nonlinear catenary components to simulate dynamic interactions. The reinforcement learning framework is designed with an adaptive latent dynamics model to capture system behavior and a robust reward estimation module to normalize multi-scale rewards. The RL-GMPC algorithm is formulated by combining MPC for short-term trajectory optimization with a terminal state value function for estimating long-term cumulative rewards, thus balancing immediate and future performance. A Markov decision process environment is constructed by defining the state variables (pantograph displacement, velocity, acceleration, and contact force), the action space (pneumatic lift force adjustment), and the reward function, which penalizes contact force deviations and abrupt control changes.  Results and Discussions   Experimental validation under Beijing-Shanghai line conditions demonstrates significant reductions in contact force standard deviations: 14.29%, 18.07%, 21.52%, and 34.87% at 290, 320, 350, and 380 km/h, respectively. The RL-GMPC algorithm outperforms conventional H∞ control and Proximal Policy Optimization (PPO) by generating smoother control inputs and suppressing high-frequency oscillations. Robustness tests under 20% random wind disturbances show a 30.17% reduction in contact force variations, confirming adaptability to dynamic perturbations. Cross-validation with different catenary configurations (Beijing-Guangzhou and Beijing-Tianjin lines) reveals consistent performance improvements, with deviations reduced by 17.04%~33.62% across speed profiles. Training efficiency analysis indicates that RL-GMPC requires 57% fewer interaction samples than PPO to achieve convergence, demonstrating superior sample efficiency.  Conclusions   The RL-GMPC algorithm integrates the predictive capabilities of model-based control with the adaptive learning strengths of reinforcement learning. By dynamically optimizing pantograph posture, it enhances contact stability across varying speeds and environmental disturbances. Its demonstrated robustness to parameter variations and external perturbations highlights its practical applicability in high-speed railway systems. This study establishes a novel framework for improving pantograph-catenary interaction quality, reducing maintenance costs, and advancing the development of next-generation high-speed trains.
Spatial Self-Attention Incorporated Imputation Algorithm for Severely Missing Multivariate Time Series
LIU Hui, FENG Haoran, MA Jiani, ZHENG Hongdang, ZHANG Lin
Available online  , doi: 10.11999/JEIT250220
Abstract:
  Objective  Multivariate time series data, characterized by their high dimensionality and temporal dynamics, are widely generated across diverse application domains, including healthcare monitoring, industrial sensor networks, and autonomous systems. However, these data are often subject to severe missingness caused by sensor malfunctions, transmission errors, or environmental disturbances, which obscures critical spatiotemporal patterns and hinders downstream analytical tasks such as anomaly detection, predictive maintenance, and decision support. Existing imputation methods, ranging from statistical approaches to machine learning models, are primarily tailored to low missing-rate scenarios. When applied to high missing-rate conditions, they face challenges such as gradient vanishing during model training, insufficient capture of spatiotemporal dependencies, and limited ability to represent complex nonlinear features, with performance deteriorating sharply as the missing rate increases. To address these limitations, this study proposes the Spatial Self-Attention Incorporated Imputation algorithm (SSAImpute), designed to enhance imputation performance specifically under severely missing conditions.  Methods  The proposed SSAImpute algorithm adopts a dual-branch Siamese architecture with adversarial fusion. Each branch comprises two core modules: a spatial self-attention-aware module and a subsequent temporal self-attention encoding module. The spatial self-attention-aware module constructs a dynamic adjacency matrix from the geolocations of data sources to explicitly quantify inter-variable spatial relationships. These spatial dependencies are then integrated with temporal features to strengthen sequence correlation modeling and enrich feature representations with embedded spatial information. The temporal self-attention encoding module employs a multi-dimensional residual attention mechanism with bidirectional temporal dependency learning. A missing-aware positional encoding scheme and a mask-adaptive self-attention mechanism are incorporated to effectively capture temporal dependencies and feature correlations, thereby mitigating severe missingness and alleviating the vanishing gradient problem. The two Siamese branches are fused through adversarial learning and dynamic weighting, which jointly refine the final imputation results. To evaluate the performance of SSAImpute against competing methods, three conventional metrics are used: Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Mean Relative Error (MRE).  Results and Discussions  Extensive experiments are conducted on four public datasets—Inter-Sensor, PeMS04, PeMS07, and PeMS11, in comparison with seven state-of-the-art time series imputation models: Mean, Median, K-Nearest Neighbor (KNN), Multi-directional Recurrent Neural Network (M-RNN), Bidirectional Recurrent Imputation for Time Series (BRITS), Transformer, and Self-Attention-based Imputation for Time Series (SAITS). The results show that the proposed method consistently improves imputation accuracy across all datasets, even under severe missingness. On the Inter-Sensor dataset, SSAImpute demonstrates superior performance compared with all competing methods. For all four time series, SSAImpute outperforms the others across all evaluation metrics, with improvements over the best-performing baseline (SAITS) of 15.9% in MAE, 0.19% in RMSE, and 16.6% in MRE for temperature; 11.1% in MAE, 1.04% in RMSE, and 13.2% in MRE for humidity; 9.8% in MAE, 10.2% in RMSE, and 24.3% in MRE for light; and 8.8% in MAE, 0.4% in RMSE, and 8.8% in MRE for voltage. On the PeMS datasets, SSAImpute also exceeds all competing methods across PeMS04, PeMS07, and PeMS11. The achieved MAE, RMSE, and MRE are 0.203, 0.328, and 22.4% for PeMS04; 0.153, 0.274, and 17.5% for PeMS07; and 0.180, 0.282, and 19.3% for PeMS11, respectively. The performance under different missing-ratio scenarios is further investigated. Although accuracy decreases exponentially with higher missingness, SSAImpute consistently outperforms the three strongest baselines. Visualization of the imputed time series further verifies its effectiveness, with reconstructed values closely aligned with the ground truth. These findings confirm the contributions of the spatial self-attention-aware module, the temporal self-attention encoding module, and the adversarial learning with dynamic weighting mechanism.  Conclusions  This study proposes a spatial self-attention-incorporated imputation method for severely missing multivariate time series data, built on a dual-branch Siamese framework. Each branch integrates a spatial self-attention-aware module, which incorporates geolocation information of the data source, followed by a temporal self-attention encoding mechanism to capture contextual dependencies. These modules jointly strengthen feature extraction of spatiotemporal dependencies, enabling more accurate reconstruction under high missingness. The proposed method provides a robust data foundation for downstream data-driven analysis and decision-making tasks in real-world applications.
Ultra-wideband Bonding Wire RF Characteristics Compensation IC and Circuit Design for Microwave Components
KONG Weidong, YAN Pengyi, LU Shaopeng, WANG Qiaonan, DENG Shixiong, LIN Peng, WANG Cong, YANG Guohui, ZHANG Kuang
Available online  , doi: 10.11999/JEIT250502
Abstract:
  Objective  In microwave modules, assembly gaps often occur between power amplifier chips and multilayer hybrid circuit boards or among different circuit units. These gaps form deep transition trenches that significantly degrade RF signal transmission quality, particularly at millimeter-wave frequencies. Bonding wires remain a critical solution for establishing electrical interconnections between RF chips and other structures. However, the inherent parasitic inductance of gold bonding wires adversely affects system performance. As RF modules increasingly operate in the Ka-band and W-band, the degradation caused by this parasitic inductance has become more pronounced. The problem is especially severe when the ground-signal return path is excessively long or when the bonding wires themselves are too long.  Methods  The impedance transformation paths of T-type and π-type matching networks are compared on the Smith chart. The analysis indicates that for a given parasitic inductance of bonding wires, the Q-circle of the π-type matching network is smaller, thereby enabling a broader matching bandwidth. A π-type matching network for chip-to-chip interconnection is realized by optimizing the bonding pad dimensions on the GaAs chip to provide capacitive loading. As the bonding pad size increases, more gold wires can be bonded to the chip, which simultaneously reduces the parasitic inductance of the wires. Additionally, a symmetric “Ground-Signal-Ground (GSG)” bonding pad structure is designed on the GaAs chip, which shortens the ground return path and further reduces the parasitic inductance of the bonding wires. By integrating these three design strategies, the proposed chip and transition structure are shown to substantially improve the performance of cross-deep-gap transitions between different circuit units in microwave modules.  Results and Discussions  The proposed chip and transition structure substantially improve the performance of cross-trench transitions between different circuit units in microwave modules (Fig. 7). Simulation results show that the interconnection architecture effectively mitigates the adverse effects of trench depth on RF characteristics. Experimental validation further confirms that the π-type matching network implemented with the designed chip achieves an ultra-wideband, high-performance cross-trench transition, with a return loss of ≥ 17 dB and an insertion loss of ≤ 0.7 dB over the DC~40 GHz frequency range.  Conclusions  Comparative analysis of impedance transformation paths between T-type and π-type matching networks demonstrates that in gold-wire bonding interconnections, the π-type configuration is more effective in mitigating the effect of bonding wire parasitic inductance on matching bandwidth, making it suitable for ultra-wideband bonded interconnection circuits. To implement the π-type matching network using GaAs technology, the bonding pad area on the chip is enlarged to provide capacitive loading and to allow additional bonding wires, thereby further reducing parasitic inductance. A GSG structure is also designed on the GaAs chip surface to modify the reference ground return path of the bonded interconnections, leading to additional reduction in parasitic inductance. By integrating these features, an ultra-wideband compensation chip is developed and applied to cross-trench transition structures in microwave modules. Experimental results indicate that for a transition structure with a trench depth of 2 mm and a width of 0.2 mm, the proposed design achieves high-performance characteristics from DC to 40 GHz, with return loss ≥ 17 dB and insertion loss ≤ 0.7 dB. When applied to interconnections between RF chips and circuit boards in microwave modules, the chip also significantly enhances the RF matching performance of bonded interconnections.
Lightweight Incremental Deployment for Computing-Network Converged AI Services
WANG Qinding, TAN bin, HUANG Guangping, DUAN Wei, YANG Dong, ZHANG Hongke
Available online  , doi: 10.11999/JEIT250663
Abstract:
  Objective   The rapid expansion of Artificial Intelligence (AI) computing services has heightened the demand for flexible access and efficient utilization of computing resources. Traditional Domain Name System (DNS) and IP-based scheduling mechanisms are constrained in addressing the stringent requirements of low latency and high concurrency, highlighting the need for integrated computing-network resource management. To address these challenges, this study proposes a lightweight deployment framework that enhances network adaptability and resource scheduling efficiency for AI services.  Methods   The AI-oriented Service IDentifier (AISID) is designed to encode service attributes into four dimensions: Object, Function, Method, and Performance. Service requests are decoupled from physical resource locations, enabling dynamic resource matching. AISID is embedded within IPv6 packets (Fig. 5), consisting of a 64-bit prefix for identification and a 64-bit service-specific suffix (Fig. 4). A lightweight incremental deployment scheme is implemented through hierarchical routing, in which stable wide-area routing is managed by ingress gateways, and fine-grained local scheduling is handled by egress gateways (Fig. 6). Ingress and egress gateways are incrementally deployed under the coordination of an intelligent control system to optimize resource allocation. AISID-based paths are encapsulated at ingress gateways using Segment Routing over IPv6 (SRv6), whereas egress gateways select optimal service nodes according to real-time load data using a weighted least-connections strategy (Fig. 8). AISID lifecycle management includes registration, query, migration, and decommissioning phases (Table 2), with global synchronization maintained by the control system. Resource scheduling is dynamically adjusted according to real-time network topology and node utilization metrics (Fig. 7).  Results and Discussions   Experimental results show marked improvements over traditional DNS/IP architectures. The AISID mechanism reduces service request initiation latency by 61.3% compared to DNS resolution (Fig. 9), as it eliminates the need for round-trip DNS queries. Under 500 concurrent requests, network bandwidth utilization variance decreases by 32.8% (Fig. 10), reflecting the ability of AISID-enabled scheduling to alleviate congestion hotspots. Computing resource variance improves by 12.3% (Fig. 11), demonstrating more balanced workload distribution across service nodes. These improvements arise from AISID’s precise semantic matching in combination with the hierarchical routing strategy, which together enhance resource allocation efficiency while maintaining compatibility with existing IPv6/DNS infrastructure (Fig. 23). The incremental deployment approach further reduces disruption to legacy networks, confirming the framework’s practicality and viability for real-world deployment.  Conclusions   This study establishes a computing-network convergence framework for AI services based on semantic-driven AISID and lightweight deployment. The key innovations include AISID’s semantic encoding, which enables dynamic resource scheduling and decoupled service access, together with incremental gateway deployment that optimizes routing without requiring major modifications to legacy networks. Experimental validation demonstrates significant improvements in latency reduction, bandwidth efficiency, and balanced resource utilization. Future research will explore AISID’s scalability across heterogeneous domains and its robustness under dynamic network conditions.
Multi-granularity Text Perception and Hierarchical Feature Interaction Method for Visual Grounding
CAI Hua, RAN Yue, FU Qiang, LI Junyan, ZHANG Chenjie, SUN Junxi
Available online  , doi: 10.11999/JEIT250387
Abstract:
  Objective  Visual grounding requires effective use of textual information for accurate target localization. Traditional methods primarily emphasize feature fusion but often neglect the guiding role of text, which limits localization accuracy. To address this limitation, a Multi-granularity Text Perception and Hierarchical Feature Interaction method for Visual Grounding (ThiVG) is proposed. In this method, the hierarchical feature interaction module is progressively incorporated into the image encoder to enhance the semantic representation of image features. The multi-granularity text-aware module is designed to generate weighted text with spatial and semantic enhancement, and a preliminary Hadamard product-based fusion strategy is applied to refine image features for cross-modal fusion. Experimental results show that the proposed method substantially improves localization accuracy and effectively alleviates the performance bottleneck arising from over-reliance on feature fusion modules in conventional approaches.  Methods  The proposed method comprises an image-text feature extraction network, a hierarchical feature interaction module, a multi-granularity text perception module, and a graphic-text cross-modal fusion and target localization network (Fig. 1). The image-text feature extraction network includes image and text branches for extracting their respective features (Fig. 2). In the image branch, text features are incorporated into the image encoder through the hierarchical feature interaction module (Fig. 3). This enables text information to filter and update image features, thereby strengthening their semantic expressiveness. The multi-granularity text perception module employs three perception mechanisms to fully extract spatial and semantic information from the text (Fig. 4). It generates weighted text, which is preliminarily fused with image features through a Hadamard product-based strategy, providing fine-grained image features for subsequent cross-modal fusion. The graphic-text cross-modal fusion module then deeply integrates image and text features using a Transformer encoder (Fig. 5), capturing their complex relationships. Finally, a Multilayer Perceptron (MLP) performs regression to predict the bounding box coordinates of the target location. This method not only achieves effective integration of image and text information but also improves accuracy and robustness in visual grounding tasks through hierarchical feature interaction and deep cross-modal fusion, offering a novel approach to complex localization challenges.  Results and Discussions  Comparison experiments demonstrate that the proposed method achieves substantial accuracy gains across five benchmark visual localization datasets (Tables 1 and 2), with particularly strong performance on the long-text RefCOCOg dataset. Although the model has a larger parameter size, comparisons of parameter counts and training-inference times indicate that its overall performance still exceeds that of traditional methods (Table 3). Ablation studies further verify the contribution of each key module (Table 4). The hierarchical feature interaction module improves the semantic representation of image features by incorporating textual information into the image encoder (Table 5). The multi-granularity text perception module enhances attention to key textual components through perception mechanisms and adaptive weighting (Table 6). By avoiding excessive modification of the text structure, it markedly strengthens the model’s capacity to process long text and complex sentences. Experiments on the number of encoder layers in the cross-modal fusion module show that a 6-layer deep fusion encoder effectively filters irrelevant background information (Table 7), yielding a more precise feature representation for the localization regression MLP. Generalization tests and visualization analyses further demonstrate that the proposed method maintains high adaptability and accuracy across diverse and challenging localization scenarios (Figs. 6, and 7).  Conclusions  This study proposes a visual grounding algorithm that integrates multi-granularity text perception with hierarchical feature interaction, effectively addressing the under-utilization of textual information and the reliance on single-feature fusion in existing approaches. Key innovations include the hierarchical feature interaction module in the image branch, which markedly enhances the semantic representation of image features; the multi-granularity text perception module, which fully exploits textual information to generate weighted text with spatial and semantic enhancement; and a preliminary Hadamard product-based fusion strategy, which provides fine-grained image representations for cross-modal fusion. Experimental results show that the proposed method achieves substantial accuracy improvements on classical vision datasets and demonstrates strong adaptability and robustness across diverse and complex localization scenarios. Future work will focus on extending this method to accommodate more diverse text inputs and further improving localization performance in challenging visual environments.
Electromagnetic Signal Feature Matching Characterization for Constant False Alarm Detection
WANG Zixin, XIANG Houhong, TIAN Bo, MA Hongwei, WANG Yuhao, ZENG Xiaolu, WANG Fengyu
Available online  , doi: 10.11999/JEIT250589
Abstract:
  Objective  Small targets such as unmanned aerial vehicles and unmanned vessels, which exhibit small Radar Cross Section (RCS) values and weak echoes, are difficult to detect due to their low observability. Traditional Constant False Alarm Rate (CFAR) detection is typically represented by the Cell-Averaged (CA) CFAR method, in which the detection threshold is determined by the statistical power parameter of the signal. However, its detection performance is constrained by the Signal-to-Noise Ratio (SNR). This study focuses on how to exploit and apply signal features beyond power parameters to achieve CFAR detection under lower SNR conditions.  Methods  After pulse compression, the envelope of a Linear Frequency Modulation (LFM) signal exhibits sinc characteristics, whereas noise retains its random nature. This difference can be used to distinguish target echoes from non-target signals. On this basis, we propose a constant false alarm detection method based on signal feature matching. First, both the ideal echo signal and the actual echo signal are processed with sliding windows of equal length to generate an ideal sample and a set of test samples. A dual-port fully connected neural network is then constructed to extract the deep feature matching degree between the ideal sample and the test samples. Finally, the constant false alarm threshold is obtained by numerically calculating the deep feature matching parameter from a large number of non-target samples compared with the standard sample.  Results and Discussions  Several sets of simulation experiments are carried out, and measured radar data from different frequency bands are applied to verify the effectiveness of the proposed method. The simulations first confirm that the method maintains stable constant false alarm characteristics (Table 1). The detection performance is then compared with traditional CA-CFAR detection, machine learning approaches, and other deep learning methods. The results indicate that, relative to CA-CFAR detection, the proposed method achieve 2–5 dB gain in equivalent SNR across different false alarm probabilities (Fig. 4). Under mismatched SNR conditions, the method continues to demonstrate robust detection performance with strong generalization capability (Fig. 5). In the processing of measured X-band radar data, the proposed method detects targets that CA-CFAR fails to identify, extending the detection range to 740 distance units, compared with 562 distance units for CA-CFAR, corresponding to an improvement of approximately 28.72% in radar detection capability (Fig. 7, 8). In the case of S-band radar data, the proposed method significantly reduces false alarms (Fig. 10, 11).  Conclusions  This study exploits the difference between target and noise signal envelopes by introducing a feature extraction network that effectively enhances target detection performance. Comparative simulation experiments and the processing of measured radar data across different frequency bands demonstrate the following: (1) the proposed method markedly improves detection performance over traditional CA-CFAR detection, yielding a 2–5 dB gain in equivalent SNR; (2) under mismatched SNR conditions, the method shows strong generalization capability, achieving better detection performance than other deep learning and machine learning approaches; (3) in X-band radar data processing, the method increases detection capability by approximately 28.72%; and (4) in S-band radar data processing, it significantly reduces false alarms. Future work will focus on accelerating the detection process to further improve efficiency.
A Two-Stage Framework for CAN Bus Attack Detection by Fusing Temporal and Deep Features
TAN Mingming, ZHANG Heng, WANG Xin, LI Ming, ZHANG Jian, YANG Ming
Available online  , doi: 10.11999/JEIT250651
Abstract:
  Objective  The Controller Area Network (CAN), the de facto standard for in-vehicle communication, is inherently vulnerable to cyberattacks. Existing Intrusion Detection Systems (IDSs) face a fundamental trade-off: achieving fine-grained classification of diverse attack types often requires computationally intensive models that exceed the resource limitations of on-board Electronic Control Units (ECUs). To address this problem, this study proposes a two-stage attack detection framework for the CAN bus that fuses temporal and deep features. The framework is designed to achieve both high classification accuracy and computational efficiency, thereby reconciling the tension between detection performance and practical deployability.  Methods  The proposed framework adopts a “detect-then-classify” strategy and incorporates two key innovations. (1) Stage 1: Temporal Feature-Aware Anomaly Detection. Two custom features are designed to quantify anomalies: Payload Data Entropy (PDE), which measures content randomness, and ID Frequency Mean Deviation (IFMD), which captures behavioral deviations. These features are processed by a Bidirectional Long Short-Term Memory (BiLSTM) network that exploits contextual temporal information to achieve high-recall anomaly detection. (2) Stage 2: Deep Feature-Based Fine-Grained Classification. Triggered only for samples flagged as anomalous, this stage employs a lightweight one-dimensional ParC1D-Net. The core ParC1D Block (Fig. 4) integrates depthwise separable one-dimensional convolution, Squeeze-and-Excitation (SE) attention, and a Feed-Forward Network (FFN), enabling efficient feature extraction with minimal parameters. Stage 1 is optimized using BCEWithLogitsLoss, whereas Stage 2 is trained with Cross-Entropy Loss.  Results and Discussions  The efficacy of the proposed framework is evaluated on public datasets. (1) State-of-the-art performance. On the Car-Hacking dataset (Table 5), an accuracy and F1-score of 99.99% are achieved, exceeding advanced baselines. On the more challenging Challenge dataset (Table 6), superior accuracy (99.90%) and a competitive F1-score (99.70% are also obtained. (2) Feature contribution analysis. Ablation studies (Tables 7 and 8) confirm the critical role of the proposed features. Removal of the IFMD feature results in the largest performance reduction, highlighting the importance of behavioral modeling. A synergistic effect is observed when PDE and IFMD are applied together. (3) Spatiotemporal efficiency. The complete model remains lightweight at only 0.39 MB. Latency tests (Table 9) demonstrate real-time capability, with average detection times of 0.62 ms on a GPU and 0.93 ms on a simulated CPU (batch size = 1). A system-level analysis (Section 3.5.4) further shows that the two-stage framework is approximately 1.65 times more efficient than a single-stage model in a realistic sparse-attack scenario.  Conclusions  This study establishes the two-stage framework as an effective and practical solution for CAN bus intrusion detection. By decoupling detection from classification, the framework resolves the trade-off between accuracy and on-board deployability. Its strong performance, combined with a minimal computational footprint, indicates its potential for securing real-world vehicular systems. Future research could extend the framework and explore hardware-specific optimizations.
YOLO-SCDI: A Vehicle Detection Algorithm Based on an Improved YOLOv8
WU Lin, CAO Wen
Available online  , doi: 10.11999/JEIT250238
Abstract:
  Objective  As a core task in computer vision, object detection is vital for intelligent transportation, supporting applications such as autonomous driving, Electronic Toll Collection (ETC), and traffic violation monitoring. However, complex urban environments—characterized by extreme weather, dense traffic occlusions, intense illumination, and reflective surfaces—pose substantial challenges, leading traditional methods to high false detection and missed detection rates. Despite recent progress, accuracy issues remain unresolved. To address these limitations, this study proposes YOLO-SCDI, a lightweight and effective vehicle detection model systematically optimized from YOLOv8 across four components: backbone network, neck structure, detection head, and loss function. These improvements significantly enhance detection accuracy and robustness in complex traffic conditions while maintaining model compactness and inference efficiency.  Methods  Building on the YOLOv8n architecture, four top-down optimization strategies are designed to balance detection accuracy, parameter efficiency, and lightweight deployment. First, to address the limited feature representation capacity, an attention-enhanced C2f-SCSA module (Fig. 4) is proposed. This module dynamically integrates local and global features through multi-scale convolutions and a dual spatial–channel attention mechanism, thereby improving the quality of input features. Second, to achieve effective multi-scale information integration while preserving both detailed and contextual features, a lightweight Cross-scale Feature Fusion Module (CCFM) is introduced into the Neck structure (Fig. 5). This results in the CCFM-Neck architecture, which reduces parameter size and enhances sensitivity to small-scale targets. Third, to mitigate the limitations of YOLOv8’s detection head—such as fixed feature fusion patterns and weak dynamic cross-scale interactions—a Dynamic Head module is incorporated. This module jointly models scale, spatial, and task attention, and includes a dynamic convolution-kernel generation network that adjusts convolution weights in real time according to input features. These improvements strengthen classification and regression feature responses, increasing the adaptability and discriminability of the detection head. Finally, because the CIoU loss function shows insufficient localization accuracy for small or irregular targets, ShapeIoU is adopted as the loss function. It is further improved using the Inner-IoU concept, which accelerates model convergence and enhances localization performance.  Results and Discussions  YOLO-SCDI is evaluated against mainstream detection models on the UA-DETRAC and BDD100K datasets. On the UA-DETRAC dataset (Table 4), YOLO-SCDI achieves an optimal balance between resource efficiency and detection performance. It requires only 2.37 M parameters and 7.6 GFLOPs—substantially fewer than competing models—while attaining 95.8% mAP@0.5, a 2.5% improvement over the baseline YOLOv8n and higher than most mainstream detectors. Under the stricter mAP@0.5:0.95 metric, YOLO-SCDI reaches 80.3%, clearly outperforming other lightweight designs. On the BDD100K dataset (Table 5), YOLO-SCDI improves mAP@0.5 and mAP@0.5:0.95 by 1.4% and 1.1%, respectively, compared with the baseline. These results are consistent with those from the UA-DETRAC dataset, confirming strong generalization and robustness. Detection results under varying illumination (Fig. 7) and adverse weather (Fig. 8) further validate its performance in realistic complex scenarios. Compared with models such as NanoDet, YOLOv12n, and YOLOv8n, YOLO-SCDI effectively reduces missed and false detections while providing higher-confidence predictions and more precise localization. Additionally, ablation studies (Table 3) confirm the contributions of the proposed C2f-SCSA, Dynamic Head, and Inner-ShapeIoU modules to performance gains. Collectively, these results demonstrate that YOLO-SCDI markedly enhances detection accuracy while maintaining a lightweight structure, thereby meeting practical requirements for vehicle detection in complex traffic environments.  Conclusions  This study proposes YOLO-SCDI, a vehicle detection algorithm built on an improved YOLOv8 framework. By optimizing the backbone network, neck structure, detection head, and loss function, the method enhances detection accuracy while substantially reducing model parameters. Experimental evaluations demonstrate that YOLO-SCDI exceeds existing approaches in both accuracy and model efficiency, making it well suited for practical vehicle detection tasks in complex traffic environments.
Recent Advances of Programmable Schedulers
ZHAO Yazhu, GUO Zehua, DOU Songshi, FU Xiaoyang
Available online  , doi: 10.11999/JEIT250657
Abstract:
  Objective  In recent years, diversified user demands, dynamic application scenarios, and massive data transmissions have imposed increasingly stringent requirements on modern networks. Network schedulers play a critical role in ensuring efficient and reliable data delivery, enhancing overall performance and stability, and directly shaping user-perceived service quality. Traditional scheduling algorithms, however, rely largely on fixed hardware, with scheduling logic hardwired during chip design. These designs are inflexible, provide coarse and static scheduling granularity, and offer limited capability to represent complex policies. Therefore, they hinder rapid deployment, increase upgrade costs, and fail to meet the evolving requirements of heterogeneous and large-scale network environments. Programmable schedulers, in contrast, leverage flexible hardware architectures to support diverse strategies without hardware replacement. Scheduling granularity can be dynamically adjusted at the flow, queue, or packet level to meet varied application requirements with precision. Furthermore, they enable the deployment of customized logic through data plane programming languages, allowing rapid iteration and online updates. These capabilities significantly reduce maintenance costs while improving adaptability. The combination of high flexibility, cost-effectiveness, and engineering practicality positions programmable schedulers as a superior alternative to traditional designs. Therefore, the design and optimization of high-performance programmable schedulers have become a central focus of current research, particularly for data center networks and industrial Internet applications, where efficient, flexible, and controllable traffic scheduling is essential.  Methods  The primary objective of current research is to design universal, high-performance programmable schedulers. Achieving simultaneous improvements across multiple performance metrics, however, remains a major challenge. Hardware-based schedulers deliver high performance and stability but incur substantial costs and typically support only a limited range of scheduling algorithms, restricting their applicability in large-scale and heterogeneous network environments. In contrast, software-based schedulers provide flexibility in expressing diverse algorithms but suffer from inherent performance constraints. To integrate the high performance of hardware with the flexibility of software, recent designs of programmable schedulers commonly adopt First-In First-Out (FIFO) or Push-In First-Out (PIFO) queue architectures. These approaches emphasize two key performance metrics: scheduling accuracy and programmability. Scheduling accuracy is critical, as modern applications such as real-time communications, online gaming, telemedicine, and autonomous driving demand strict guarantees on packet timing and ordering. Even minor errors may result in increased latency, reduced throughput, or connection interruptions, compromising user experience and service reliability. Programmability, by contrast, enables network devices to adapt to diverse scenarios, supporting rapid deployment of new algorithms and flexible responses to application-specific requirements. Improvements in both accuracy and programmability are therefore essential for developing efficient, reliable, and adaptable network systems, forming the basis for future high-performance deployments.  Results and Discussions  The overall packet scheduling process is illustrated in (Fig. 1), where scheduling is composed of scheduling algorithms and schedulers. At the ingress or egress pipelines of end hosts or network devices, scheduling algorithms assign a Rank value to each packet, determining the transmission order based on relative differences in Rank. Upon arrival at the traffic manager, the scheduler sorts and forwards packets according to their Rank values. Through the joint operation of algorithms and schedulers, packet scheduling is executed while meeting quality-of-service requirements. A comparative analysis of the fundamental principles of FIFO and PIFO scheduling mechanisms (Fig. 2) highlights their differences in queue ordering and disorder control. At present, most studies on programmable schedulers build upon these two foundational architectures (Fig. 3), with extensions and optimizations primarily aimed at improving scheduling accuracy and programmability. Specific strategies include admission control, refinement of scheduling algorithms, egress control, and advancements in data structures and queue mechanisms. On this basis, the current research progress on programmable schedulers is reviewed and systematically analyzed. Existing studies are compared along three key dimensions: structural characteristics, expressive capability, and approximation accuracy (Table 1).  Conclusions  Programmable schedulers, as a key technology for next-generation networks, enable flexible traffic management and open new possibilities for efficient packet scheduling. This review has summarized recent progress in the design of programmable schedulers across diverse application scenarios. The background and significance of programmable schedulers within the broader packet scheduling process were first clarified. An analysis of domestic and international literature shows that most current studies focus on FIFO-based and PIFO-based architectures to improve scheduling accuracy and programmability. The design approaches of these two architectures were examined, the main technical methods for enhancing performance were summarized, and their structural characteristics, expressive capabilities, and approximation accuracy were compared, highlighting respective advantages and limitations. Potential improvements in existing research were also identified, and future development directions were discussed. Nevertheless, the design of a universal, high-performance programmable scheduler remains a critical challenge. Achieving optimal performance across multiple metrics while ensuring high-quality network services will require continued joint efforts from both academia and industry.
Collaborative Inference for Large Language Models Against Jamming Attacks
LIN Zhiping, XIAO Liang, CHEN Hongyi, XU Xiaoyu, LI Jieling
Available online  , doi: 10.11999/JEIT250675
Abstract:
  Objective  Collaborative inference with Large Language Models (LLMs) is employed to enable mobile devices to offload multi-modal data, including images, text, video, and environmental information such as temperature and humidity, to edge servers. This offloading improves the performance of inference tasks such as human-computer question answering, logical reasoning, and decision support. Jamming attacks, however, increase transmission latency and packet loss, which reduces task completion rates and slows inference. A reinforcement learning-based collaborative inference scheme is proposed to enhance inference speed, accuracy, and task completion under jamming conditions. LLMs with different sparsity levels and quantization precisions are deployed on edge servers to meet heterogeneous inference requirements across tasks.  Methods  A reinforcement learning-based collaborative inference scheme is proposed to enhance inference accuracy, speed, and task completion under jamming attacks. The scheme jointly selects the edge servers, sparsity rates and quantization levels of LLMs, as well as the transmit power and channels for data offloading, based on task type, data volume, channel gains, and received jamming power. A policy risk function is formulated to quantify the probability of inference task failure given offloading latency and packet loss rate, thereby reducing the likelihood of unsafe policy exploration. Each edge server deploys LLMs with varying sparsity rates and quantization precisions, derived from layer-wise unstructured pruning and model parameter quantization, to process token vectors of multi-modal data including images, text, video, and environmental information such as temperature and humidity. This configuration is designed to meet diverse requirements for inference accuracy and speed across different tasks. The LLM inference system is implemented with mobile devices offloading images and text to edge servers for human-computer question answering and driving decision support. The edge servers employ a vision encoder and tokenizer to transform the received sensing data into token vectors, which serve as inputs to the LLMs. Pruning and parameter quantization are applied to the foundation model LLaVA-1.5-7B, generating nine LLM variants with different sparsity rates and quantization precisions to accommodate heterogeneous inference demands.  Results and Discussions  Experiments are conducted with three vehicles offloading images (i.e., captured traffic scenes) and texts (i.e., user prompts) using a maximum transmit power of 100 mW on 5.170~5.330 MHz frequency channels. The system is evaluated against a smart jammer that applies Q-learning to block one of the 20 MHz channels within this band. The results show consistent performance gains over benchmark schemes. Faster responses and more accurate driving advice are achieved, enabled by reduced offloading latency and lower packet loss in image transmission, which allow the construction of more complete traffic scenes. Over 20 repeated runs, inference speed is improved by 20.3%, task completion rate by 14.1%, and accuracy by 12.2%. These improvements are attributed to the safe exploration strategy, which prevents performance degradation and satisfies diverse inference requirements across tasks.  Conclusions  This paper proposed a reinforcement learning-based collaborative inference scheme that jointly selects the edge servers, sparsity rates and quantization levels of LLMs, as well as the transmit power and offloading channels, to counter jamming attacks. The inference system deploys nine LLM variants with different sparsity rates and quantization precisions for human-computer question answering and driving decision support, thereby meeting heterogeneous requirements for accuracy and speed. Experimental results demonstrate that the proposed scheme provides faster responses and more reliable driving advice. Specifically, it improves inference speed by 20.3%, task completion rate by 14.1%, and accuracy by 12.2%, achieved through reduced offloading latency and packet loss compared with benchmark approaches.
Parametric Holographic MIMO Channel Modeling and Its Bayesian Estimation
YUAN Zhengdao, GUO Yabo, GAO Dawei, GUO Qinghua, HUANG Chongwen, LIAO Guisheng
Available online  , doi: 10.11999/JEIT250436
Abstract:
  Objective  Holographic Multiple-Input Multiple-Output (HMIMO), based on continuous-aperture antennas and programmable metasurfaces, is regarded as a cornerstone of 6G wireless communication. Its potential to overcome the limitations of conventional massive MIMO is critically dependent on accurate channel modeling and estimation. Three major challenges remain: (1) oversimplified electromagnetic propagation models, such as far-field approximations, cause severe mismatches in near-field scenarios; (2) statistical models fail to characterize the coupling between channel coefficients, user positions, and random orientations; and (3) the high dimensionality of parameter spaces results in prohibitive computational complexity. To address these challenges, a hybrid parametric-Bayesian framework is proposed in which neural networks, factor graphs, and convex optimization are integrated. Precise channel estimation, user position sensing, and angle decoupling in near-field HMIMO systems are thereby achieved. The methodology provides a pathway toward high-capacity 6G applications, including Integrated Sensing And Communication (ISAC).  Methods  A hybrid channel estimation method is proposed to decouple the “channel-coordinate-angle” parameters and to enable joint estimation of channel coefficients, coordinates, and angles under random user orientations. A neural network is first employed to capture the nonlinear relationship between holographic channel characteristics and the relative coordinates of the base station and user. The trained network is then embedded into a factor graph, where global optimization is performed. The neural network is dynamically approximated through Taylor expansion, allowing bidirectional message propagation and iterative refinement of parameter estimates. To address random user orientations, Euler angle rotation theory is introduced. Finally, convex optimization is applied to estimate the rotation mapping matrix, resulting in the decoupling of coordinate and angle parameters and accurate channel estimation.  Results and Discussions  The simulations evaluate the performance of different algorithms under varying key parameters, including Signal-to-Noise Ratio (SNR), pilot length L, and base station antenna number M. Two performance metrics are considered: Normalized Mean Square Error (NMSE) of channel estimation and user positioning accuracy, with the Cramér–Rao Lower Bound (CRLB) serving as the theoretical benchmark. At an SNR of 10 dB, the proposed method achieves a channel NMSE below –40 dB, outperforming Least Squares (LS) estimation and approximate model-based approaches. Under high SNR conditions, the NMSE converges toward the CRLB, confirming near-optimal performance (Fig. 5a). The proposed channel model demonstrates superior performance over “approximate methods” due to its enhanced characterization of real-world channels. Moreover, the positioning error gap between the proposed method and the “parallel bound” narrows to nearly 3 dB at high SNR, confirming the accuracy of angle estimation and the effectiveness of parameter decoupling (Fig. 5b). Moreover, the proposed method maintains performance close to the theoretical bounds when system parameters, such as user antenna number N, base station antenna number M, and pilot length L, are varied, demonstrating strong robustness (Figs. 68). These results also show that the Euler angle rotation-based estimation effectively compensates for coordinate offsets induced by random user orientations.  Conclusions  This study proposes a framework for HMIMO channel estimation by integrating neural networks, factor graphs, and convex optimization. The main contributions are threefold. First, Euler angles and coordinate mapping are incorporated into the parameterized channel model through factorization and factor graphs, enabling channel modeling under arbitrary user antenna orientations. Second, neural networks and convex optimization are embedded as factor nodes in the graph, allowing nonlinear function approximation and global optimization. Third, bidirectional message passing between neural network and convex optimization nodes is realized through Taylor expansion, thereby achieving joint decoupling and estimation of channel parameters, coordinates, and angles. Simulation results confirm that the proposed framework achieves higher accuracy—exceeding benchmarks by more than 3 dB, and demonstrates strong robustness across a range of scenarios. Future work will extend the method to multi-user environments, incorporate polarization diversity, and address hardware impairments such as phase noise, with the aim of supporting practical deployment in 6G systems.
Multi-Mode Anti-Jamming for UAV Communications: A Cooperative Mode-Based Decision-Making Approach via Two-Dimensional Transfer Reinforcement Learning
WANG Shiyu, WANG Ximing, KE Zhenyi, LIU Dianxiong, LIU Jize, DU Zhiyong
Available online  , doi: 10.11999/JEIT250566
Abstract:
  Objective  With the widespread application of Unmanned Aerial Vehicles (UAVs) in military reconnaissance, logistics, and emergency communications, ensuring the security and reliability of UAV communication systems has become a critical challenge. Wireless channels are highly vulnerable to diverse jamming attacks. Traditional anti-jamming techniques, such as Frequency-Hopping Spread Spectrum (FHSS), are limited in dynamic spectrum environments and may be compromised by advanced machine learning algorithms. Furthermore, UAVs operate under strict constraints on onboard computational power and energy, which hinders the real-time use of complex anti-jamming algorithms. To address these challenges, this study proposes a multi-mode anti-jamming framework that integrates Intelligent Frequency Hopping (IFH), Jamming-based Backscatter Communication (JBC), and Energy Harvesting (EH) to strengthen communication resilience in complex electromagnetic environments. A Multi-mode Transfer Deep Q-Learning (MT-DQN) method is further proposed, enabling two-dimensional transfer to improve learning efficiency and adaptability under resource constraints. By leveraging transfer learning, the framework reduces computational load and accelerates decision-making, thereby allowing UAVs to counter jamming threats effectively even with limited resources.  Methods  The proposed framework adopts a multi-mode anti-jamming architecture that integrates IFH, JBC, and EH to establish a comprehensive defense strategy of “avoiding, utilizing, and converting” interference. The system is formulated as a Markov Decision Process (MDP) to dynamically optimize the selection of anti-jamming modes and communication channels. To address the challenges of high-dimensional state-action spaces and restricted onboard computational resources, a two-dimensional transfer reinforcement learning framework is developed. This framework comprises a cross-mode strategy-sharing network for extracting common features across different anti-jamming modes (Fig. 3) and a parallel network for cross-task transfer learning to adapt to variable task requirements (Fig. 4). The cross-mode strategy-sharing network accelerates convergence by reusing experiences, whereas the cross-task transfer learning network enables knowledge transfer under different task weightings. The reward function is designed to balance communication throughput and energy consumption. It guides the UAV to select the optimal anti-jamming strategy in real time based on spectrum sensing outcomes and task priorities.  Results and Discussions  The simulation results validate the effectiveness of the proposed MT-DQN. The dynamic weight allocation mechanism exhibits strong cross-task transfer capability (Fig. 6), as weight adjustments enable rapid convergence toward the corresponding optimal reward values. Compared with conventional Deep Reinforcement Learning (DRL) algorithms, the proposed method achieves a 64% faster convergence rate while maintaining the probability of communication interruption below 20% in dynamic jamming environments (Fig. 7). The framework shows robust performance in terms of throughput, convergence rate, and adaptability to variations in jamming patterns. In scenarios with comb-shaped and sweep-frequency jamming, the proposed method yields higher normalized throughput and faster convergence, exceeding baseline DQN and other transfer learning-based approaches. The results also indicate that MT-DQN improves stability and accelerates policy optimization during jamming pattern switching (Fig. 7), highlighting its adaptability to abrupt changes in jamming patterns through transfer learning.  Conclusions  This study proposes a multi-modal anti-jamming framework that integrates IFH, JBC, and EH, thereby enhancing the communication capability of UAVs. The proposed solution shifts the paradigm from traditional jamming avoidance toward active jamming exploitation, repurposing jamming signals as covert carriers to overcome the limitations of conventional frequency-hopping systems. Simulation results confirm the advantages of the proposed method in throughput performance, convergence rate, and environmental adaptability, demonstrating stable communication quality even under complex electromagnetic conditions. Although DRL approaches are inherently constrained in handling completely random jamming without intrinsic patterns, this work improves adaptability to dynamic jamming through transfer learning and cross-modal strategy sharing. These findings provide a promising approach for countering complex jamming threats in UAV networks. Future work will focus on validating the proposed algorithm in hardware implementations and enhancing the robustness of DRL methods under highly non-stationary, though not entirely unpredictable, jamming conditions such as pseudo-random or adaptive interference.
Quasi-Vortex Electromagnetic Wave Radar Forward Looking based on Echo Phase Weighting
SHU Gaofeng, WEI Yixin, LI Ning
Available online  , doi: 10.11999/JEIT250542
Abstract:
  Objective  Forward-looking radar imaging plays a critical role in multiple applications. Numerous algorithms have been proposed to enhance azimuth resolution; however, improvement remains difficult due to the limitations imposed by antenna aperture. Existing high-resolution techniques, including synthetic aperture radar and Doppler beam sharpening, rely on Doppler bandwidth and inevitably create blind spots in the forward-looking region. Vortex electromagnetic waves carrying orbital angular momentum offer potential in forward-looking scenarios because of the orthogonality between different orbital angular momentum modes. In conventional vortex electromagnetic wave imaging, a Uniform Circular Array (UCA) is used to generate and transmit multi-mode vortex electromagnetic waves. Yet, the UCA-generated waves suffer from main lobe divergence, which disperses energy and weakens echo signals, while multi-mode transmission increases system complexity. To address these issues, this paper proposes a Quasi-Circular Array (QCA) that reduces system complexity, produces vortex electromagnetic waves with more concentrated main lobes, and preserves phase linearity. In addition, a post-processing method based on echo phase weighting is introduced. By applying phase modulation to the single-mode echo received by each antenna element, a complete equivalent multi-mode echo is synthesized. The proposed method enhances azimuth resolution and exhibits strong anti-noise performance.  Methods  To obtain clear images under low Signal-to-Noise Ratio (SNR) conditions, a phase modulation echo post-processing method combined with a QCA is proposed. The QCA first generates a single-mode vortex electromagnetic wave to illuminate the region of interest. Each element of the array then receives and stores the echo. Phase modulation is subsequently applied to the stored echo to generate signals of specific modes, thereby synthesizing an equivalent multi-mode echo with enhanced amplitude that preserves target information. This approach demonstrates strong potential for practical applications in forward-looking radar imaging under low SNR conditions.  Results and Discussions  When noise is added to the echo and imaging is performed (Figure 11), the proposed method achieves superior results under noisy conditions. As noise intensity increases, a clear target can still be reconstructed at a SNR of –10 dB. Even when the SNR is reduced to –15 dB and the target is submerged in noise, the contour features of the reconstructed target remain distinguishable. These results demonstrate that the method has strong anti-noise performance. In addition, when imaging is performed within a smaller mode range, the azimuth resolution achieved by the proposed method improves by an average factor of 2.2 compared with the traditional method (Figure 9). The improvements in resolution and anti-noise performance can be attributed to two factors: (1) The vortex electromagnetic waves generated by the QCA experience reduced destructive interference due to the asymmetric spatial distribution of array elements, producing waves with more concentrated main lobes, lower side lobes, and higher radiation gain. (2) Applying phase modulation in echo processing reduces the pulse repetition frequency of the vortex electromagnetic wave at the transmitting end, thereby lowering system complexity.  Conclusions  This study proposes a method capable of effective imaging under low SNR conditions. The echo expression of the electric field generated by the QCA is derived, and the radiation gain and phase characteristics of the quasi-vortex electromagnetic wave are analyzed. In addition, an echo post-processing method based on phase modulation is introduced. Simulation results demonstrate that, compared with the traditional UCA method, the proposed approach generates vortex electromagnetic waves with more concentrated main lobes, lower side lobes, and higher gain, while improving azimuth resolution by a factor of 2.2. Even at a SNR of –15 dB, the reconstructed imaging results remain distinguishable.
A Multi-class Local Distribution-based Weighted Oversampling Algorithm for Multi-class Imbalanced Datasets
TAO Xinmin, XU Annan, SHI Lihang, LI Junxuan, GUO Xinyue, ZHANG Yanping
Available online  , doi: 10.11999/JEIT250381
Abstract:
  Objective  Classification with imbalanced datasets remains one of the most challenging problems in machine learning. In addition to class imbalance, such datasets often contain complex factors including class overlap, small disjuncts, outliers, and low-density regions, all of which can substantially degrade classifier performance, particularly in multi-class settings. To address these challenges simultaneously, this study proposes the Multi-class Local Distribution-based Weighted Oversampling Algorithm (MC-LDWO).  Methods  The MC-LDWO algorithm first constructs hyperspheres centered on dynamically determined minority classes, with radii estimated from the distribution of each class. Within these hyperspheres, minority class samples are selected for oversampling according to their local distribution, and an adaptive weight allocation strategy is designed using local density metrics. This ensures that samples in low-density regions and near class boundaries are assigned higher probabilities of being oversampled. Next, a low-density vector is computed from the local distribution of both majority and minority classes. A random vector is then introduced and integrated with the low-density vector, and a cutoff threshold is applied to determine the generation sites of synthetic samples, thereby reducing class overlap during boundary oversampling. Finally, an improved decomposition strategy tailored for multi-class imbalance is employed to further enhance classification performance in multi-class imbalanced scenarios.  Results and Discussions  The MC-LDWO algorithm dynamically identifies the minority and combined majority class sample sets and constructs hyperspheres centered on each minority class sample, with radii determined by the distribution of the corresponding minority class. These hyperspheres guide the subsequent oversampling process. A trade-off parameter (\begin{document}$ \beta $\end{document}) is introduced to balance the influence of local densities between the combined majority and minority classes. Experimental results on KEEL datasets show that this approach effectively prevents class overlap during boundary oversampling while assigning higher oversampling weights to critical minority samples located near boundaries and in low-density regions. This improves boundary distribution and simultaneously addresses within-class imbalance. When the trade-off parameter is set to 0.5, MC-LDWO achieves a balanced consideration of both boundary distribution and the diverse densities present in minority classes due to data difficulty factors, thereby supporting improved performance in downstream classification tasks (Fig. 10).  Conclusions  Comparative results with other state-of-the-art oversampling algorithms demonstrate that: (1) The MC-LDWO algorithm effectively prevents overlap when strengthening decision boundaries by setting the cutoff threshold (\begin{document}$ T $\end{document}) and adaptively assigns oversampling weights according to two local density indicators for the minority and combined majority classes within the hypersphere. This approach addresses within-class imbalance caused by data difficulty factors and enhances boundary distribution. (2) By jointly considering density and boundary distribution, and setting the trade-off parameter to 0.5, the proposed algorithm can simultaneously mitigate within-class imbalance and reinforce the boundary information of minority classes. (3) When applied to highly imbalanced datasets characterized by complex decision boundaries and data difficulty factors such as outliers and small disjuncts, MC-LDWO significantly improves the boundary distribution of each minority class while effectively managing within-class imbalance, thereby enhancing the performance of subsequent classifiers.
A Survey on Physical Layer Security in Near-Field Communication
XU Yongjun, LI Jing, LUO Dongxin, WANG Ji, LI Xingwang, YANG Long, CHEN Li
Available online  , doi: 10.11999/JEIT250336
Abstract:
  Significance   Traditional wireless communication systems have relied on far-field plane-wave models to support wide-area coverage and long-distance transmission. However, emerging Sixth-Generation (6G) applications—such as extended reality, holographic communication, pervasive intelligence, and smart factories—demand ultra-high bandwidth, ultra-low latency, and sub-centimeter-level localization accuracy. These requirements exceed the spatial multiplexing gains and interference suppression achievable under far-field assumptions. Enabled by extremely large-scale antenna arrays and terahertz technologies, the near-field region has expanded to hundreds of meters, where spherical-wave propagation enables precise beam focusing and flexible spatial resource management. The additional degrees of freedom in the angle and distance domains, however, give rise to new Physical Layer Security (PLS) challenges, including joint angle–distance eavesdropping, beam-split-induced information leakage caused by frequency-dependent focusing, and security–interference conflicts in hybrid near- and far-field environments. This paper provides a comprehensive survey of near-field PLS techniques, advancing theoretical understanding of spherical-wave propagation and associated threat models while offering guidance for designing robust security countermeasures and informing the development of future 6G security standards.  Progress   This paper presents a comprehensive survey of recent advances in PLS for near-field communications in 6G networks, with an in-depth discussion of key enabling technologies and optimization methodologies. Core security techniques, including beam focusing, Artificial Noise (AN), and multi-technology integration, are first examined in terms of their security objectives. Beam focusing exploits ultra-large-scale antenna arrays and the spherical-wave propagation characteristics of near-field communication to achieve precise spatial confinement, thereby reducing information leakage. AN introduces deliberately crafted noise toward undesired directions to hinder eavesdropping. Multi-technology integration combines terahertz communications, Reconfigurable Intelligent Surfaces (RIS), and Integrated Sensing And Communication (ISAC), markedly enhancing overall security performance. Tailored strategies are then analyzed for different transmission environments, including Line-of-Sight (LoS), Non-Line-of-Sight (NLoS), and hybrid near–far-field conditions. In LoS scenarios, beamforming optimization strengthens interference suppression. In NLoS scenarios, RIS reconstructs transmission links, complicating unauthorized reception. For hybrid near–far-field environments, multi-beam symbol-level precoding spatially distinguishes users and optimizes beamforming patterns, ensuring robust security for mixed-distance user groups. Finally, critical challenges are highlighted, including complex channel modeling, tradeoffs between security and performance, and interference management in converged multi-network environments. Promising directions for future research are also identified, such as Artificial Intelligence (AI)-assisted security enhancement, cooperative multi-technology schemes, and energy-efficient secure communications in near-field systems.  Conclusions  This paper provides a comprehensive survey of PLS techniques for near-field communications, with particular emphasis on enabling technologies and diverse transmission scenarios. The fundamentals and system architecture of near-field communications are first reviewed, highlighting their distinctions from far-field systems and their unique channel characteristics. Representative PLS approaches are then examined, including beam focusing, AN injection, and multi-technology integration with RIS and ISAC. Secure transmission strategies are further discussed for LoS, NLoS, and hybrid near–far-field environments. Finally, several open challenges are identified, such as accurate modeling of complex channels, balancing security and performance, and managing interference in multi-network integration. Promising research directions are also outlined, including hybrid near–far-field design and AI-enabled security. These directions are expected to provide theoretical foundations for advancing and standardizing near-field communication security in future 6G networks.  Prospects   Research on PLS for near-field communications remains at an early stage, with no unified or systematic framework established to date. As communication scenarios become increasingly diverse and complex, future studies should prioritize hybrid far-field and near-field environments, where channel coupling and user heterogeneity raise new security challenges. AI-driven PLS techniques show strong potential for adaptive optimization and improved resilience against adversarial threats. In parallel, integrating near-field PLS with advanced technologies such as RIS and ISAC can deliver joint improvements in security, efficiency, and functionality. Moreover, low-power design will be essential to balance security performance with energy efficiency, enabling the development of high-performance, intelligent, and sustainable near-field secure communication systems.
A Ku-band Circularly Polarized Leaky-wave Antenna Loaded with Parasitic Slots
HUANG Zhiyuan, ZHANG Yunhua, ZHAO Xiaowen
Available online  , doi: 10.11999/JEIT250347
Abstract:
This paper proposes a Ku-band circularly polarized Leaky-Wave Antenna (LWA) based on a Substrate Integrated Waveguide (SIW). A parasitic slot, with the same configuration as the main radiation slot but reduced in size, is employed to address the open-stopband problem and enhance impedance matching. The radiation slot excites Circularly Polarized (CP) waves, while the parasitic slot simultaneously broadens the Axial Ratio (AR) bandwidth and suppresses the open-stopband effect. A prototype antenna is designed, fabricated, and measured. The results demonstrate that the antenna achieves a 32% 3-dB AR bandwidth from 12.6 GHz to 17.4 GHz, with CP beam scanning from –49° to +14°. The simulated and measured results are in good agreement. In addition, the realized gain remains stable across the operating band. Compared with existing works, the proposed design achieves the widest scanning range.  Objective  Compared with traditional phased array antennas, frequency-scanning antennas have extensive applications in both military and civilian fields owing to their advantages of low profile, low cost, and lightweight design. CP waves offer superior anti-interference performance compared with linearly polarized waves. As a representative frequency-scanning antenna, the LWA has attracted sustained global research interest. This study focuses on the investigation of a Ku-band Circularly Polarized Leaky-Wave Antenna (CP-LWA), with emphasis on wide-bandwidth and wide-scanning techniques, as well as methods for achieving circular polarization. The aim is to provide potential design concepts for next-generation mobile communication and radar system antennas.  Methods   The fan-shaped slot is modified based on previous work, and an additional size-reduced parasitic slot of the same shape as the main slot is introduced. The parasitic slots cancel the reflected waves generated by the main radiating slot, thereby suppressing the Open-Stop-Band (OSB) effect, and they also enlarge the effective radiating aperture, which improves radiation efficiency and impedance matching. By exploiting the metallic boundary of the conductors, the parasitic slots enhance CP performance and broaden the AR bandwidth. To validate the proposed design, an antenna consisting of 12 main slots and 11 parasitic slots is designed, simulated, and measured.  Results and Discussions  A prototype is designed, fabricated, and measured in a microwave anechoic chamber to validate the proposed antenna. Both simulated and measured S11 values remain below –10 dB across the entire Ku-band. The measured S11 is slightly higher in the low-frequency range (12~13 GHz) and slightly lower in the high-frequency range (16~18 GHz), while maintaining an overall consistent trend with the simulations, except for a frequency shift of approximately 0.2 GHz toward lower frequencies. For the AR bandwidth, the simulated and measured 3-dB AR bandwidths are 32.7% (12.8~17.8 GHz) and 32.0% (12.6~17.4 GHz), respectively. The realized gains are on average 0.6 dB lower than the simulated values across the AR bandwidth, likely due to measurement system errors and fabrication tolerances. The simulated and measured peak gains reach 14.26 dB and 13.65 dB, respectively, with maximum gain variations of 2.91 dB and 2.85 dB. The measured AR and gain results therefore show strong agreement with the simulations. The measured sidelobe level increases on average by approximately 0.65 dB. The simulated CP scanning range extends from –47° to +17°, while the measured range narrows slightly to –49° to +14°. The frequency shift of the LWA is analyzed, and based on the simulated effect of variations in εr on the scanning patterns, the shift toward lower frequencies is attributed to the actual dielectric constant of the substrate being smaller than the nominal value of 2.2 specified by the manufacturer.  Conclusions  This paper proposes a Ku-band CP-LWA based on a SIW. The antenna employs etched slots consisting of fan-shaped radiation slots and size-reduced parasitic slots. The radiation slots excite circular polarization due to their inherent geometric properties, while the parasitic slots suppress the CP effect and broaden the CP bandwidth. Measurements confirm that the proposed LWA achieves a wide 3-dB AR bandwidth of 12.6~17.4 GHz (32%) with a CP beam scanning range from –49° to +14°. Meanwhile, the antenna demonstrates stable gain performance across the entire AR bandwidth.
Research on ECG Pathological Signal Classification Empowered by Diffusion Generative Data
GE Beining, CHEN Nuo, JIN Peng, SU Xin, LU Xiaochun
Available online  , doi: 10.11999/JEIT250404
Abstract:
  Objective  ElectroCardioGram (ECG) signals are key indicators of human health. However, their complex composition and diverse features make visual recognition prone to errors. This study proposes a classification algorithm for ECG pathological signals based on data generation. A Diffusion Generative Network (DGN), also known as a diffusion model, progressively adds noise to real ECG signals until they approach a noise distribution, thereby facilitating model processing. To improve generation speed and reduce memory usage, a Knowledge Distillation-Diffusion Generative Network (KD-DGN) is proposed, which demonstrates superior memory efficiency and generation performance compared with the traditional DGN. This work compares the memory usage, generation efficiency, and classification accuracy of DGN and KD-DGN, and analyzes the characteristics of the generated data after lightweight processing. In addition, the classification effects of the original MIT-BIH dataset and an extended dataset (MIT-BIH-PLUS) are evaluated. Experimental results show that convolutional networks extract richer feature information from the extended dataset generated by DGN, leading to improved recognition performance of ECG pathological signals.  Methods  The generative network-based ECG signal generation algorithm is designed to enhance the performance of convolutional networks in ECG signal classification. The process begins with a Gaussian noise-based image perturbation algorithm, which obscures the original ECG data by introducing controlled randomness. This step simulates real-world variability, enabling the model to learn more robust representations. A diffusion generative algorithm is then applied to reconstruct and reproduce the data, generating synthetic ECG signals that preserve the essential characteristics of the original categories despite the added noise. This reconstruction ensures that the underlying features of ECG signals are retained, allowing the convolutional network to extract more informative features during classification. To improve efficiency, the approach incorporates knowledge distillation. A teacher-student framework is adopted in which a lightweight student model is trained from the original, more complex teacher ECG data generation model. This strategy reduces computational requirements and accelerates the data generation process, improving suitability for practical applications. Finally, two comparative experiments are designed to validate the effectiveness and accuracy of the proposed method. These experiments evaluate classification performance against existing approaches and provide quantitative evidence of its advantages in ECG signal processing.  Results and Discussions  The data generation algorithm yields ECG signals with a Signal-to-Noise Ratio (SNR) comparable to that of the original data, while presenting more discernible signal features. The student model constructed through knowledge distillation produces ECG samples with the same SNR as those generated by the teacher model, but with substantially reduced complexity. Specifically, the student model achieves a 50% reduction in size, 37.5% lower memory usage, and a 57% shorter runtime compared with the teacher model (Fig. 6). When the convolutional network is trained with data generated by the KD-DGN, its classification performance improves across all metrics compared with a convolutional network trained without KD-DGN. Precision reaches 95.7%, and the misidentification rate is reduced to approximately 3% (Fig. 9).  Conclusions  The DGN provides an effective data generation strategy for addressing the scarcity of ECG datasets. By supplying additional synthetic data, it enables convolutional networks to extract more diverse class-specific features, thereby improving recognition performance and reducing misidentification rates. Optimizing DGN with knowledge distillation further enhances efficiency, while maintaining SNR equivalence with the original DGN. This optimization reduces computational cost, conserves machine resources, and supports simultaneous task execution. Moreover, it enables the generation of new data without LOSS, allowing convolutional networks to learn from larger datasets at lower cost. Overall, the proposed approach markedly improves the classification performance of convolutional networks on ECG signals. Future work will focus on further algorithmic optimization for real-world applications.
Multi-target Behavior and Intent Prediction on the Ground Under Incomplete Perception Conditions
ZHU Xinyi, PING Peng, HOU Wanying, SHI Quan, WU Qi
Available online  , doi: 10.11999/JEIT250322
Abstract:
  Objective  Modern battlefield environments, characterized by complex and dynamically uncertain target behaviors combined with information asymmetry, present significant challenges for intent prediction. Conventional methods lack robustness in processing incomplete data, rely on oversimplified behavioral models, and fail to capture tactical intent semantics or adapt to rapidly evolving multi-target coordinated scenarios. These limitations restrict their ability to meet the demands of real-time recognition of high-value target intent and comprehensive ground target situational awareness. To address these challenges, this study proposes a Threat Field-integrated Gated Recurrent Unit model (TF-GRU), which improves prediction accuracy and robustness through threat field modeling, dynamic data repair, and multi-target collaboration, thereby providing reliable support for battlefield decision-making.  Methods  The TF-GRU framework integrates static and dynamic threat field modeling with a hybrid Particle Filtering (PF) and Dynamic Time Warping (DTW) strategy. Static threat fields quantify target-specific threats (e.g., tanks, armored vehicles, artillery) using five factors: enemy-friend distance, range, firepower, defense, and mobility. Gaussian and exponential decay models are employed to describe spatial threat diffusion across different target categories. Dynamic threat fields incorporate real-time kinematic variables (velocity, acceleration, orientation) and temporal decay, allowing adaptive updates of threat intensity. To address incomplete sensor data, a PF-DTW switching mechanism dynamically alternates between short-term PF (N = 1 000 particle) and long-term historical trajectory matching (DTW with β = 50). Collaborative PF introduces neighborhood angular constraints to refine multi-target state estimation. The GRU architecture is further enhanced with Mish activation, adaptive Xavier initialization, and threat-adaptive gating, ensuring effective fusion of trajectory and threat features.  Results and Discussions  Experiments were conducted on a simulated dataset comprising 40 trajectories and 144,000 timesteps. Under complete data conditions, the TF-GRU model achieved the highest accuracy on both the training and test sets, reaching 94.7% and 92.9%, respectively, indicating strong fitting capability and generalization performance (Fig. 10). After integrating static and dynamic threat fields, model accuracy increased from 72% (trajectory-only input) to 83%, accompanied by substantial improvements in F1 scores and reductions in predictive uncertainty (Fig. 6). In scenarios with 30% missing data, TF-GRU maintained an accuracy of 86.2%, outperforming comparative models and demonstrating superior robustness (Fig. 10). These results confirm that the PF-DTW mechanism effectively reduces the adverse effects of both short-term and long-term data loss, while the collaborative PF strategy strengthens multi-target prediction through neighborhood synergy (η = 0.6). This combination enables robust threat field reconstruction and reliable intent inference (Figs. 89).  Conclusions  The TF-GRU model effectively addresses the challenges of intent prediction in complex battlefield environments with incomplete data through threat field modeling, the PF-DTW dynamic repair mechanism, and multi-target collaboration. It achieves high accuracy and robustness, providing reliable support for situational awareness and command decision-making. Future work will focus on applying the model to real-world datasets and enhancing computational efficiency to facilitate practical deployment.
Low-complexity Ordered Statistic Decoding Algorithm Based on Skipping Mechanisms
WANG Qianfan, GUO Yangeng, SONG Linqi, MA Xiao
Available online  , doi: 10.11999/JEIT250447
Abstract:
  Objective  Ultra-Reliable Low-Latency Communication (URLLC) in 5G and the emerging Hyper-Reliable Low-Latency Communication (HRLLC) in 6G impose exceptionally stringent requirements on both reliability and end-to-end delay. These requirements create opportunities and challenges for short-length channel codes, particularly in scenarios where Maximum-Likelihood (ML) or near-ML decoding is desirable but computational complexity and latency are prohibitive. Ordered Statistic Decoding (OSD) is a universal near-ML decoding technique that can closely approach finite-length performance bounds. However, its re-encoding step suffers from combinatorial explosion, resulting in impractical complexity in high-throughput and low-latency systems. The excessive number of Test-Error-Pattern (TEP) re-encodings fundamentally restricts the deployment of OSD in URLLC and HRLLC contexts. To address this bottleneck, we design multiple efficient skip mechanisms that substantially reduce re-encoding operations while maintaining negligible performance degradation.  Methods  Three complementary skipping mechanisms are developed to prune the OSD re-encoding search: (1) Soft-information based skipping. Two criteria are introduced—Trivial and Dynamic Approximate Ideal (DAI), to compare the soft metric of each TEP against the minimum soft weight in the current list. Candidates with excessively large soft weights, which are unlikely to be correct, are skipped. Unlike prior work that evaluates only the first TEP at each Hamming weight increment, both criteria are applied to every candidate. The Trivial criterion ensures no performance loss by skipping only when a TEP’s soft metric exceeds the best-so-far. The DAI criterion incorporates an expected residual soft-weight compensation term over non-basis bits, enabling more aggressive skipping with minimal performance degradation. (2) Extra-parity skipping. The search dimension is expanded from \begin{document}$ k $\end{document} to \begin{document}$ k+\delta $\end{document} by appending the \begin{document}$ \delta $\end{document} most reliable non-basis bit positions to the test vector. Additional parity checks arising from the extended generator matrix eliminate invalid TEPs. Any candidate failing these extra parity constraints is bypassed. (3) Joint skipping. This approach integrates the two preceding mechanisms. Each partial TEP \begin{document}$ ({\boldsymbol{e}}_{\mathrm{L}},{\boldsymbol{e}}_{\delta })\in {\mathbb{F}}_{2}^{k+\delta } $\end{document} is first tested using the DAI rule and then subjected to the extra-parity check. Only candidates passing both criteria are re-encoded.  Results and Discussions  Extensive simulations on extended BCH \begin{document}$ \left[\mathrm{128,64}\right] $\end{document} and BCH \begin{document}$ \left[\mathrm{127,64}\right] $\end{document} codes over the BPSK-AWGN channel demonstrate the efficacy of the proposed skipping mechanisms. Soft-information skipping: When compared with conventional OSD using maximum flipping order \begin{document}$ t=4 $\end{document}, the Trivial rule is found to reduce average re-encodings by 50%~90% across the SNR range. The DAI rule achieves an additional 60%~99% reduction beyond the Trivial rule. At SNR = 3 dB, the average number of re-encodings decreases from approximately \begin{document}$ 6.7\times {10}^{5} $\end{document} to \begin{document}$ 1.2\times {10}^{3} $\end{document}, with negligible degradation in Frame-Error Rate (FER) (Fig. 1). Extra-parity skipping: For \begin{document}$ \delta =4 $\end{document}, over \begin{document}$ 90 \% $\end{document} of re-encodings are eliminated uniformly across SNR values, thereby reducing dependence on channel conditions. This reduction is achieved without significant FER loss (Fig. 2). Joint skipping: The combined mechanism demonstrates superior performance over individual schemes. It reduces average re-encodings by an additional ~40% compared with the DAI rule alone, and by more than \begin{document}$ > 99.9 \% $\end{document} compared with extra-parity alone in high-SNR regimes. In this region, re-encodings decrease from \begin{document}$ \sim 6.7\times {10}^{5} $\end{document} to fewer than 100, while FER remains nearly identical to that of baseline OSD (Fig. 3). The joint skipping mechanism is further evaluated on BCH codes with different rates, including \begin{document}$ \left[\mathrm{127,36}\right] $\end{document}, \begin{document}$ \left[\mathrm{127,64}\right] $\end{document}, and \begin{document}$ \left[\mathrm{127,92}\right] $\end{document}. In all cases, substantial reductions in re-encodings are consistently achieved with negligible performance degradation (Fig. 4). A comparative analysis with state-of-the-art schemes—including Probabilistic Sufficient/Necessary Conditions (PSC/PNC), Fast OSD (FOSD), and Order-Skipping OSD (OS-OSD)—shows that the proposed joint skipping OSD with \begin{document}$ \delta =4 $\end{document} achieves the lowest re-encoding count. Up to two orders of magnitude fewer re-encodings are observed relative to OS-OSD at low SNR, and superiority over FOSD is maintained at moderate SNR, while error-correction performance is preserved across all tested SNRs (Fig. 5).  Conclusions  To address the stringent reliability and latency requirements of 5G URLLC and future 6G HRLLC, this work presents novel skipping mechanisms for OSD that substantially reduce re-encoding complexity. For offline pre-computed TEPs, the soft-information, extra-parity, and joint skipping rules eliminate more than \begin{document}$ 99 \% $\end{document} of redundant re-encodings in typical operating regimes with negligible degradation in Frame-Error Rate (FER). In particular, the proposed joint skipping mechanism lowers the average re-encoding count from approximately \begin{document}$ 6.7\times {10}^{5} $\end{document} to only tens in the high-SNR region, thereby meeting practical latency constraints while preserving near-ML performance. These findings demonstrate the potential of the proposed skipping framework to enable high-performance short-block decoding in next-generation HRLLC.
Combine the Pre-trained Model with Bidirectional Gated Recurrent Units and Graph Convolutional Network for Adversarial Word Sense Disambiguation
ZHANG Chunxiang, SUN Ying, GAO Kexin, GAO Xueyao
Available online  , doi: 10.11999/JEIT250386
Abstract:
  Objective  In Word Sense Disambiguation (WSD), the Linguistically-motivated bidirectional Encoder Representation from Transformer (LERT) is employed to capture rich semantic representations from large-scale corpora, enabling improved contextual understanding of word meanings. However, several challenges remain. Current WSD models are not sufficiently sensitive to temporal and spatial dependencies within sequences, and single-dimensional features are inadequate for representing the diversity of linguistic expressions. To address these limitations, a hybrid network is constructed by integrating LERT, Bidirectional Gated Recurrent Units (Bi-GRU), and Graph Convolutional Network (GCN). This network enhances the modeling of structured text and contextual semantics. Nevertheless, generalization and robustness remain problematic. Therefore, an adversarial training algorithm is applied to improve the overall performance and resilience of the WSD model.  Methods  An adversarial WSD method is proposed based on a pre-trained model, combining Bi-GRU and GCN. First, word forms, parts of speech, and semantic categories of the neighboring words of an ambiguous term are input into the LERT model to obtain the CLS sequence and token sequence. Second, cross-attention is applied to fuse the global semantic information extracted by Bi-GRU from the token sequence with the local semantic information derived from the CLS sequence. Sentences, word forms, parts of speech, and semantic categories are then used as nodes to construct a disambiguation feature graph, which is subsequently input into GCN to update the feature information of the nodes. Third, the semantic category of the ambiguous word is determined through the interpolated prediction layer and semantic classification layer. Fourth, subtle continuous perturbations are generated by computing the gradient of the dynamic word vectors in the input. These perturbations are added to the original word vector matrix to create adversarial samples, which are used to optimize the LERT+Bi-GRU+CA+GCN (LBGCA-GCN) model. A cross-entropy loss function is applied to measure the performance of the LBGCA-GCN model on adversarial samples. Finally, the loss from the network is combined with the loss from AT to optimize the LBGCA-GCN model..  Results and Discussions  When the FreeLB algorithm is applied, stronger adversarial perturbations are generated, and the LBGCA-GCN-AT model achieves the best performance (Table 2). As the number of perturbation steps increases, the strength of AT improves. However, when the number of steps exceeds a certain threshold, the LBGCA-GCN+AT(LBGCA-GCN-AT) model begins to overfit. The Free Large-Batch (FreeLB) algorithm demonstrates strong robustness with three perturbation steps (Table 3). The cross-attention mechanism, which fuses the token sequence with the CLS sequence, yields significant performance gains in complex semantic scenarios (Fig. 3). By incorporating AT, the LBGCA-GCN-AT model achieves notable improvements across multiple evaluation metrics (Table 4).  Conclusions  This study presents an adversarial WSD method based on a pre-trained model, integrating Bi-GRU and GCN to address the weak generalization ability and robustness of conventional WSD models. LERT is used to transform discriminative features into dynamic word vectors, while cross-attention fuses the global semantic information extracted by Bi-GRU from the token sequence with the local semantic information derived from the CLS sequence. This fusion generates more complete node representations for the disambiguation feature graph. A GCN is then applied to update the relationships among nodes within the feature graph. The interpolated prediction layer and semantic classification layer are used to determine the semantic category of ambiguous words. To further improve robustness, the gradient of the dynamic word vector is computed and perturbed to generate adversarial samples, which are used to optimize the LBGCA-GCN model. The network loss is combined with the AT loss to refine the model. Experiments conducted on the SemEval-2007 Task #05 and HealthWSD datasets examine multiple factors affecting model performance, including adversarial algorithms, perturbation steps, and sequence fusion methods. Results demonstrate that introducing AT improves the model’s ability to handle real-world noise and perturbations. The proposed method not only enhances robustness and generalization but also strengthens the capacity of WSD models to capture subtle semantic distinctions.
Constructing Two Classes of Maximum Distance Separable Entanglement-Assisted Quantum Error-Correcting Codes by Using Twisted Generalized Reed-Solomon Codes
PAN Xin, GAO Jian
Available online  , doi: 10.11999/JEIT250258
Abstract:
  Objective  With the rapid development of quantum communication and computation, efficient error-correction technologies are critical for ensuring the reliability of quantum systems. Maximal Distance Separable (MDS) Entanglement-Assisted Quantum Error-Correcting Codes (EAQECCs) with flexible code lengths and larger minimum distances offer significant advantages in enhancing quantum system robustness. However, classical quantum codes face limitations in parameter flexibility and minimum distance constraints. This study addresses these challenges by leveraging Twisted Generalized Reed-Solomon (TGRS) codes to construct novel MDS EAQECCs, aiming to improve performance in complex quantum communication scenarios. In this paper, TGRS codes are innovatively applied to construct MDS EAQECCs. Different from polynomial construction methods, we determine the dimension of the Hermitian Hull through special matrix rank analysis, and successfully construct \begin{document}$ q $\end{document}-ary MDS EAQECCs with a minimum distance exceeding\begin{document}$ q+1 $\end{document}. Two construction schemes with flexible code lengths are proposed. Notably, the minimum distances of the constructed \begin{document}$ q $\end{document}-ary codes all exceed \begin{document}$ q+1 $\end{document}. Table 2 systematically summarizes the known MDS EAQECCs with a length less than \begin{document}$ {q}^{2} $\end{document} and a minimum distance exceeding \begin{document}$ q+1 $\end{document}. The novelty of the construction schemes in this paper is highlighted through parameter comparison. Reasoning shows that our schemes achieve flexible adjustment of the code length while maintaining the advantage of the minimum distance.  Methods  The proposed approach integrates TGRS codes with algebraic techniques to design MDS EAQECCs. Two families of MDS EAQECCs are constructed by using TGRS codes over finite fields. The key steps are as follows: (1) TGRS Code Construction: Utilizing twisted polynomials and hook parameters to generate TGRS codes with adjustable dimensions and minimum distances. (2) Hermitian Hull Analysis: Applying matrix rank analysis to determine the dimensions of the Hermitian Hull of the constructed codes, which is crucial for satisfying the Singleton bound in quantum codes. (3) Twisted Operation Optimization: Employing twisted operations to transform the constructed MDS EAQECCs into ME-MDS EAQECCs.  Results and Discussions  This paper constructs two families of MDS EAQECCs using TGRS codes and gives certain twisted conditions under which the codes are ME-MDS EAQECCs. Compared with other known codes, these new codes have more flexible parameters and significant advantages in terms of code length, minimum distance, and maximum entanglement state. This paper constructs \begin{document}$ q $\end{document}-ary EAQECCs with \begin{document}$ [[i(q-1), $\end{document}\begin{document}$ i(q-1)-2j-q-t+2v+2,q+j+1;q-t+2v+2]{]}_{q} $\end{document} when \begin{document}$ i $\end{document} is odd and \begin{document}$ [[i(q-1),i(q-1)- 2j-q-t+ $\end{document}\begin{document}$ 2v+3,q+j+1;q-t+2v+3]{]}_{q} $\end{document} when \begin{document}$ i $\end{document} is even. Based on Theorem 1 and Theorem 2, several types of MDS EAQECCs are obtained, and their parameters are listed in Table 1.Theorems 3-6 give the existence conditions of specific \begin{document}$ q $\end{document}-ary EAQECCs under different parameter settings. Furthermore, this paper upgrades the MDS EAQECCs to ME-MDS EAQECCs with \begin{document}$ \left[\right[i(q-1),i(q-1)-q-j,q+j+1;q+j]{]}_{q} $\end{document} by a twisted operation. Meanwhile, this paper constructs \begin{document}$ q $\end{document}-ary EAQECCs with \begin{document}$ \left[\right[i(q+1),(i-1)(q+1)-s,q+2;q-s+1]{]}_{q} $\end{document}. Moreover, this paper upgrades the MDS EAQECCs to ME-MDS EAQECCs with \begin{document}$ [[i(q+1),(i-1) $\end{document}\begin{document}$ (q+1),q+2;q+1]{]}_{q} $\end{document} by a twisted operation. Theorem 7 gives the dimension of the Hermitian Hull of this RS code after the twisted operation on its generating matrix. Similar to the first construction, a new twisted operation is applied, which upgrades the MDS EAQECCs to ME-MDS EAQECCs with specific parameters. Theorems 8, 9 give the existence conditions of specific \begin{document}$ q $\end{document}-ary EAQECCs under different parameter settings based on this construction method.  Conclusions  Two families of MDS EAQECCs are constructed using TGRS codes, and the parameters of known MDS EAQECCs are systematically summarized. Comparative analysis reveals that the EAQECCs proposed in this paper offer the following advantages: Compared with the known \begin{document}$ q $\end{document}-ary MDS EAQECCs (Table 2), the parameters of the MDS EAQECCs constructed here are new and have not been covered by previous studies; The codes not only enable flexible code-length adjustments but also achieve a minimum distance that significantly exceeds traditional theoretical bounds; Under specific twisted conditions, the constructed MDS EAQECCs are upgraded to ME-MDS EAQECCs. By introducing the twisted operation on RS codes, more flexible parameter combinations are obtained. This provides greater flexibility in the design of quantum error-correcting codes, enabling better adaptation to different quantum communication requirements. This improvement further optimizes the performance of quantum error-correcting codes, enhancing the entanglement-assisted error-correcting ability and improving the overall efficiency of quantum systems. These research results indicate that TGRS codes are important theoretical tools for constructing high-performance EAQECCs with excellent parameters. They play a pivotal part in advancing the development of quantum communication technology. Moreover, they offer a firm theoretical underpinning for the practical implementation of quantum error-correcting codes. Future research can focus on further exploring the potential of TGRS codes in constructing more advanced quantum error-correcting codes and expanding their applications in different quantum communication scenarios.
Global–local Co-embedding and Semantic Mask-driven Aging Approach
LIU Yaohui, LIU Jiaxin, SUN Peng, SHEN Zhe, LANG Yubo
Available online  , doi: 10.11999/JEIT250430
Abstract:
  Objective  Facial age progression has become increasingly important in applications such as criminal investigation and digital identity authentication, making it a key research area in computer vision. However, existing mainstream facial age progression networks face two primary limitations. First, they tend to overemphasize the embedding of age-related features, often at the expense of preserving identity-consistent multi-scale attributes. Second, they fail to effectively eliminate interference from non-age-related elements such as hair and glasses, leading to suboptimal performance in complex scenarios. To address these challenges, this study proposes a global–local co-embedding and semantic mask-driven aging method. The global–local co-embedding strategy improves the accuracy of input portrait reconstruction while reducing computational cost during the embedding phase. In parallel, a semantic mask editing mechanism is introduced to remove non-age-related features—such as hair and eyewear—thereby enabling more accurate embedding of age-related characteristics. This dual strategy markedly enhances the model’s capacity to learn and represent age-specific attributes in facial imagery.  Methods  A Global–Local Collaborative Embedding (GLCE) strategy is proposed to achieve high-quality latent space mapping of facial images. Distinct learning objectives are assigned to separate latent subspaces, which enhances the representation of fine-grained facial features while preserving identity-specific information. Therefore, identity consistency is improved, and both training time and computational cost are reduced, increasing the efficiency of feature extraction. To address interference from non-age-related elements, a semantic mask-driven editing mechanism is employed. Semantic segmentation and image inpainting techniques are integrated to accurately remove regions such as hair and glasses that hinder precise age modeling. A differentiable generator, DsGAN, is introduced to align the transferred latent codes with the embedded identity-preserving codes. Through this alignment, the expression of age-related features is enhanced, and identity information is better retained during the age progression process.  Results and Discussions  Experimental results on benchmark datasets, including CCAD and CelebA, demonstrate that GLS-Age outperforms existing methods such as IPCGAN, CUSP, SAM, and LATS in identity confidence assessment. The age distributions of the generated portraits are also more closely aligned with those of the target age groups. Qualitative analysis further shows that, in cases with hair occlusion, GLS-Age produces more realistic wrinkle textures and enables more accurate embedding of age-related features compared with other methods. Simultaneously, it significantly improves the identity consistency of the synthesized facial images.  Conclusions  This study addresses core challenges in facial age progression, including identity preservation, inadequate detail modeling, and interference from non-age-related factors. A novel Global–Local collaborative embedding and Semantic mask-driven Aging method (GLS-Age) is proposed to resolve these limitations. By employing a differentiated latent space learning strategy, the model achieves hierarchical decoupling of structural and textural features. When integrated with semantic-guided portrait editing and a differentiable generator for latent space alignment, GLS-Age markedly enhances both the fidelity of age feature expression and the consistency of identity retention. The method demonstrates superior generalization and synthesis quality across multiple benchmark datasets, effectively reproducing natural wrinkle patterns and age-related facial changes. These results confirm the feasibility and advancement of GLS-Age in facial age synthesis tasks. Furthermore, this study establishes a compact, high-quality dataset focused on Asian facial portraits, supporting further research in image editing and face generation within this demographic. The proposed method not only contributes technical support to practical applications such as cold case resolution and missing person identification in public security but also offers a robust data and modeling framework for advancing human age-based simulation technologies. Future work will focus on enhancing controllable editing within latent spaces, improving anatomical plausibility in skull structure transformations, and strengthening model performance across extreme age groups, including infants and the elderly. These efforts aim to expand the application of facial age progression in areas such as forensic analysis, humanitarian family search, and social security systems.
Research on Fusion Localization Algorithm of Indoor UWB and IMU Assisted by GPR Error Calibration
MA Xinpeng, CHEN Yu, CUI Zhicheng, LI Xingguang, CUI Wei
Available online  , doi: 10.11999/JEIT241145
Abstract:
  Objective  Ultra-WideBand (UWB) ranging in confined indoor environments is prone to coarse ranging errors and Gaussian noise, which substantially degrade localization accuracy for both static and dynamic targets, even under Line-of-Sight (LOS) conditions. In addition, during indoor operations of wheeled vehicles, obstacles often obstruct the LOS between onboard UWB tags and anchors, resulting in Non-Line-of-Sight (NLOS) propagation. NLOS-induced interference results in abrupt fluctuations in range measurements, which severely compromise the estimation of the vehicle’s motion state.  Methods   To address these challenges, this study proposes a Gaussian Process Regression (GPR)-calibrated indoor UWB/Inertial Measurement Unit (IMU) fusion localization algorithm (GIU-EKF). The proposed approach offers two key features. First, a UWB ranging error model for the entire two-dimensional localization area is established by collecting UWB measurements at a limited set of reference points under LOS conditions. This model is applied to correct LOS-related ranging biases. Second, by leveraging the low short-term drift characteristics of inertial measurements, the algorithm fuses IMU and UWB data to mitigate NLOS ranging errors.  Results and Discussions  Under LOS conditions, a GPR-based error calibration model is constructed by sampling and analyzing UWB ranging deviations at known reference coordinates. This model captures the statistical association between two-dimensional spatial positions and the corresponding ranging errors. For any queried location, the model generates a set of probabilistic range estimates, with the final range value obtained by weighting nearby sample points according to their normalized likelihoods. This enables real-time suppression of LOS-related ranging errors. A threshold-based detection mechanism identifies NLOS conditions when the UWB range increment exceeds a predefined threshold. In NLOS scenarios, a subordinate Extended Kalman Filter (EKF) fuses UWB range data with short-term IMU measurements to compensate for NLOS-induced ranging errors during motion. The corrected range data are then incorporated into a primary EKF to update the vehicle’s motion state estimation. Experimental results demonstrate that the proposed GPR-based coarse error correction reduces localization errors by 64% and 58% for static and dynamic tags under LOS conditions, respectively. In three representative NLOS scenarios, the GIU-EKF algorithm maintains reliable motion state estimation for low-speed targets, achieving an average localization error of 7.5 cm. For tags moving at speeds between 0.2 m/s and 0.8 m/s, localization errors remain below 10 cm. The robustness of the proposed algorithm under extreme conditions is also validated. As shown in Section 4.3.2, the algorithm maintains stable velocity vector estimation even when the wheeled vehicle experiences alternating occlusions between single-anchor and multi-anchor configurations. Under low-speed conditions (2.2 cm/s), the localization error remains as low as 6.7 cm. Section 4.3.3 further verifies the algorithm’s convergence capability under large initial deviations. When subjected to initial heading errors between 5° and 50°, or a combined 1.5-meter position offset and 10° heading deviation, the proposed method consistently converges to the true position within a 2-meter travel distance.  Conclusions  This study presents a GPR-assisted indoor UWB/IMU fusion localization algorithm. By independently suppressing LOS and NLOS ranging errors from four UWB anchors, the proposed approach enhances localization accuracy in complex indoor environments. Under LOS conditions, a GPR-based correction mitigates coarse UWB ranging errors. In NLOS scenarios, short-term inertial estimates compensate for anomalous UWB measurements. A subordinate EKF adaptively balances observation uncertainties from the two sensing modalities, maintaining motion state observability when the tag moves slowly in confined spaces. This design avoids long-term drift accumulation, which is often observed in tightly coupled systems that treat IMU data as a strong prior, particularly when using low-cost inertial sensors. Experimental results demonstrate that the proposed algorithm achieves sub-10 cm localization accuracy under both LOS and NLOS conditions. During low-speed operations, the system maintains convergence of both velocity and position estimates. Furthermore, even with significant initial motion state biases, the algorithm consistently converges to the true trajectory. These findings indicate that the proposed method effectively meets the operational requirements of mobile robots in narrow indoor environments. However, practical application still requires further attention to two key aspects: efficient and autonomous collection of coordinate samples for GPR model training, and integration of real-time localization outputs with vehicle path planning and motion control systems.
Multi-Objective Optimization of UAV-Assisted Wireless Power Transfer Mobile Edge Computing System
LIU Jianhua, LI Guohua, LIU Jiajia, TU Xiaoguang
Available online  , doi: 10.11999/JEIT250026
Abstract:
  Objective   With the rapid growth of data volume at the edge of the Internet of Things (IoT), traditional centralized computing architectures are inadequate to meet real-time processing requirements. Due to the limited computing and storage capabilities of IoT devices, data overflow frequently occurs when handling large volumes of information. Meanwhile, the reliance on battery power exacerbates energy shortages, further hindering continuous operation and data processing. Unmanned Aerial Vehicles (UAVs), with their flexible deployment capabilities, are increasingly being integrated into distributed computing environments to improve IoT data processing efficiency. However, the stochastic and dynamic nature of IoT user demands presents significant challenges for existing resource scheduling schemes, which struggle to effectively coordinate UAV deployments. Moreover, most studies focus on single-objective optimization, making it difficult to simultaneously balance energy harvesting, energy consumption, system latency, and data transmission rates under complex environmental conditions. To address these challenges, this paper proposes an optimization framework for a UAV-assisted, wireless-powered Mobile Edge Computing (MEC) system based on an improved Multi-Objective Deep Deterministic Policy Gradient (MODDPG) algorithm. The proposed method jointly optimizes task offloading, energy harvesting, flying energy consumption, and system latency, enhancing the overall performance and feasibility of IoT systems.  Methods   This paper designs a UAV-assisted, wireless-powered MEC system model, where the UAV follows a “fly-hover-communicate” protocol and operates in full-duplex mode during hovering to simultaneously perform data collection and energy transmission. The system model integrates a propulsion energy consumption model and a nonlinear energy harvesting model, formulating a multi-objective optimization problem based on task generation, energy state, and wireless channel conditions of IoT devices. To address this optimization problem, an improved MODDPG algorithm is proposed. A multi-dimensional reward function is constructed to maximize the data transmission rate and total harvested energy while minimizing system energy consumption and task offloading latency within a unified reinforcement learning framework. The agent continuously interacts with the dynamic environment to optimize the UAV’s flight trajectory, task offloading ratio, and energy transmission strategy. The proposed method is implemented and validated through extensive simulations conducted on the Python platform under various experimental settings.  Results and Discussions  To evaluate the performance of the proposed algorithm, comparative experiments are conducted against four other control strategies. The impact of different data collection coverage radii on the proposed strategy is analyzed (Fig. 3, Fig. 4). In terms of total data transmission rate, the \begin{document}${P_{\rm{MODDPG}}}$\end{document} strategy consistently ranks second to the maximum throughput strategy across all coverage settings, while significantly outperforming the other three strategies (Fig. 3(a)). Regarding total harvested energy, \begin{document}${P_{\rm{MODDPG}}}$\end{document} achieves the highest energy collection in all scenarios, with performance improving as the coverage radius increases (Fig. 3(b)), demonstrating its capability to enhance energy transmission coverage and system efficiency. In terms of system energy consumption, the \begin{document}${P_{{V_{\max }}}}$\end{document} strategy exhibits the highest energy usage, whereas \begin{document}${P_{\rm{MODDPG}}}$\end{document} effectively controls flight energy consumption, maintaining significantly lower energy usage compared to high-speed flying strategies (Fig. 3(c)). For average latency, \begin{document}${P_{\rm{MODDPG}}}$\end{document} achieves lower latency than traditional flight strategies and the TD3-based method in most cases, though it remains slightly higher than the maximum throughput strategy (Fig. 3(d)). In terms of service performance, \begin{document}${P_{\rm{MODDPG}}}$\end{document} demonstrates superior results across multiple metrics, including the number of uploading users, average harvested energy, average number of charged users per hovering instance, and the number of users experiencing data overflow, with particularly significant advantages in energy collection and reducing data overflow rates. (Fig. 4(a), Fig. 4(b), Fig. 4(c), Fig. 4(d)). In experiments with varying IoT device densities (Fig. 5, Fig. 6), \begin{document}${P_{\rm{MODDPG}}}$\end{document} exhibits strong environmental adaptability. As device density increases and resource competition intensifies, \begin{document}${P_{\rm{MODDPG}}}$\end{document} intelligently optimizes flight paths and resource scheduling, maintaining high data transmission rates and energy harvesting levels while effectively suppressing system latency and data overflow. Furthermore, simulation results under different task load scenarios (small and large tasks) (Fig. 7) indicate that \begin{document}${P_{\rm{MODDPG}}}$\end{document} can flexibly adjust its strategy based on task load variations, maximizing resource utilization. Even under heavy load conditions, the proposed algorithm maintains superior data rates, energy harvesting efficiency, and system stability, effectively coping with dynamic traffic variations. Notably, \begin{document}${P_{\rm{MODDPG}}}$\end{document} not only delivers outstanding optimization performance but also demonstrates significantly faster convergence during training compared to traditional methods such as Proximal Policy Optimization (PPO) and TD3. Overall,\begin{document}${P_{\rm{MODDPG}}}$\end{document} outperforms existing strategies in multi-objective optimization, system energy efficiency enhancement, task scheduling flexibility, and environmental adaptability.  Conclusions   This paper addresses the critical challenges of energy limitations, data overflow, and high latency in IoT environments by proposing an optimization framework for a UAV-assisted wireless-powered MEC system, which integrates UAV, MEC, and Wireless Power Transfer (WPT) technologies. In the proposed system, IoT devices’ data upload demands are updated in real time, with the UAV sequentially accessing each device based on the priority of its demands using a "fly-hover-communicate" protocol. An improved MODDPG algorithm, based on deep reinforcement learning, is introduced to jointly optimize data transmission rate, energy harvesting, energy consumption, and system latency. Training results demonstrate that, under all optimization scenarios, the proposed scheme consistently achieves the highest total harvested energy while simultaneously balancing the other three optimization objectives. Compared to other baseline strategies, the proposed method exhibits better adaptability by dynamically adjusting to environmental changes and varying device priorities, ultimately achieving coordinated multi-objective optimization and superior performance across different conditions. Future research will focus on extending this work to address multi-UAV cooperative task allocation, resource management, energy harvesting coordination, and multi-UAV trajectory planning.
Task Segmentation and Computing Resource Allocation Method Driven by Path Prediction in Internet of Vehicles
HUO Ru, LÜ Kecheng, HUANG Tao
Available online  , doi: 10.11999/JEIT250135
Abstract:
  Objective  The rapid development of 5G and the Internet of Vehicles (IoV) has accelerated the demand for intelligent services and large-scale data processing. However, limited computational capacity at vehicle terminals and frequent topology changes due to high mobility reduce transmission efficiency and reliability. While current task offloading and resource allocation methods meet basic processing needs, they fall short in supporting fine-grained task segmentation and in adapting to the dynamic characteristics of vehicular networks. To address these limitations, this study proposes an optimization approach for task segmentation and computing resource allocation based on Reinforcement Learning (RL). By combining the time-series modeling capabilities of advanced Deep Learning (DL) models with the resource scheduling flexibility of a Multi-Agent Deep Reinforcement Learning (MADRL) algorithm, this method enables intelligent, prediction-driven task segmentation and allocation. It supports decoupled processing and multi-objective optimization of latency and resource use. The proposed framework enhances offloading efficiency and resource utilization in Vehicular Edge Computing (VEC) environments and provides theoretical support for real-time resource management in highly dynamic vehicular networks.  Methods  The proposed method integrates vehicle trajectory prediction, dynamic task segmentation, and adaptive resource scheduling. For trajectory prediction, historical movement data are preprocessed through trajectory mapping. A Transformer-based model captures long-range temporal dependencies to achieve accurate path prediction. Based on predicted routes and estimated dwell times in different edge zones, a segmentation strategy divides tasks into subtasks to distribute load evenly across edge servers. The resource allocation challenge under this segmentation strategy is modeled as a Markov Decision Process (MDP), with each edge server acting as an agent that interacts with its environment. The state space includes task sizes, server loads, and available resources, while the action space defines the scheduling decisions each agent may select. The reward function is designed to minimize delay and improve efficiency, promoting cooperative behavior across agents. Through continual environmental feedback, the MADRL framework adaptively adjusts resource allocation in response to real-time task demands and server workloads.  Results and Discussions  The Transformer-based path prediction model outperforms baseline models (Table 1), achieving a Mean Squared Error (MSE) of 0.0116, substantially lower than that of LSTM (0.0246) and BiLSTM (0.0323). To identify the optimal prediction horizon, the model is evaluated across different prediction steps (Fig. 5). The reward increases from 2 to 5 steps but declines beyond 5 due to increased transmission delays caused by frequent edge node handovers. Thus, 5-step prediction is adopted as the optimal setting. The proposed algorithm is assessed against three baseline strategies: binary offloading, random segmentation, and alternating convex search. Under a 5-step horizon and a five-task vehicle scenario (Fig. 6), the proposed method reduces average task latency by 54.1% compared with random segmentation and by 16.5% compared with alternating convex search. By tuning the weight coefficient (Fig. 7), the method demonstrates balanced performance across latency and resource efficiency objectives. Specifically, at a coefficient of 0.8, it achieves a convergence reward of 39.224, exceeding that of alternating convex search (35.872) by 9.4%. Scalability is evaluated under varying vehicle numbers (Fig. 8). The proposed method consistently yields the highest convergence reward across all cases. With four vehicles, it reaches a reward of 41.059, 7.3% higher than alternating convex search (38.277). With seven vehicles, it maintains a reward of 35.630, exceeding the baseline’s 32.524 by 9.6%. Overall, the proposed method exhibits superior latency reduction, higher resource efficiency, and better adaptability across different vehicle densities, validating its effectiveness and scalability for task offloading and resource allocation in VEC systems.  Conclusions  This study proposes a task segmentation and computing resource allocation method tailored for VEC scenarios. By exploiting the time-series modeling capacity of Transformer networks, the accuracy of vehicle trajectory prediction is enhanced. A dynamic task segmentation strategy is then applied to optimize workload distribution across Mobile Edge Computing (MEC) servers. The use of a MADRL-based cooperative resource allocation approach effectively reduces task latency while improving resource use. Simulation results confirm the method’s effectiveness in minimizing processing delay and enhancing system efficiency. Future research will explore the role of task prioritization in resource allocation and investigate strategies to sustain performance under increased vehicle counts and task complexity in large-scale deployments.
Acoustic DOA Estimation in Underwater Environments by Integrating Spatial Domain Wiener Filtering and Convolutional Neural Networks
XING Chuanxi, HUANG Tinglong, TAN Guangzhi, LI Weiqiang
Available online  , doi: 10.11999/JEIT250141
Abstract:
  Objective  Human activities in shallow-sea regions have increased with the progressive development and utilization of marine environments. However, interference from noise sources such as water currents, vessels, and marine organisms substantially reduces the Signal-to-Noise Ratio (SNR) of received signals, leading to performance degradation in conventional Direction of Arrival (DOA) estimation algorithms. These limitations affect accurate target localization and detection, which are essential for underwater acoustic applications. This paper proposes a method that combines Wiener filtering with deep learning, specifically an improved Convolutional Neural Network (CNN), to enhance the robustness and accuracy of DOA estimation in shallow-sea environments by addressing challenges such as noise interference and grid mismatch.  Methods  The proposed algorithm integrates Wiener filtering and deep learning to optimize DOA estimation. First, Wiener filtering is applied to the covariance matrix of received signals to reduce noise. Considering the complexity of underwater acoustic environments, this preprocessing step is essential for preserving spatial signal features and enhancing the effectiveness of network training. Second, an improved CNN architecture is designed to perform both classification and regression tasks, enabling accurate DOA angle estimation while correcting off-grid errors. A fractional label correction strategy is introduced to mitigate grid mismatch, allowing prediction errors to be compensated and estimated angles to align more closely with true values. The network is trained using a large-scale dataset of simulated underwater acoustic signals, and its feasibility is verified through real-world sea trial data.  Results and Discussions  The proposed method is validated using both simulation data and sea trial data, demonstrating superior performance under low SNR conditions. Simulation results show that incorporating Wiener filtering improves overall network performance by 25.20% (Fig. 4). The proposed algorithm achieves lower estimation errors in low SNR environments (Fig. 4, Fig. 5). Compared with the MUSIC and ESPRIT algorithms, the Root Mean Square Error (RMSE) is reduced by 93.42% and 92.14%, respectively, at an SNR of –20 dB. Moreover, as the SNR increases, the algorithm exhibits a gradual reduction in estimation error (Fig. 6). In sea trial experiments, the algorithm accurately identifies target azimuths and achieves lower estimation errors compared to conventional methods (Fig. 9, Table 1). These results indicate that the proposed method meets the requirements for robustness and accuracy in practical underwater acoustic applications.  Conclusions  This paper proposes a method for direction estimation in shallow-water environments that integrates Wiener filtering with a deep learning-based CNN to address the performance degradation caused by low SNR conditions in conventional algorithms. The combination of noise reduction preprocessing with CNN model training enhances both network efficiency and estimation accuracy. Simulation and sea trial experiments yield the following results: (1) Overall network performance is improved by 25.20% through Wiener filtering optimization; (2) The improved binary cross-entropy loss function enables the network to perform both classification and regression tasks for DOA estimation; (3) The algorithm demonstrates higher robustness to noise interference and grid mismatch, particularly under low SNR and limited snapshot conditions, making it suitable for practical deployment. However, the network is trained using datasets with known source numbers. Future research will focus on network designs that accommodate more diverse source conditions, including blind source scenarios and adaptive noise reduction models.
Cross Modal Hashing of Medical Image Semantic Mining for Large Language Model
LIU Qinghai, WU Qianlin, LUO Jia, TANG Lun, XU Liming
Available online  , doi: 10.11999/JEIT250529
Abstract:
  Objective  A novel cross-modal hashing framework driven by Large Language Models (LLMs) is proposed to address the semantic misalignment between medical images and their corresponding textual reports. The objective is to enhance cross-modal semantic representation and improve retrieval accuracy by effectively mining and matching semantic associations between modalities.  Methods  The generative capacity of LLMs is first leveraged to produce high-quality textual descriptions of medical images. These descriptions are integrated with diagnostic reports and structured clinical data using a dual-stream semantic enhancement module, designed to reinforce inter-modality alignment and improve semantic comprehension. A structural similarity-guided hashing scheme is then developed to encode both visual and textual features into a unified Hamming space, ensuring semantic consistency and enabling efficient retrieval. To further enhance semantic alignment, a prompt-driven attention template is introduced to fuse image and text features through fine-tuned LLMs. Finally, a contrastive loss function with hard negative mining is employed to improve representation discrimination and retrieval accuracy.  Results and Discussions  Experiments are conducted on a multimodal medical dataset to compare the proposed method with existing cross-modal hashing baselines. The results indicate that the proposed method significantly outperforms baseline models in terms of precision and Mean Average Precision (MAP) (Table 3; Table 4). On average, a 7.21% improvement in retrieval accuracy and a 7.72% increase in MAP are achieved across multiple data scales, confirming the effectiveness of the LLM-driven semantic mining and hashing approach.  Conclusions  Experiments are conducted on a multimodal medical dataset to compare the proposed method with existing cross-modal hashing baselines. The results indicate that the proposed method significantly outperforms baseline models in terms of precision and Mean Average Precision (MAP) (Table 3; Table 4). On average, a 7.21% improvement in retrieval accuracy and a 7.72% increase in MAP are achieved across multiple data scales, confirming the effectiveness of the LLM-driven semantic mining and hashing approach.
BIRD1445: Large-scale Multimodal Bird Dataset for Ecological Monitoring
WANG Hongchang, XIAN Fengyu, XIE Zihui, DONG Miaomiao, JIAN Haifang
Available online  , doi: 10.11999/JEIT250647
Abstract:
  Objective  With the rapid advancement of Artificial Intelligence (AI) and growing demands in ecological monitoring, high-quality multimodal datasets have become essential for training and deploying AI models in specialized domains. Existing bird datasets, however, face notable limitations, including challenges in field data acquisition, high costs of expert annotation, limited representation of rare species, and reliance on single-modal data. To overcome these constraints, this study proposes an efficient framework for constructing large-scale multimodal datasets tailored to ecological monitoring. By integrating heterogeneous data sources, employing intelligent semi-automatic annotation pipelines, and adopting multi-model collaborative validation based on heterogeneous attention fusion, the proposed approach markedly reduces the cost of expert annotation while maintaining high data quality and extensive modality coverage. This work offers a scalable and intelligent strategy for dataset development in professional settings and provides a robust data foundation for advancing AI applications in ecological conservation and biodiversity monitoring.  Methods  The proposed multimodal dataset construction framework integrates multi-source heterogeneous data acquisition, intelligent semi-automatic annotation, and multi-model collaborative verification to enable efficient large-scale dataset development. The data acquisition system comprises distributed sensing networks deployed across natural reserves, incorporating high-definition intelligent cameras, custom-built acoustic monitoring devices, and infrared imaging systems, supplemented by standardizedpublic data to enhance species coverage and modality diversity. The intelligent annotation pipeline is built upon four core automated tools: (1) spatial localization annotation leverages object detection algorithms to generate bounding boxes; (2) fine-grained classification employs Vision Transformer models for hierarchical species identification; (3) pixel-level segmentation combines detection outputs with SegGPT models to produce instance-level masks; and (4) multimodal semantic annotation uses Qwen large language models to generate structured textual descriptions. To ensure annotation quality and minimize manual verification costs, a multi-scale attention fusion verification mechanism is introduced. This mechanism integrates seven heterogeneous deep learning models, each with different feature perception capacities across local detail, mid-level semantic, and global contextual scales. A global weighted voting module dynamically assigns fusion weights based on model performance, while a prior knowledge-guided fine-grained decision module applies category-specific accuracy metrics and Top-K model selection to enhance verification precision and computational efficiency.  Results and Discussions  The proposed multi-scale attention fusion verification method dynamically assesses data quality based on heterogeneous model predictions, forming the basis for automated annotation validation. Through optimized weight allocation and category-specific verification strategies, the collaborative verification framework evaluates the effect of different model combinations on annotation accuracy. Experimental results demonstrate that the optimal verification strategy—achieved by integrating seven specialized models—outperforms all baseline configurations across evaluation metrics. Specifically, the method attains a Top-1 accuracy of 95.39% on the CUB-200-2011 dataset, exceeding the best-performing single-model baseline, which achieves 91.79%, thereby yielding a 3.60% improvement in recognition precision. The constructed BIRD1445 dataset, comprising 3.54 million samples spanning 1,445 bird species and four modalities, outperforms existing datasets in terms of coverage, quality, and annotation accuracy. It serves as a robust benchmark for fine-grained classification, density estimation, and multimodal learning tasks in ecological monitoring.  Conclusions  This study addresses the challenge of constructing large-scale multimodal datasets for ecological monitoring by integrating multi-source data acquisition, intelligent semi-automatic annotation, and multi-model collaborative verification. The proposed approach advances beyond traditional manual annotation workflows by incorporating automated labeling pipelines and heterogeneous attention fusion mechanisms as the core quality control strategy. Comprehensive evaluations on benchmark datasets and real-world scenarios demonstrate the effectiveness of the method: (1) the verification strategy improves annotation accuracy by 3.60% compared to single-model baselines on the CUB-200-2011 dataset; (2) optimal trade-offs between precision and computational efficiency are achieved using Top-K = 3 model selection, based on performance–complexity alignment; and (3) in large-scale annotation scenarios, the system ensures high reliability across 1,445 species categories. Despite its effectiveness, the current approach primarily targets species with sufficient data. Future work should address the representation of rare and endangered species by incorporating advanced data augmentation and few-shot learning techniques to mitigate the limitations posed by long-tail distributions.
Adversarial Transferability Attack on Deep Neural Networks Through Spectral Coefficient Decay
QIAN Yaguan, KONG Yaxin, CHEN Kecheng, SHEN Yunkai, BAO Qiqi, JI Shouling
Available online  , doi: 10.11999/JEIT250157
Abstract:
  Objective  The rapid advancement of machine learning has accelerated the development of artificial intelligence applications based on big data. Deep Neural Networks (DNNs) are now widely used in real-world systems, including autonomous driving and facial recognition. Despite their advantages, DNNs remain vulnerable to adversarial attacks, particularly in image recognition, which exposes significant security risks. Attackers can generate carefully crafted adversarial examples that appear visually indistinguishable from benign inputs but cause incorrect model predictions. In critical applications, such attacks may lead to system failures. Studying adversarial attack strategies is crucial for designing effective defenses and improving model robustness in practical deployments.  Methods  This paper proposes a black-box adversarial attack method termed Spectral Coefficient Attenuation (SCA), based on Spectral Coefficient Decay (SCD). SCA perturbs spectral coefficients during the iterative generation of adversarial examples by attenuating the amplitude of each frequency component. This attenuation introduces diverse perturbation patterns and reduces the dependency of adversarial examples on the source model, thereby improving their transferability across models.  Results and Discussions  This study generated adversarial examples using three naturally trained deep models based on Convolutional Neural Network (CNN) architectures and evaluated their attack success rates against several widely used convolutional image recognition models. Compared with MI-FGSM, the proposed method achieved average increases in attack success rates of 4.7%, 4.1%, and 4.9%, respectively. To further assess the effectiveness of the SCA algorithm, additional experiments were conducted on vision models based on Transformer architectures. In these settings, the SCA method yielded average success rate improvements of 4.6%, 4.5%, and 3.7% compared with MI-FGSM.  Conclusions  The proposed method attenuates the energy of individual spectral components during the iterative generation of adversarial examples in the frequency domain. This reduces reliance on specific features of the source model, mitigates the risk of overfitting, and enhances black-box transferability across models with different architectures. Extensive experiments on the ImageNet validation set confirm the effectiveness of the SCA method. Whether targeting CNN-based or Transformer-based models, the SCA algorithm consistently achieves high transferability and robust attack performance.
Depression Screening Method Driven by Global-Local Feature Fusion
ZHANG Siyong, QIU Jiefan, ZHAO Xiangyun, XIAO Kejiang, CHEN Xiaofu, MAO Keji
Available online  , doi: 10.11999/JEIT250035
Abstract:
  Objective  Depression is a globally prevalent mental disorder that poses a serious threat to the physical and mental health of millions of individuals. Early screening and diagnosis are essential to reducing severe consequences such as self-harm and suicide. However, conventional questionnaire-based screening methods are limited by their dependence on the reliability of respondents’ answers, their difficulty in balancing efficiency with accuracy, and the uneven distribution of medical resources. New auxiliary screening approaches are therefore needed. Existing Artificial Intelligence (AI) methods for depression detection based on facial features primarily emphasize global expressions and often overlook subtle local cues such as eye features. Their performance also declines in scenarios where partial facial information is obscured, for instance by masks, and they raise privacy concerns. This study proposes a Global-Local Fusion Axial Network (GLFAN) for depression screening. By jointly extracting global facial and local eye features, this approach enhances screening accuracy and robustness under complex conditions. A corresponding dataset is constructed, and experimental evaluations are conducted to validate the method’s effectiveness. The model is deployed on edge devices to improve privacy protection while maintaining screening efficiency, offering a more objective, accurate, efficient, and secure depression screening solution that contributes to mitigating global mental health challenges.  Methods  To address the challenges of accuracy and efficiency in depression screening, this study proposes GLFAN. For long-duration consultation videos with partial occlusions such as masks, data preprocessing is performed using OpenFace 2.0 and facial keypoint algorithms, combined with peak detection, clustering, and centroid search strategies to segment the videos into short sequences capturing dynamic facial changes, thereby enhancing data validity. At the model level, GLFAN adopts a dual-branch parallel architecture to extract global facial and local eye features simultaneously. The global branch uses MTCNN for facial keypoint detection and enhances feature extraction under occlusion using an inverted bottleneck structure. The local branch detects eye regions via YOLO v7 and extracts eye movement features using a ResNet-18 network integrated with a convolutional attention module. Following dual-branch feature fusion, an integrated convolutional module optimizes the representation, and classification is performed using an axial attention network.  Results and Discussions  The performance of GLFAN is evaluated through comprehensive, multi-dimensional experiments. On the self-constructed depression dataset, high accuracy is achieved in binary classification tasks, and non-depression and severe depression categories are accurately distinguished in four-class classification. Under mask-occluded conditions, a precision of 0.72 and a precision of 0.690 are obtained for depression detection. Although these values are lower than the precision of 0.87 and precision of 0.840 observed under non-occluded conditions, reliable screening performance is maintained. Compared with other advanced methods, GLFAN achieves higher recall and F1 scores. On the public AVEC2013 and AVEC2014 datasets, the model achieves lower Mean Absolute Error (MAE) values and shows advantages in both short- and long-sequence video processing. Heatmap visualizations indicate that GLFAN dynamically adjusts its attention according to the degree of facial occlusion, demonstrating stronger adaptability than ResNet-50. Edge device tests further confirm that the average processing delay remains below 17.56 milliseconds per frame, and stable performance is maintained under low-bandwidth conditionsThe performance of GLFAN is evaluated through comprehensive, multi-dimensional experiments. On the self-constructed depression dataset, high accuracy is achieved in binary classification tasks, and non-depression and severe depression categories are accurately distinguished in four-class classification. Under mask-occluded conditions, a precision of 0.72 and a recall of 0.690 are obtained for depression detection. Although these values are lower than the precision of 0.87 and recall of 0.840 observed under non-occluded conditions, reliable screening performance is maintained. Compared with other advanced methods, GLFAN achieves higher recall and F1 scores. On the public AVEC2013 and AVEC2014 datasets, the model achieves lower Mean Absolute Error (MAE) values and shows advantages in both short- and long-sequence video processing. Heatmap visualizations indicate that GLFAN dynamically adjusts its attention according to the degree of facial occlusion, demonstrating stronger adaptability than ResNet-50. Edge device tests further confirm that the average processing delay remains below 17.56 frame/s, and stable performance is maintained under low-bandwidth conditions.  Conclusions  This study proposes a depression screening approach based on edge vision technology. A lightweight, end-to-end GLFAN is developed to address the limitations of existing screening methods. The model integrates global facial features extracted via MTCNN with local eye-region features captured by YOLO v7, followed by effective feature fusion and classification using an Axial Transformer module. By emphasizing local eye-region information, GLFAN enhances performance in occluded scenarios such as mask-wearing. Experimental validation using both self-constructed and public datasets demonstrates that GLFAN reduces missed detections and improves adaptability to short-duration video inputs compared with existing models. Grad-CAM visualizations further reveal that GLFAN prioritizes eye-region features under occluded conditions and shifts focus to global facial features when full facial information is available, confirming its context-specific adaptability. The model has been successfully deployed on edge devices, offering a lightweight, efficient, and privacy-conscious solution for real-time depression screening.
Edge Network Data Scheduling Optimization Method Integrating Improved Jaya and Cluster Center Selection Algorithm
YANG Wensheng, PAN Chengsheng
Available online  , doi: 10.11999/JEIT250317
Abstract:
  Objective  The rapid advancement of technologies such as artificial intelligence and the Internet of Things has placed increasing strain on traditional centralized cloud computing architectures, which struggle to meet the communication and computational demands of large-scale data processing. Due to the physical separation between cloud servers and end-users, data transmission typically incurs considerable latency and energy consumption. Therefore, edge computing—by deploying computing and storage resources closer to users, has emerged as a viable paradigm for supporting data-intensive and latency-sensitive applications. However, effectively addressing the challenges of data-intensive services in edge computing environments, such as efficient edge node clustering and resource scheduling, remains a key issue. This study proposes a data scheduling optimization method for edge networks that integrates an improved Jaya algorithm with a cluster center selection strategy. Specifically, for data-intensive services, the method partitions edge nodes into clusters and identifies optimal cluster centers. Data are first aggregated at these centers before being transmitted to the cloud. By leveraging cluster-based aggregation, the method facilitates more efficient data scheduling and improved resource management in edge environments.  Methods  The proposed edge network data scheduling optimization method comprises two core components: a shortest-path selection algorithm based on an improved Jaya algorithm and an optimal cluster center selection algorithm. The scheduling framework accounts for both the shortest communication paths among edge nodes and the availability of network resources. The improved Jaya algorithm incorporates a cosine-based nonlinear decay function and a multi-stage search strategy to dynamically optimize inter-node paths. The nonlinear decay function modulates the variation of random factors across iterations, allowing adaptive adjustment of the algorithm’s exploration capacity. This mechanism helps prevent premature convergence and reduces the likelihood of becoming trapped in local optima during the later optimization stages. To further enhance performance, a multi-stage search strategy divides the optimization process into two phases: an exploration phase during early iterations, which prioritizes global search across the solution space, and an exploitation phase during later iterations, which refines solutions locally. This staged approach improves the trade-off between convergence speed and solution accuracy, increasing the algorithm’s robustness in complex edge network environments. Based on the optimized paths and available bandwidth, a criterion is established for selecting the initial cluster center. Subsequently, a selection scheme for additional cluster centers is formulated by evaluating inter-cluster center distances. Finally, a partitioning method assigns edge nodes to their respective clusters based on the optimized topology.  Results and Discussions  The simulation experiments comprise two parts: performance evaluation of the improved Jaya algorithm (Jaya*) and analysis of the cluster partitioning scheme. To assess convergence speed and optimization accuracy, three benchmark test functions are used to compare Jaya* with four existing algorithms: Simulated Annealing (SA), Genetic Algorithm (GA), Ant Colony Optimization (ACO), and the standard Jaya algorithm. Building on these results, two additional experiments—cluster center selection and cluster partitioning—are conducted to evaluate the feasibility and effectiveness of the proposed optimal cluster center selection algorithm for resource scheduling. A parameter sensitivity analysis using the multi-modal Rastrigin function is performed to investigate the effects of different population sizes and maximum iteration counts on optimization accuracy and stability (Table 2 and Table 3). The optimal configuration is determined to be \begin{document}$ {\text{po}}{{\text{p}}_{{\text{size}}}} = 50 $\end{document} and \begin{document}$ {t_{\max }} = 500 $\end{document}, which achieves a favorable balance between accuracy and computational efficiency. Subsequently, a multi-algorithm comparison experiment is carried out under consistent conditions. The improved Jaya algorithm outperforms the four alternatives in convergence speed and optimization accuracy across three standard functions: Sphere (Fig. 4), Rastrigin (Fig. 5), and Griewank (Fig. 6). The algorithm also demonstrates superior stability. Its convergence trajectory is characterized by a rapid initial decline followed by gradual stabilization in later stages. Based on these findings, the cluster center selection algorithm is applied to tactical edge networks of varying scales—25, 38, and 50 nodes (Fig. 7). The parameter mi is calculated (Fig. 8), and various numbers of cluster centers are set to complete center selection and cluster member assignment (Table 5). Evaluation using the Average Sum of Squared Errors (AvgSSE) under different cluster center counts reveals that the minimum AvgSSE for all three network sizes occurs when the number of cluster centers is 4 (Table 6), indicating that this configuration yields the optimal clustering outcome. Therefore, the proposed method effectively selects cluster centers and derives the optimal clustering configuration (Fig. 9), while maintaining low clustering error and enhancing the efficiency and accuracy of resource scheduling. Finally, in a 38-node edge network scenario with four cluster centers, a multi-algorithm cluster partitioning comparison is conducted (Table 7). The improved Jaya algorithm achieves the best AvgSSE result of 16.22, significantly outperforming the four baseline algorithms. These results demonstrate its superiority in convergence precision and global search capability.  Conclusions  To address data resource scheduling challenges in edge computing environments, this study proposes an edge network data scheduling optimization method that integrates an improved Jaya algorithm with a cluster center selection strategy. The combined approach achieves high clustering accuracy, robustness, and generalization performance. It effectively enhances path planning precision and central node selection, leading to improved data transmission performance and resource utilization in edge networks.
Optimized Design of Non-Transparent Bridge for Heterogeneous Interconnects in Hyper-converged Infrastructure
ZHENG Rui, SHEN Jianliang, LV Ping, DONG Chunlei, SHAO Yu, ZHU Zhengbin
Available online  , doi: 10.11999/JEIT250272
Abstract:
  Objective  The integration of heterogeneous computing resource clusters into modern Hyper-Converged Infrastructure (HCI) systems imposes stricter performance requirements in latency, bandwidth, throughput, and cross-domain transmission stability. Traditional HCI systems primarily rely on the Ethernet TCP/IP protocol, which exhibits inherent limitations, including low bandwidth efficiency, high latency, and limited throughput. Existing PCIe Switch products typically employ Non-Transparent Bridges (NTBs) for conventional dual-system connections or intra-server communication; however, they do not meet the performance demands of heterogeneous cross-domain transmission within HCI environments. To address this limitation, a novel Dual-Mode Non-Transparent Bridge Architecture (D-MNTBA) is proposed to support dual transmission modes. D-MNTBA combines a fast transmission mode via a bypass mechanism with a stable transmission mode derived from the Traditional Data Path Architecture (TDPA), thereby aligning with the data characteristics and cross-domain streaming demands of HCI systems. Hardware-level enhancements in address and ID translation schemes enable D-MNTBA to support more complex mappings while minimizing translation latency. These improvements increase system stability and effectively support the cross-domain transmission of heterogeneous data in HCI systems.  Methods  To overcome the limitations of traditional single-pass architectures and the bypass optimizations of the TDPA, the proposed D-MNTBA incorporates both a fast transmission path and a stable transmission path. This dual-mode design enables the NTB to leverage the data characteristics of HCI systems for telegram-based streaming, thereby reducing dependence on intermediate protocols and data format conversions. The stable transmission mode ensures reliable message delivery, while the fast transmission mode—enhanced through hardware-level optimizations in address and ID translation—supports high-real-time cross-domain communication. This combination improves overall transmission performance by reducing both latency and system overhead. To meet the low-latency demands of the bypass transmission path, the architecture implements hardware-level enhancements to the address and ID conversion modules. The address translation module is expanded with a larger lookup table, allowing for more complex and flexible mapping schemes. This enhancement enables efficient utilization of non-contiguous and fragmented address spaces without compromising performance. Simultaneously, the ID conversion module is optimized through multiple conversion strategies and streamlined logic, significantly reducing the time required for ID translation.  Results and Discussions  Address translation in the proposed D-MNTBA is validated through emulation within a constructed HCI environment. The simulation log for indirect address translation shows no errors or deadlocks, and successful hits are observed on BAR2/3. During dual-host disk access, packet header addresses and payload content remain consistent, with no packet loss detected (Fig. 14), indicating that indirect address translation is accurately executed under D-MNTBA. ID conversion performance is evaluated by comparing the proposed architecture with the TDPA implemented in the PEX8748 chip. The switch based on D-MNTBA exhibits significantly shorter ID conversion times. A maximum reduction of approximately 34.9% is recorded, with an ID conversion time of 71 ns for a 512-byte payload (Fig. 15). These findings suggest that the ID function mapping method adopted in D-MNTBA effectively reduces conversion latency and enhances system performance. Throughput stability is assessed under sustained heavy traffic with payloads ranging from 256 to 2048 bytes. The maximum throughputs of D-MNTBA, the Ethernet card, and PEX8748 are measured at 1.36 GB/s, 0.97 GB/s, and 0.9 GB/s, respectively (Fig. 16). Compared to PEX8748 and the Ethernet architecture, D-MNTBA improves throughput by approximately 51.1% and 40.2%, respectively, and shows the slowest degradation trend, reflecting superior stability in heterogeneous cross-domain transmission. Bandwidth comparison reveals that D-MNTBA outperforms TDPA and the Ethernet card, with bandwidth improvements of approximately 27.1% and 19.0%, respectively (Fig. 17). These results highlight the significant enhancement in cross-domain transmission performance achieved by the proposed architecture in heterogeneous environments.  Conclusions  This study proposes a Dual-Mode D-MNTBA to address the challenges of heterogeneous interconnection in HCI systems. By integrating a fast transmission path enabled by a bypass architecture with the stable transmission path of the TDPA, D-MNTBA accommodates the specific data characteristics of cross-domain transmission in heterogeneous environments and enables efficient message routing. D-MNTBA enhances transmission stability while improving system-wide performance, offering robust support for high-real-time cross-domain transmission in HCI. It also reduces latency and overhead, thereby improving overall transmission efficiency. Compared with existing transmission schemes, D-MNTBA achieves notable gains in performance, making it a suitable solution for the demands of heterogeneous domain interconnects in HCI systems. However, the architectural enhancements, particularly the bypass design and associated optimizations, increase logic resource utilization and power consumption. Future work should focus on refining hardware design, layout, and wiring strategies to reduce logic complexity and resource consumption without compromising performance.
A Survey on System and Architecture Optimization Techniques for Mixture-of-Experts Large Language Models
WANG Zehao, ZHU Zhenhua, XIE Tongxin, WANG Yu
Available online  , doi: 10.11999/JEIT250407
Abstract:
The Mixture-of-Experts (MoE) framework has become a pivotal approach for enhancing the knowledge capacity and inference efficiency of Large Language Models (LLMs). Conventional methods for scaling dense LLMs have reached significant limitations in training and inference due to computational and memory constraints. MoE addresses these challenges by distributing knowledge representation across specialized expert sub-networks, enabling parameter expansion while maintaining efficiency through sparse expert activation during inference. However, the dynamic nature of expert activation introduces substantial challenges in resource management and scheduling, necessitating targeted optimization at both the system and architectural levels. This survey focuses on the deployment of MoE-based LLMs. It first reviews the definitions and developmental trajectory of MoE, followed by an in-depth analysis of current system-level optimization strategies and architectural innovations tailored to MoE. The paper concludes by summarizing key findings and proposing prospective optimization techniques for MoE-based LLMs.  Significance   The MoE mechanism offers a promising solution to the computational and memory limitations of dense LLMs. By distributing knowledge representation across specialized expert sub-networks, MoE facilitates model scaling without incurring prohibitive computational costs. This architecture alleviates the bottlenecks associated with training and inference in traditional dense models, marking a notable advance in LLM research. Nonetheless, the dynamic expert activation patterns inherent to MoE introduce new challenges in resource scheduling and management. Overcoming these challenges requires targeted system- and architecture-level optimizations to fully harness the potential of MoE-based LLMs.  Progress   Recent advancements in MoE-based LLMs have led to the development of various optimization strategies. At the system level, approaches such as automatic parallelism, communication–computation pipelining, and communication operator fusion have been adopted to reduce communication overhead. Memory management has been improved through expert prefetching, caching mechanisms, and queue scheduling policies. To address computational load imbalance, both offline scheduling methods and runtime expert allocation strategies have been proposed, including designs that leverage heterogeneous CPU–GPU architectures. In terms of hardware architecture, innovations include dynamic adaptation to expert activation patterns, techniques to overcome bandwidth limitations, and near-memory computing schemes that improve deployment efficiency. In parallel, the open-source community has developed supporting tools and frameworks that facilitate the practical deployment and optimization of MoE-based models.  Conclusions  This survey presents a comprehensive review of system and architectural optimization techniques for MoE-based LLMs. It highlights the importance of reconciling parameter scalability with computational efficiency through the MoE framework. The dynamic nature of expert activation poses significant challenges in scheduling and resource management, which this survey systematically addresses. By evaluating current optimization techniques across both system and hardware layers, the paper offers key insights into the state of the field. It also proposes directions for future work, providing a reference for researchers and practitioners seeking to improve the performance and scalability of MoE-based models. The findings emphasize the need for continued innovation across algorithm development, system engineering, and architectural design to fully realize the potential of MoE in real-world applications.  Prospects   Future research on MoE-based LLMs is expected to advance the integration of algorithm design, system optimization, and hardware co-design. Key research directions include resolving load imbalance and maximizing resource utilization through adaptive expert scheduling algorithms, refining system frameworks to support dynamic sparse computation more effectively, and exploring hardware paradigms such as near-memory computing and hierarchical memory architectures. These developments aim to deliver more efficient and scalable MoE model deployments by fostering deeper synergy between software and hardware components.
Hybrid Far-Near Field Channel Estimation for XL-RIS Assisted Communication Systems
SHAO Kai, HUA Fanyu, WANG Guangyu
Available online  , doi: 10.11999/JEIT250306
Abstract:
  Objective  With the rapid development of sixth-generation mobile communication, Extra-Large Reconfigurable Intelligent Surfaces (XL-RIS) have attracted significant attention due to their potential to enhance spectral efficiency, expand coverage, and reduce energy consumption. However, conventional channel estimation methods, primarily based on Far-Field (FF) or near-field (NF) models, face limitations in addressing the hybrid far-NF environment that arises from the coexistence of NF spherical waves and FF planar waves in XL-RIS deployments. These limitations restrict the intelligent control capability of RIS technology due to inaccurate channel modeling and reduced estimation accuracy. To address these challenges, this paper constructs a hybrid-field channel model for XL-RIS and proposes a robust channel estimation method to resolve parameter estimation challenges under coupled FF and NF characteristics, thereby improving channel estimation accuracy in complex propagation scenarios.  Methods  For channel estimation in XL-RIS-aided communication systems, several key challenges must be addressed, including the modeling of hybrid far-NF cascaded channels, separation of FF and NF channel components, and individual parameter estimation. To capture the hybrid-field effects of XL-RIS, a hybrid-field cascaded channel model is constructed. The RIS-to-User Equipment (UE) channel is modeled as a hybrid far-NF channel, whereas the Base Station (BS)-to-RIS channel is characterized under the FF assumption. A unified representation of FF and NF models is established by introducing equivalent cascaded angles for the angle of departure and angle of arrival on the RIS side. The XL-RIS hybrid-field cascaded channel is parameterized through BS arrival angles, RIS-UE cascaded angles, and distances. To reduce the computational complexity of joint parameter estimation, a Two-Stage Hybrid-Field (TS-HF) channel estimation scheme is proposed. In the first stage, the BS arrival angle is estimated using the MUltiple SIgnal Classification (MUSIC) algorithm. In the second stage, a Hybrid-Field forward spatial smoothing Rank-reduced MUSIC (HF-RM) algorithm is proposed to estimate the parameters of the RIS-UE hybrid-field channel. The received signals are pre-processed using a forward spatial smoothing technique to mitigate multipath coherence effects. Subsequently, the Rank-reduced MUSIC (RM) algorithm is applied to separately estimate the FF and NF angle parameters, as well as the NF distance parameter. During this stage, a power spectrum comparison scheme is designed to distinguish FF and NF angles based on power spectral characteristics, thereby providing high-precision angular information to support NF distance estimation. Finally, channel attenuation is estimated using the least squares method. To validate the effectiveness of the proposed hybrid-field channel estimation scheme, comparative analyses are conducted against FF, NF, and the proposed TS-HF-RM schemes. The FF estimation approximates the hybrid-field channel using a FF channel model and estimates FF angle parameters with the MUSIC algorithm, referred to as the TS-FF-M scheme. The NF estimation applies a NF channel model to characterize the hybrid channel and estimates angle and distance parameters using the RM algorithm, referred to as the TS-NF-RM scheme. To further evaluate the estimation performance, additional benchmark schemes are considered, including the Two-Stage Near-Field Orthogonal Matching Pursuit (TS-NOMP) scheme, the Two-Stage Hybrid Orthogonal Matching Pursuit with Prior (TS-HOMP-P) scheme that requires prior knowledge of FF and NF quantities, and the Two-Stage Hybrid Orthogonal Matching Pursuit with No Prior (TS-HOMP-NP) scheme that operates without requiring such prior information.  Results and Discussions  Compared with the TS-FF-M and TS-NF-RM schemes, the proposed TS-HF-RM approach achieves effective separation and accurate estimation of both FF and NF components by jointly modeling the hybrid-field channel. The method consistently demonstrates superior estimation accuracy across a wide range of Signal-to-Noise Ratio (SNR) conditions (Fig. 4). These results confirm both the necessity of hybrid-field channel modeling and the effectiveness of the proposed estimation scheme. Experimental findings show that the TS-HF-RM approach significantly improves channel estimation performance in XL-RIS-assisted communication systems. Further comparative analysis reveals that the TS-HF-RM scheme outperforms TS-NOMP and TS-HOMP-P by mitigating power leakage effects and overcoming limitations associated with unknown path numbers through distinct processing of FF and NF components. Without requiring prior knowledge of the propagation environment, the proposed method achieves lower Normalized Mean Square Error (NMSE) while demonstrating improved robustness and estimation precision (Fig. 5). Although TS-HOMP-NP also operates without prior field information, the TS-HF-RM scheme provides superior parameter resolution, attributed to its subspace decomposition principle. Additionally, both the TS-HF-RM and TS-HOMP-P schemes exhibit improved performance as the number of pilot signals increases. However, TS-HF-RM consistently outperforms TS-HOMP-P under low-SNR conditions (0 dB). At high SNR (10 dB) with a limited number of pilot signals (<280), TS-HOMP-P temporarily achieves better performance due to its higher sensitivity to SNR. Nevertheless, the proposed TS-HF-RM approach demonstrates greater stability and adaptability under low-SNR and resource-constrained conditions (Fig. 6).  Conclusions  This study addresses the challenge of hybrid-field channel estimation for XL-RIS by constructing a hybrid-field cascaded channel model and proposing a two-stage estimation scheme. The HF-RM algorithm is specifically designed for accurate hybrid component estimation in the second stage. Theoretical analysis and simulation results demonstrate the following: (1) The hybrid-field model reduces inaccuracies associated with traditional single-field assumptions, providing a theoretical foundation for reliable parameter estimation in complex propagation environments; (2) The proposed TS-HF-RM algorithm enables high-resolution parameter estimation with effective separation of FF and NF components, achieving lower NMSE compared to hybrid-field OMP-based methods.
An Optimization Design Method for Zero-Correlation Zone Sequences Based on Newton’s Method
HU Enbo, LIU Tao, LI Yubo
Available online  , doi: 10.11999/JEIT250394
Abstract:
  Objective  Sequences with favorable correlation properties are widely applied in radar and communication systems. Sequence sets with zero or low correlation characteristics enhance radar resolution, target detection, imaging quality, and information acquisition, while also improving the omnidirectional transmission capability of massive multiple-input multiple-output (MIMO) systems. Designing aperiodic Zero Correlation Zone (ZCZ) sequence sets with excellent correlation performance is therefore critical for both wireless communication and radar applications. For example, aperiodic Z-Complementary Set (ZCS) sequence sets are often used in omnidirectional precoding for MIMO systems, whereas aperiodic ZCZ sequence sets are employed in integrated MIMO radar-communication systems. These ZCZ sequence sets are thus valuable across a range of system applications. However, most prior studies rely on analytical construction methods, which impose constraints on parameters such as sequence length and the number of sequences, thereby limiting design flexibility and practical applicability. This study proposes a numerical optimization approach for designing ZCS and aperiodic ZCZ sequence sets with improved correlation properties and greater parametric flexibility. The method minimizes the Complementary Peak Sidelobe Level (CPSL) and Weighted Peak Sidelobe Level (WPSL) using Newton’s method to achieve superior sequence performance.  Methods  This study proposes an optimization-based design method using Newton’s method to construct both aperiodic ZCS sequence sets and aperiodic ZCZ sequence sets with low sidelobe levels and flexible parameters. The optimization objective is first formulated using the CPSL and WPSL. The problem is then reformulated as an equivalent system of nonlinear equations, which is solved using Newton’s method. To reduce computation time, partial derivatives are approximated using numerical differentiation techniques. A loop iteration strategy is employed to address multiple constraints during the optimization process. To ensure algorithmic convergence, Armijo’s rule is used for step size selection, promoting stable descent of the objective function along the defined search direction.  Results and Discussions  The aperiodic ZCS sequence set is constructed using Newton’s method. As the number of sequences increases, the CPSL progressively decreases, falling below –300 dB when \begin{document}$M \geqslant 2$\end{document}. The proposed method yields better sidelobe performance than the improved Iterative Twisted Approximation (ITORX) algorithm (Fig. 1). The performance of ZCS sequences generated by both methods is evaluated under different ZCZ conditions. While both approaches achieve low CPSL, Newton’s method yields aidelobe levels closer to the ideal value (Fig. 2). Convergence behavior is assessed using CPSL and the number of iterations. The improved ITROX algorithm typically requires around 20000 iterations to converge, with increasing iterations as ZCZ size grows. In contrast, Newton’s method achieves rapid convergence within approximately 10 iterations (Figs. 3 and 4). The aperiodic ZCZ sequence set constructed using Newton’s method exhibits autocorrelation and cross-correlation peak sidelobe levels below –300 dB within the ZCZ. Moreover, Newton’s method achieves the lowest WPSL, offering the best overall performance among all tested methods (Fig. 5). The smooth convergence curves further confirm the algorithm’s stability when applied to aperiodic ZCZ sequence construction (Fig. 6).  Conclusions  This study proposes an optimization-based algorithm for designing aperiodic ZCS and aperiodic ZCZ sequence sets using Newton’s method, aiming to address the limitations of fixed parameters and high peak sidelobe levels found in existing approaches. Two optimization problems are formulated by minimizing the WPSL and CPSL, respectively. To simplify computation, the optimization tasks are converted into systems of nonlinear equations, which are solved using Newton’s method. The Jacobian matrix is computed via numerical differentiation to reduce computational cost. A loop iteration strategy is introduced to meet multiple constraints in the construction of aperiodic ZCZ sequences. Simulation results confirm that the proposed method yields sequence sets with excellent correlation properties and flexible parameter configurations. By tuning the weighting coefficients, low sidelobe levels can be achieved in specific regions of interest, accommodating different application requirements. The combination of flexible design parameters and favorable correlation performance makes the proposed sequences suitable for a wider range of practical scenarios.
Continuous Federation of Noise-resistant Heterogeneous Medical Dialogue Using the Trustworthiness-based Evaluation
LIU Yupeng, ZHANG Jiang, TANG Shichen, MENG Xin, MENG Qingfeng
Available online  , doi: 10.11999/JEIT250057
Abstract:
  Objective   To address the key challenges of client model heterogeneity, data distribution heterogeneity, and text noise in medical dialogue federated learning, this paper proposes a trustworthiness-based, noise-resistant heterogeneous medical dialogue federated learning method, termed FedRH. FedRH enhances robustness by improving the objective function, aggregation strategy, and local update process, among other components, based on credibility evaluation.  Methods   Model training is divided into a local training stage and a heterogeneous federated learning stage. During local training, text noise is mitigated using a symmetric cross-entropy loss function, which reduces the risk of overfitting to noisy text. In the heterogeneous federated learning stage, an adaptive aggregation mechanism incorporates clean, noisy, and heterogeneous client texts by evaluating their quality. Local parameter updates consider both local and global parameters simultaneously, enabling continuous adaptive updates that improve resistance to both random and structured (syntax/semantic) noise and model heterogeneity. The main contributions are threefold: (1) A local noise-resistant training strategy that uses symmetric cross-entropy loss to prevent overfitting to noisy text during local training; (2) A heterogeneous federated learning approach based on client trustworthiness, which evaluates each client’s text quality and learning effectiveness to compute trust scores. These scores are used to adaptively weight clients during model aggregation, thereby reducing the influence of low-quality data while accounting for text heterogeneity; (3) A local continuous adaptive aggregation mechanism, which allows the local model to integrate fine-grained global model information. This approach reduces the adverse effects of global model bias caused by heterogeneous and noisy text on local updates.  Results and Discussions   The effectiveness of the proposed model is systematically validated through extensive, multi-dimensional experiments. The results indicate that FedRH achieves substantial improvements over existing methods in noisy and heterogeneous federated learning scenarios (Table 2, Table 3). The study also presents training process curves for both heterogeneous models (Figure 3) and isomorphic models (Figure 6), supplemented by parameter sensitivity analysis, ablation experiments, and a case study.  Conclusions   The proposed FedRH framework significantly enhances the robustness of federated learning for medical dialogue tasks in the presence of heterogeneous and noisy text. The main conclusions are as follows: (1) Compared to baseline methods, FedRH achieves superior performance in client-side models under heterogeneous and noisy text conditions. It demonstrates improvements across multiple metrics, including precision, recall, and factual consistency, and converges more rapidly during training. (2) Ablation experiments confirm that both the symmetric cross-entropy-based local training strategy and the credibility-weighted heterogeneous aggregation approach contribute to performance gains.
A Collaborative Detection Method for Bauxite Quality Parameters Based on the Fusion of G-DPN and Near-Infrared Spectroscopy
ZOU Liang, REN Kelong, WU Hao, XU Zhibin, TAN Zhiyi, LEI Meng
Available online  , doi: 10.11999/JEIT250240
Abstract:
  Objective  Bauxite is a critical non-metallic mineral resource used in aluminum production, ceramic manufacturing, and refractory material processing. As global demand for aluminum and its derivatives continues to rise, improving the efficiency of bauxite resource utilization is essential. Accurate determination of quality parameters supports the reduction of waste from low-grade ores during smelting and improves overall process optimization. However, traditional chemical analyses are time-consuming, costly, complex, and subject to human error. Existing rapid testing methods, often based on machine learning, typically predict individual quality indicators and overlook correlations among multiple parameters. Deep learning, particularly multi-task learning, offers a solution to this limitation. Near-InfraRed (NIR) spectroscopy, a real-time, non-destructive analytical technique, is especially suited for assessing mineral quality. This study proposes a multi-indicator collaborative detection model—Gate-Depthwise Pointwise Network (G-DPN)—based on NIR spectroscopy to enable the simultaneous prediction of multiple bauxite quality parameters. The proposed approach addresses the limitations of conventional methods and supports efficient, accurate, and cost-effective real-time quality monitoring in industrial settings.  Methods  To accurately model the nonlinear relationships between NIR spectral features and bauxite quality parameters while leveraging inter-parameter correlations, this study proposes a dedicated representation model, G-DPN. The model incorporates large-kernel DepthWise Convolution (DWConv) to extract long-range dependencies within individual spectral channels, and PointWise Convolution (PWConv) to enable inter-channel feature fusion. A Spatial Squeeze-and-Excitation (sSE) mechanism is introduced to enhance spatial feature weighting, and residual connections support the integration of deep features. To further improve task differentiation, a Custom Gate Control (CGC) module is added to separate shared and task-specific features. Orthogonal constraints are applied within this module to reduce feature redundancy. Gate-controlled fusion enables each branch to focus on extracting task-relevant information while preserving shared representations. Additionally, quality parameter labels are normalized to address scale heterogeneity, allowing the model to establish a stable nonlinear mapping between spectral inputs and multiple output parameters.  Results and Discussions  This study applies large convolution kernels in DWConv to capture long-range dependencies within individual spectral channels (Fig. 3). Compared with conventional small-sized kernels (e.g., 3×3), which increase the receptive field but exhibit limited focus on critical spectral regions, large kernels enable more concentrated activation in key bands, thereby enhancing model sensitivity (Fig. 4). Empirical results confirm that the use of large kernels improves prediction accuracy (Table 6). Furthermore, compared to Transformer-based models, DWConv with large kernels achieves comparable accuracy with fewer parameters, offering computational efficiency. The CGC module effectively disentangles shared and task-specific features while applying orthogonal constraints to reduce redundancy. Its dynamic fusion mechanism enables adaptive feature sharing across tasks without compromising task-specific learning, thereby mitigating task interference and accounting for sample correlations (Fig. 6). Relative to conventional multi-task learning frameworks, the CGC-based architecture demonstrates superior performance in multi-parameter prediction (Table 6).  Conclusions  This study proposes a deep learning approach that integrates large-kernel DWConv and a CGC module for multi-parameter prediction of bauxite quality using NIR spectroscopy. DWConv captures long-range dependencies within spectral channels, while the CGC module leverages inter-parameter correlations to enhance feature sharing and reduce task interference. This design mitigates the effects of spectral peak overlap and establishes a robust nonlinear mapping between spectral features and quality parameters. Experiments on 424 bauxite samples show that the proposed G-DPN model achieves \begin{document}${R^2}$\end{document} values of 0.9226, 0.9377, and 0.9683 for aluminum, silicon, and iron content, respectively—outperforming conventional machine learning and existing deep learning methods. These results highlight the potential of combining NIR spectroscopy with G-DPN for accurate, efficient, and scalable mineral quality analysis, contributing to the sustainable utilization of bauxite resources.
OTFS Communication Link Construction and Three-Dimensional Constellation Encryption Design
MA Yingjie, LIU Yueheng, ZHAO Geng, ZHAO Mingjing, WANG Dan
Available online  , doi: 10.11999/JEIT250181
Abstract:
  Objective  With the advancement of Sixth-Generation (6G) communication technologies, the demand for highly reliable and secure transmission in high-speed mobile scenarios has grown significantly. Orthogonal Time Frequency Space (OTFS) modulation has emerged as a promising solution due to its superior transmission reliability in high-mobility environments. However, research on the security aspects of OTFS systems remains limited. To address this challenge, this study proposes a Three-Dimensional (3D) constellation encryption scheme for OTFS communication links. The objective is to enhance the security of OTFS systems while preserving communication performance by designing a novel 3D constellation structure and implementing physical-layer dynamic encryption based on a Coupled Map Lattice (CML) spatiotemporal chaotic system. This research provides a secure and efficient transmission strategy for future wireless communication systems.  Methods  The integration of 3D constellation encryption with OTFS modulation enhances the security of OTFS systems. First, a novel 3D constellation is designed, where 16 constellation points are uniformly distributed on the surface of a sphere centered at the origin. By optimizing the spatial distribution of constellation points, the minimum Euclidean Distance (MED) is increased and the average transmission power is reduced, thereby improving Bit Error Rate (BER) performance. Second, a dynamic encryption scheme for the 3D constellation over the OTFS link is proposed. In this scheme, pseudo-random sequences are generated using the CML system, quantized into 16-bit binary sequences, and employed to control both the selection of mapping rules and the rotational encryption of the 3D constellation. The encryption process operates as follows: first, the mapping rules are dynamically selected based on the quantized chaotic sequences; second, the rotation angles of the constellation points around the X, Y, and Z axes are determined by the same chaotic sequences. This dual encryption mechanism enhances the system’s security while maintaining reliable communication performance.  Results and Discussions  The performance of the proposed scheme is evaluated through extensive simulations. The results show that the optimized 3D constellation achieves a 22% improvement in MED compared with traditional 2D 16-quadrature amplitude modulation (16QAM). Compared with existing 3D constellations, the proposed design increases the MED by 5%, reduces average transmission power by 11%, and improves the Constellation Figure of Merit (CMF) by 10% (Table 2, Table 3). Superior communication performance is also demonstrated, with a BER improvement of approximately 0.6 dB over existing methods (Fig. 5, Fig. 6). In terms of security, the encryption scheme provides a key space of 10120, effectively resisting brute-force attacks. Key sensitivity tests confirm that even minor changes to the initial key result in completely different decrypted images (Fig. 7), verifying the robustness of the encryption mechanism. The encrypted images exhibit uniform histogram distributions (Fig. 8) and correlation coefficients near zero (Table 6), indicating strong resistance to statistical attacks. Additionally, the scheme maintains reliable performance under noisy conditions, with decrypted images preserving high quality even at low Signal-to-Noise Ratios (SNR) (Fig. 9).  Conclusions  This study combines OTFS communication links with 3D constellation encryption to realize physical-layer security for OTFS systems, enhancing both communication performance and transmission security. The proposed 3D constellation improves the MED by 5%, reduces average power by 11%, and increases the CMF by 10%, thereby enhancing BER performance. The dual dynamic encryption mechanism based on a CML system offers strong security. Simulation results show that the proposed scheme achieves a favorable balance between communication performance and security. BER performance improves by approximately 0.6 dB, while the large key space, high key sensitivity, strong resistance to statistical attacks, and good robustness collectively ensure enhanced security. The combination of OTFS modulation and 3D constellation encryption not only improves spectrum resource utilization but also strengthens channel confidentiality, enhancing resilience to interference and potential attacks. This scheme improves transmission security without compromising communication performance, offering broad application prospects in fields such as the Internet of Things, unmanned aerial vehicles, satellite communication, and ocean communication. The proposed approach is expected to provide more secure and reliable transmission solutions for future wireless communication systems.
Precise Hand Joint Motion Analysis Driven by Complex Physiological Information
YAN Jiaqing, LIU Gengchen, ZHOU Qingqi, XUE Weiqi, ZHOU Weiao, TIAN Yunzhi, WANG Jiaju, DONG Zhekang, LI Xiaoli
Available online  , doi: 10.11999/JEIT250033
Abstract:
  Objective  The human hand is a highly dexterous organ essential for performing complex tasks. However, dysfunction due to trauma, congenital anomalies, or disease substantially impairs daily activities. Restoring hand function remains a major challenge in rehabilitation medicine. Virtual Reality (VR) technology presents a promising approach for functional recovery by enabling hand pose reconstruction from surface ElectroMyoGraphy (sEMG) signals, thereby facilitating neural plasticity and motor relearning. Current sEMG-based hand pose estimation methods are limited by low accuracy and coarse joint resolution. This study proposes a new method to estimate the motion of 15 hand joints using eight-channel sEMG signals, offering a potential improvement in rehabilitation outcomes and quality of life for individuals with hand impairment.  Methods  The proposed method, termed All Hand joints Posture Estimation (AHPE), incorporates a continuous denoising network that combines sparse attention and multi-channel attention mechanisms to extract spatiotemporal features from sEMG signals. A dual-decoder architecture estimates both noisy hand poses and the corresponding correction ranges. These outputs are subsequently refined using a Bidirectional Long Short-Term Memory (BiLSTM) network to improve pose accuracy. Model training employs a composite loss function that integrates Mean Squared Error (MSE) and Kullback-Leibler (KL) divergence to enhance joint angle estimation and capture inter-joint dependencies. Performance is evaluated using the NinaproDB8 and NinaproDB5 datasets, which provide sEMG and hand pose data for single-finger and multi-finger movements, respectively.  Results and Discussions  The AHPE model outperforms existing methods—including CNN-Transformer, DKFN, CNN-LSTM, TEMPOnet, and RPC-Net—in estimating hand poses from multi-channel sEMG signals. In within-subject validation (Table 1), AHPE achieves a Root Mean Squared Error (RMSE) of 2.86, a coefficient of determination (R2) of 0.92, and a Mean Absolute Deviation (MAD) of 1.79° for MetaCarPophalangeal (MCP) joint rotation angle estimation. In between-subject validation (Table 2), the model maintains high accuracy with an RMSE of 3.72, an R2 of 0.88, and an MAD of 2.36°, demonstrating strong generalization. The model’s capacity to estimate complex hand gestures is further confirmed using the NinaproDB5 dataset. Estimated hand poses are visualized with the Mano Torch hand model (Fig. 4, Fig. 5). The average R2 values for finger joint extension estimation are 0.72 (thumb), 0.692 (index), 0.696 (middle), 0.689 (ring), and 0.696 (little finger). Corresponding RMSE values are 10.217°, 10.257°, 10.290°, 10.293°, and 10.303°, respectively. A grid error map (Fig. 6) highlights prediction accuracy, with red regions indicating higher errors.  Conclusions  The AHPE model offers an effective approach for estimating hand poses from sEMG signals, addressing key challenges such as signal noise, high dimensionality, and inter-individual variability. By integrating mixed attention mechanisms with a dual-decoder architecture, the model enhances both accuracy and robustness in multi-joint hand pose estimation. Results confirm the model’s capacity to reconstruct detailed hand kinematics, supporting its potential for applications in hand function rehabilitation and human-machine interaction. Future work will aim to improve robustness under real-world conditions, including sensor noise and environmental variation.
Research on Station Centrality and Cascade Failure Invulnerability of Urban Rail Transit Networks
RUI Xiaobin, LIN Weihan, JI Jiaxin, WANG Zhixiao
Available online  , doi: 10.11999/JEIT250182
Abstract:
  Objective  Research on node centrality in rail transit networks is essential for ensuring operational safety. Identifying critical stations enables the development of preventive strategies and mitigates the effects of station failures. Existing studies highlight two key determinants of station importance: static topology and dynamic passenger flow. However, most current approaches treat these factors separately, leading to biased estimations of node importance. To address this limitation, this study proposes a novel node centrality measure that integrates static topology and dynamic passenger flow. The method combines topology-based centrality—derived from PageRank and a modified K-shell algorithm—with passenger centrality, which is based on station inflow and outflow volumes. A reinforcement mechanism ensures that passenger centrality consistently amplifies topology-based centrality, balancing the influence of both components. Using cascade failure simulations and real-world data from the Shanghai Metro, the proposed method reliably identifies key stations. These findings offer practical guidance for the design and maintenance of robust metro systems, enhancing their resilience to cascading failures and improving overall safety and stability.  Methods  The proposed method integrates static topology and dynamic passenger flow to evaluate the centrality of urban rail transit stations, addressing the limitations of existing approaches in identifying key stations. It consists of three components: static topology centrality, dynamic passenger flow centrality, and an integration strategy. (1) Static topology centrality is computed using a combination of PageRank and an improved K-core method. This hybrid approach captures both connectivity and node importance based on iterative removal order, mitigating the loss of resolution caused by the long-tail degree distribution typical in transit networks. (2) Dynamic passenger flow centrality assigns separate weights to inbound and outbound flows to account for congestion effects and directional asymmetry—factors often overlooked in previous models. The weights are derived from average boarding and alighting times and adjusted for flow variations across morning peak, evening peak, and off-peak periods. (3) Integration strategy: An exponential function combines the two centrality measures, ensuring that passenger flow consistently amplifies topology-derived importance. This design improves sensitivity to dynamic changes while preserving structural significance. The integrated centrality metric enhances network resilience by supporting targeted protection of critical stations, based on both static and dynamic characteristics.  Results and Discussions  This study investigates the vulnerability of Shanghai’s urban rail transit network by simulating cascading failures and identifying key stations. Using dynamic passenger flow data from three representative weekdays across 14 subway lines—comprising 289 stations and 335 edges—a load-capacity model is applied to assess node importance based on each station’s effect on network stability during cascading failure events. The results (Fig. 1, Fig. 2, Fig. 3) demonstrate that the proposed method consistently and effectively identifies key stations, outperforming five benchmark approaches. When assessing passenger flow loss, passenger flow centrality alone proves more informative than static topology centrality alone (Fig. 3). Moreover, the influence of passenger flow centrality is more pronounced during morning and evening peak periods, highlighting the role of temporal dynamics in station vulnerability. These findings highlight the importance of incorporating dynamic passenger flow data into vulnerability assessments to better capture real-world operational risks. The ablation study (Table 1) confirms that the integrated centrality—combining static topology and dynamic flow—offers superior performance over single-factor methods. Prioritizing the protection of stations identified by this approach can substantially improve the network’s resilience to cascading failures and enhance overall system safety.  Conclusions  This study investigates station centrality in urban rail transit networks and analyzes the cascading failure effects triggered by key node disruptions using a load-capacity model. The proposed method enhances the ranking of station importance by jointly capturing network structure and usage patterns, offering practical value for the design and maintenance of safe and resilient metro systems.Experiments on the Shanghai Metro network show that the method effectively identifies critical stations, with improvements observed in network average efficiency, connectivity, and reduced passenger flow loss. The results indicate the following: (1) Accurate identification of key stations requires the integration of both static topology and dynamic passenger flow; relying on either alone limits precision. (2) Failures at key stations can induce substantial cascading failures, highlighting the need to prioritize their protection to improve system resilience. (3) Future research should focus on developing more effective strategies for integrating static and dynamic centrality measures to extend applicability across different urban transit networks.
Breakthrough in Solving NP-Complete Problems Using Electronic Probe Computers
XU Jin, YU Le, YANG Huihui, JI Siyuan, ZHANG Yu, YANG Anqi, LI Quanyou, LI Haisheng, ZHU Enqiang, SHI Xiaolong, WU Pu, SHAO Zehui, LENG Huang, LIU Xiaoqing
Available online  , doi: 10.11999/JEIT250352
Abstract:
This study presents a breakthrough in addressing NP-complete problems using a newly developed Electronic Probe Computer (EPC60). The system employs a hybrid serial–parallel computational model and performs large-scale parallel operations through seven probe operators. In benchmark tests on 3-coloring problems in graphs with 2,000 vertices, EPC60 achieves 100% accuracy, outperforming the mainstream solver Gurobi, which succeeds in only 6% of cases. Computation time is reduced from 15 days to 54 seconds. The system demonstrates high scalability and offers a general-purpose solution for complex optimization problems in areas such as supply chain management, finance, and telecommunications.  Objective   NP-complete problems pose a fundamental challenge in computer science. As problem size increases, the required computational effort grows exponentially, making it infeasible for traditional electronic computers to provide timely solutions. Alternative computational models have been proposed, with biological approaches—particularly DNA computing—demonstrating notable theoretical advances. However, DNA computing systems continue to face major limitations in practical implementation.  Methods  Computational Model: EPC is based on a non-Turing computational model in which data are multidimensional and processed in parallel. Its database comprises four types of graphs, and the probe library includes seven operators, each designed for specific graph operations. By executing parallel probe operations, EPC efficiently addresses NP-complete problems.Structural Features:EPC consists of four subsystems: a conversion system, input system, computation system, and output system. The conversion system transforms the target problem into a graph coloring problem; the input system allocates tasks to the computation system; the computation system performs parallel operations via probe computation cards; and the output system maps the solution back to the original problem format.EPC60 features a three-tier hierarchical hardware architecture comprising a control layer, optical routing layer, and probe computation layer. The control layer manages data conversion, format transformation, and task scheduling. The optical routing layer supports high-throughput data transmission, while the probe computation layer conducts large-scale parallel operations using probe computation cards.  Results and Discussions  EPC60 successfully solved 100 instances of the 3-coloring problem for graphs with 2,000 vertices, achieving a 100% success rate. In comparison, the mainstream solver Gurobi succeeded in only 6% of cases. Additionally, EPC60 rapidly solved two 3-coloring problems for graphs with 1,500 and 2,000 vertices, which Gurobi failed to resolve after 15 days of continuous computation on a high-performance workstation.Using an open-source dataset, we identified 1,000 3-colorable graphs with 1,000 vertices and 100 3-colorable graphs with 2,000 vertices. These correspond to theoretical complexities of O(1.3289n) for both cases. The test results are summarized in Table 1.Currently, EPC60 can directly solve 3-coloring problems for graphs with up to n vertices, with theoretical complexity of at least O(1.3289n).On April 15, 2023, a scientific and technological achievement appraisal meeting organized by the Chinese Institute of Electronics was held at Beijing Technology and Business University. A panel of ten senior experts conducted a comprehensive technical evaluation and Q&A session. The committee reached the following unanimous conclusions:1. The probe computer represents an original breakthrough in computational models.2. The system architecture design demonstrates significant innovation.3. The technical complexity reaches internationally leading levels.4. It provides a novel approach to solving NP-complete problems.Experts at the appraisal meeting stated, “This is a major breakthrough in computational science achieved by our country, with not only theoretical value but also broad application prospects.” In cybersecurity, EPC60 has also demonstrated remarkable potential. Supported by the National Key R&D Program of China (2019YFA0706400), Professor Xu Jin’s team developed an automated binary vulnerability mining system based on a function call graph model. Evaluation of the system using the Modbus Slave software showed over 95% vulnerability coverage, far exceeding the 75 vulnerabilities detected by conventional depth-first search algorithms. The system also discovered a previously unknown flaw, the “Unauthorized Access Vulnerability in Changyuan Shenrui PRS-7910 Data Gateway” (CNVD-2020-31406), highlighting EPC60’s efficacy in cybersecurity applications.The high efficiency of EPC60 derives from its unique computational model and hardware architecture. Given that all NP-complete problems can be polynomially reduced to one another, EPC60 provides a general-purpose solution framework. It is therefore expected to be applicable in a wide range of domains, including supply chain management, financial services, telecommunications, energy, and manufacturing.  Conclusions   The successful development of EPC offers a novel approach to solving NP-complete problems. As technological capabilities continue to evolve, EPC is expected to demonstrate strong computational performance across a broader range of application domains. Its distinctive computational model and hardware architecture also provide important insights for the design of next-generation computing systems.
Research on an EEG-based Neurofeedback System for the Auxiliary Intervention of Post-Traumatic Stress Disorder
TAN Lize, DING Peng, WANG Fan, LI Na, GONG Anmin, NAN Wenya, LI Tianwen, ZHAO Lei, FU Yunfa
Available online  , doi: 10.11999/JEIT250093
Abstract:
  Objective  The ElectroEncephaloGram (EEG)-based Neurofeedback Regulation (ENR) system is designed for real-time modulation of dysregulated stress responses to reduce symptoms of Post-Traumatic Stress Disorder (PTSD) and anxiety. This study evaluates the system’s effectiveness and applicability using a series of neurofeedback paradigms tailored for both PTSD patients and healthy participants.  Methods  Employing real-time EEG monitoring and feedback, the ENR system targets the regulation of alpha wave activity, to alleviate mental health symptoms associated with dysregulated stress responses. The system integrates MATLAB and Unity3D to support a complete workflow for EEG data acquisition, processing, storage, and visual feedback. Experimental validation includes both PTSD patients and healthy participants to assess the system’s effects on neuroplasticity and emotional regulation. Primary assessment indices include changes in alpha wave dynamics and self-reported reductions in stress and anxiety.  Results and Discussions  Compared with conventional therapeutic methods, the ENR system shows significant potential in reducing symptoms of PTSD and anxiety. During functionality tests, the system effectively captures and regulates alpha wave activity, enabling real-time and efficient neurofeedback. Dynamic adjustment of feedback thresholds and task paradigms allows participants to improve stress responses and emotional states following training. Quantitative data indicate clear enhancements in EEG pattern modulation, while qualitative assessments reflect improvements in participants’ self-reported stress and anxiety levels.  Conclusion  This study presents an effective and practical EEG-based neurofeedback regulation system that proves applicable and beneficial for both individuals with PTSD and healthy participants. The successful implementation of the system provides a new technological approach for mental health interventions and supports ongoing personalized neuroregulation strategies. Future research should explore broader applications of the system across neurological conditions to fully assess its efficacy and scalability.
Personalized Federated Learning Method Based on Collation Game and Knowledge Distillation
SUN Yanhua, SHI Yahui, LI Meng, YANG Ruizhe, SI Pengbo
Available online  , doi: 10.11999/JEIT221203
Abstract:
To overcome the limitation of the Federated Learning (FL) when the data and model of each client are all heterogenous and improve the accuracy, a personalized Federated learning algorithm with Collation game and Knowledge distillation (pFedCK) is proposed. Firstly, each client uploads its soft-predict on public dataset and download the most correlative of the k soft-predict. Then, this method apply the shapley value from collation game to measure the multi-wise influences among clients and quantify their marginal contribution to others on personalized learning performance. Lastly, each client identify it’s optimal coalition and then distill the knowledge to local model and train on private dataset. The results show that compared with the state-of-the-art algorithm, this approach can achieve superior personalized accuracy and can improve by about 10%.
The Range-angle Estimation of Target Based on Time-invariant and Spot Beam Optimization
Wei CHU, Yunqing LIU, Wenyug LIU, Xiaolong LI
Available online  , doi: 10.11999/JEIT210265
Abstract:
The application of Frequency Diverse Array and Multiple Input Multiple Output (FDA-MIMO) radar to achieve range-angle estimation of target has attracted more and more attention. The FDA can simultaneously obtain the degree of freedom of transmitting beam pattern in angle and range. However, its performance is degraded due to the periodicity and time-varying of the beam pattern. Therefore, an improved Estimating Signal Parameter via Rotational Invariance Techniques (ESPRIT) algorithm to estimate the target’s parameters based on a new waveform synthesis model of the Time Modulation and Range Compensation FDA-MIMO (TMRC-FDA-MIMO) radar is proposed. Finally, the proposed method is compared with identical frequency increment FDA-MIMO radar system, logarithmically increased frequency offset FDA-MIMO radar system and MUltiple SIgnal Classification (MUSIC) algorithm through the Cramer Rao lower bound and root mean square error of range and angle estimation, and the excellent performance of the proposed method is verified.
Circuit and System Design
A Battery Internal-Short-Circuit Fault Diagnosis Method Combining Battery Phase Plane with Conformer-BiGRU Network
MAO Lin, ZHANG Haixin, HE Zhiwei, GAO Mingyu, DONG Zhekang
Available online  , doi: 10.11999/JEIT250313
Abstract:
  Objective  New Energy Vehicles (NEVs) have gained rapid popularity in recent years due to their environmental benefits and high efficiency. However, as their market share continues to grow, concerns regarding frequent malfunctions and safety risks have also increased. Among these issues, Internal Short Circuit (ISC) faults are particularly concerning due to their strong concealment and the potential for severe consequences. Without accurate diagnosis and timely intervention, ISC faults can result in serious safety incidents. Therefore, developing efficient and reliable diagnostic methods for ISC faults is of practical significance.  Methods  A novel ISC fault diagnosis method is proposed for battery packs by combining an improved battery phase plane approach with a Conformer-BiGRU network. First, the improved battery phase plane method is employed to extract two-dimensional features from voltage sequences, providing deeper spatial and structural information. Second, a Conformer-BiGRU network is employed to learn features from the voltage data. The network integrates a CNN branch for local feature extraction and a Transformer branch for global representation. A feature coupling unit fuses the outputs of both branches, which are then passed to a BiGRU module to classify individual cells within the battery pack and detect ISC faults.  Results and Discussions  The proposed method is evaluated using fault data collected from an experimental platform. The results demonstrate that the improved battery phase plane effectively distinguishes between normal and faulty batteries within a two-dimensional plane (Figure 6) and further confirm its capability to detect ISC faults with varying severity under different data volumes (Figure 9). Using the Conformer-BiGRU network for fault diagnosis, the method achieves classification accuracies of 94.30%, 92.77%, and 94.85% under FUDS, UDDS, and US06 operating conditions, respectively (Table 3), significantly exceeding the performance of comparative models. Additionally, the feature extraction module contributes to an overall performance improvement of approximately 2.04% (Table 5). These findings indicate that the proposed method exhibits strong robustness (Table 6 and Figure 11) and offers a promising approach for enhancing the safety of NEVs.  Conclusions  This study proposes a novel method for diagnosing ISC faults in battery packs by integrating an improved battery phase plane approach with a Conformer-BiGRU network. The main contributions are as follows: First, the improved battery phase plane method enhances the separability of different fault states in two-dimensional space by incorporating both voltage and its first-order differential distribution, addressing the limitations of conventional one-dimensional feature extraction. Second, a hybrid Conformer-BiGRU architecture is developed, in which the Conformer module captures local discharge characteristics, while the BiGRU module models temporal dependencies. These features are integrated through a feature coupling unit to achieve cross-level feature fusion. Third, an experimental ISC fault dataset with varying severity levels is established using a self-built testing platform. Experimental results demonstrate average diagnostic accuracy, recall, and F1-scores of 91.26%, 85.17%, and 88.09%, respectively, across three international driving cycles. Although laboratory testing verifies the effectiveness of the proposed method, real-world application requires targeted optimization. This includes adapting BiGRU parameters during the migration of the Improved Battery Phase Plane (IBPP) module and refining the Conformer’s local perception weights through transfer learning to enhance feature decoupling. Future research focuses on improving diagnostic performance under concurrent fault scenarios to enhance engineering robustness in complex operating conditions.
A P-band Dual-polarized Ultra-thin Absorptive-transmissive Electromagnetic Surface Using Frequency Selective Surface
SUN Daifei, YANG Huanhuan, LI Tong, LIAO Jiawei, WU Tianhao, ZOU Jing, YANG Qi, CAO Xiangyu
Available online  , doi: 10.11999/JEIT250309
Abstract:
  Objective  Frequency Selective Surfaces (FSS), as artificial ElectroMagnetic (EM) periodic structures, regulate the transmission and reflection of EM waves. Radomes integrating FSS can protect antennas, preserve the aerodynamic profile of radio-frequency systems, shape spatial scattering field distributions, and suppress backward Radar Cross Section (RCS). However, when illuminated by multiple radars, such radomes often fail to maintain low detectability due to their inability to achieve bistatic low RCS. Recent efforts have focused on developing absorptive structures based on FSS, where active FSS-based absorbers offer adaptive tunability across frequency and time domains. Nonetheless, achieving absorption in the P-band remains challenging due to the inherent limitations of existing dielectric materials. While FSS bandpass properties are frequently employed in radomes and the tunability of active FSS supports the design of reconfigurable absorbers, the two functionalities have largely been pursued independently, resulting in limited multifunctional surface designs. This study proposes a P-band ultra-thin absorber using FSS composed of cascaded unit cells with gradually curved meander lines. By exploiting the distinct equivalent circuit characteristics of absorbing and transmitting FSS structures, an integrated system is developed that enables both EM wave transmission and tunable wideband absorption in the P-band.  Methods  This paper proposes a novel design method for a dual-polarized, ultra-thin absorptive-transmissive EM surface element operating in the P-band, based on the FSS technique. The method uses cascaded elements with a gradient-bending structure to increase the effective current path length and incorporates lumped components to achieve wideband tunable absorption at low frequencies. By analyzing the equivalent circuit characteristics of both absorptive and transmissive FSS-based elements, an integrated absorptive-transmissive structure is developed. The difference in their equivalent circuits effectively suppresses mutual coupling, enabling the relatively independent design of absorptive and transmissive functions. To demonstrate this approach, a dual-polarized ultra-thin EM surface element is designed that simultaneously exhibits high transmittance and tunable wideband absorptivity. The step-by-step design process is presented, and the operating mechanism of the proposed element is thoroughly analyzed.  Results and Discussions  Both simulation and experimental results confirm that the proposed integrated element achieves dual-polarized absorptive performance in the P-band and dual-polarized transmissive performance in the C-band. The element features an ultra-thin profile, requires few lumped components, and exhibits a broad operational bandwidth. Notably, the proposed method combines equivalent circuit modeling with field-based analysis to facilitate the design of multifunctional EM surfaces, thereby streamlining the integration of absorptive and transmissive functionalities within a single structure.  Conclusions  Structural absorbers face a fundamental trade-off between achieving efficient low-frequency absorption and maintaining a low profile or lightweight design, making the development of P-band absorbing surfaces particularly challenging. FSS, commonly used in radomes, also offer potential for tunable absorber design. This study integrates both functionalities to develop a multifunctional EM surface capable of simultaneous wave absorption and transmission, based on an FSS architecture. When irradiated from one direction, the surface achieves absorptivity above 0.9 across 0.34~1.1 GHz. When irradiated from the opposite direction, it exhibits transmittivity exceeding 0.8 over 4.26~4.48 GHz. These two functions operate with relative independence. The proposed structure features a wide operational bandwidth, ultra-thin profile, and minimal reliance on electronic components. The method enables not only effective P-band absorption but also the integrated design of multifunctional EM surfaces. It offers strong scalability and holds significant potential for future applications.
Dynamic Distribution Adaptation with Higher-order Moment Matching for Battery Pack Multi-fault Diagnosis
GUAN Siwei, HE Zhiwei, DONG Zhekang, TONG Hongtao, MA Shenhui, GAO Mingyu
Available online  , doi: 10.11999/JEIT250226
Abstract:
  Objective  Electric vehicle battery pack fault diagnosis is challenged by diverse operating conditions, the scarcity of fault data, and the domain shift caused by the non-Gaussian distribution of battery features. Conventional fault diagnosis methods struggle to address multiple fault types, lack the capability for fault isolation, and fail to account for distribution shifts between training and test data. Domain adaptation approaches enable robust multi-fault diagnosis across operating conditions without relying on accurate cell models or abundant labeled data. However, current methods remain limited. (1) They typically assume that aligning global and fine-grained subdomain distributions is equally important, which may not hold in practice. (2) Knowledge transfer cannot be fully achieved by aligning only low-order statistical features; higher-order statistical features are needed to capture the non-Gaussian characteristics of battery discharge profiles. To address these issues, a method is proposed in which global domains and subdomains are dynamically aligned while higher-order statistical moments are extracted to represent complex non-Gaussian distributions, thereby achieving fine-grained domain alignment and effective knowledge transfer.  Methods  This study proposes a dynamic distribution adaptation method with higher-order moment matching for multi-fault diagnosis of battery packs. The approach consists of three components: (1) Dynamic distribution adaptation. A feature extractor based on a one-dimensional convolutional network with residual connectivity and a multilayer perceptron classifier is constructed. The global distributions of source and target domains are aligned using Maximum Mean Discrepancy (MMD), while subdomain distributions of similar faults are aligned using Local Maximum Mean Discrepancy (LMMD). A dynamic factor is introduced to automatically adjust the relative weights of global and local alignment according to the inter-domain discrepancy, thereby adapting to distribution shifts under different operating conditions. (2) Higher-order moment matching. To address the non-Gaussian characteristics of battery data, higher-order statistical moment matching is incorporated into MMD. Computational complexity in high-dimensional tensors is reduced by random sampling, which enables fine-grained domain alignment across multi-order statistics and enhances the transferability of non-Gaussian distribution features. (3) Multi-fault diagnosis with domain adaptation. Experimental data from three standard vehicle operating conditions are used to jointly optimize classification loss and domain adaptation loss. This enables the diagnosis of multiple faults, including internal short circuit, sensor drift/noise, and battery inconsistency, across operating conditions while reducing reliance on manual annotation. By dynamically integrating global and local feature alignment, the method improves generalization performance under complex operating conditions and non-Gaussian distribution scenarios.  Results and Discussions  Systematic experiments validate the superiority of the proposed dynamic distribution adaptation with higher-order moment matching for multi-fault diagnosis in electric vehicle battery packs. As shown in Table 3, the results from six transfer tasks under three operating conditions demonstrate that the proposed method achieves an average F1 score of 94.9%, which is 13.3% higher than that of the best-performing baseline model (DSAN). The confusion matrix in Fig. 6 indicates that the method achieves the lowest misclassification rate in distinguishing similar faults. Feature visualization results (Fig. 7) show that the method effectively reduces cross-domain feature distances of similar faults by dynamically adjusting the weights of global and local distribution alignment. Moreover, it successfully captures non-Gaussian discharge characteristics through higher-order moment matching, thereby achieving fine-grained domain adaptation. In terms of efficiency, the proposed method attains an average diagnosis time of 0.404 3 seconds (Table 4), satisfying real-time on-board application requirements. Nonetheless, optimization of computational resource consumption remains necessary for deployment on edge devices. Importantly, the method does not require labeled data from the target domain and overcomes the generalization bottleneck of traditional methods under domain shift and non-Gaussian conditions. However, some cross-domain features (Fig. 7) are not completely overlapped, and lightweight model design is still required for practical implementation on edge devices.  Conclusions  The battery pack is recognized as a critical component of electric vehicles, and reliable multi-fault diagnosis is regarded as essential for safe operation. Considering the unknown and diverse nature of real operating conditions, fault diagnosis is investigated across three driving cycles: UDDS, FUDS, and US06. A dynamic distribution adaptation with higher-order moment matching (DDAMD) is proposed for diagnosing multiple faults in series-connected battery packs. The method dynamically evaluates the relative importance of conditional and marginal distributions to align source and target domains, while non-Gaussian features from charge-discharge curves are effectively extracted for fine-grained alignment. Experimental results across six transfer tasks confirm that DDAMD achieves the highest diagnostic performance. Detailed analyses present diagnostic accuracy for each fault type as well as the diagnostic speed, while feature visualization further improves interpretability by demonstrating how the algorithm extracts domain-invariant and discriminative fault features across domains. Future research will extend this work in two directions: (1) incorporating additional operating conditions and a broader set of fault categories, and (2) exploring transfer tasks from simulation to real-world applications to facilitate data acquisition and labeling.
Image and Intelligent Information Processing
Weakly Supervised Image Semantic Segmentation Based on Multi-Seeded Information Aggregation and Positive-Negative Hybrid Learning
SANG Yu, LIU Tong, MA Tianjiao, LI Le, LI Siman, LIU Yunan
Available online  , doi: 10.11999/JEIT250112
Abstract:
  Objective   The rapid development of deep learning techniques, particularly Convolutional Neural Networks (CNN), has led to notable advances in semantic segmentation, enabling applications in medical imaging, autonomous driving, and remote sensing. However, conventional semantic segmentation tasks typically rely on large numbers of pixel-level annotated images, which is both time-consuming and expensive. To address this limitation, Weakly Supervised Semantic Segmentation (WSSS) using image-level labels has emerged as a promising alternative. This approach aims to reduce annotation costs while maintaining or enhancing segmentation performance, thus supporting broader adoption of semantic segmentation techniques. Most existing methods focus on optimizing Class Activation Mapping (CAM) to generate high-quality seed regions, with further refinement through post-processing. However, the resulting seed labels often contain varying degrees of noise. To mitigate the effect of noisy labels on the segmentation network and to efficiently extract accurate information by leveraging multiple complementary seed sources, this study proposes a weakly supervised semantic segmentation method based on multi-seed information aggregation and positive-negative hybrid learning. The proposed approach improves segmentation performance by integrating complementary information from different seeds while reducing noise interference.  Methods   Building on the idea that combining multiple seeds can effectively extract accurate information, this study proposes a weakly supervised image semantic segmentation method based on multi-seed information aggregation and positive-negative hybrid learning. The approach employs a generalized classification network to generate diverse seed regions by varying the input image scale and modifying the Dropout layer to randomly deactivate neurons with different probabilities. This process enables the extraction of complementary information from multiple sources. Subsequently, a semantic segmentation network is trained using a hybrid positive-negative learning strategy based on the category labels assigned to each pixel across these seeds. Clean labels, identified with high confidence, guide the segmentation network through a positive learning process, where the model learns that “the input image belongs to its assigned labels.” Conversely, noisy labels are addressed using two complementary strategies. Labels determined as incorrect are trained under the principle that “the input image does not belong to its assigned labels,” representing a form of positive learning for error suppression. Additionally, an indirect negative learning strategy is applied, whereby the network learns that “the input image does not belong to its complementary labels,”further mitigating the influence of noisy labels. To reduce the adverse effects of noisy labels, particularly the tendency of conventional cross-entropy loss to assign higher prediction confidence to such labels, a prediction constraint loss is introduced. This loss function enhances the model’s predictive accuracy for reliable labels while reducing overfitting to incorrect labels. The overall framework effectively suppresses noise interference and improves the segmentation network’s performance.  Results and Discussions   The proposed weakly supervised image semantic segmentation method based on multi-seed information aggregation and positive-negative hybrid learning generates diverse seeds by randomly varying the Dropout probability and input image scale, with Conditional Random Field (CRF) optimization applied to further refine seed quality. To limit noise introduction while maintaining the effectiveness of positive-negative hybrid learning, six complementary seeds are selected (Table 5). The integration of multi-source information from these seeds enhances segmentation performance, as demonstrated in (Table 7) . Pixel labels within these seeds are classified as clean or noisy based on a defined confidence threshold. The segmentation network is subsequently trained using a positive-negative hybrid learning strategy, which suppresses the influence of noisy labels and improves segmentation accuracy. Experimental results confirm that positive-negative hybrid learning effectively reduces label noise and enhances segmentation performance (Table 8). The proposed method was validated on the PASCAL VOC 2012 and MS COCO 2014 datasets. With a CNN-based segmentation network, the mean Intersection over Union (mIoU) reached 72.5% and 40.8%, respectively. When using a Transformer-based segmentation network, the mIoU improved to 76.8% and 46.7% (Table 1, Table 3). These results demonstrate that the proposed method effectively enhances segmentation accuracy while controlling the influence of noisy labels.  Conclusions   This study addresses the challenge of inaccurate seed labels in WSSS based on image-level annotations by proposing a multi-seed label differentiation strategy that leverages complementary information to improve seed quality. In addition, a positive-negative hybrid learning approach is introduced to enhance segmentation performance and mitigate the influence of erroneous pixel labels on the segmentation model. The proposed method achieves competitive results on the PASCAL VOC 2012 and MS COCO 2014 datasets. Specifically, the mIoU reaches 72.5% and 40.8%, respectively, using a CNN-based segmentation network. With a Transformer-based segmentation network, the mIoU further improves to 76.8% and 46.7%. These results demonstrate the effectiveness of the proposed method in improving segmentation accuracy while reducing noise interference. Although the method does not yet achieve ideal label precision, label differentiation combined with positive-negative hybrid learning effectively suppresses misinformation propagation and outperforms approaches based on single-seed generation and positive learning alone.
SR-FDN: A Frequency-Domain Diffusion Network for Image Detail Restoration in Super-Resolution
LI Xiumei, DING Linlin, SUN Junmei, BAI Huang
Available online  , doi: 10.11999/JEIT250224
Abstract:
  Objective  Image Super-Resolution (SR) is a critical computer vision task aimed at reconstructing High-Resolution (HR) images from Low-Resolution (LR) inputs, with broad applications in fields such as medical imaging and satellite imaging. Recently, diffusion-based SR methods have attracted significant attention due to their generative capability and strong performance in restoring fine image details. Existing diffusion model-based SR approaches have demonstrated potential in recovering textures and structures, with some methods focusing on spatial domain features and others utilizing frequency domain information. Spatial domain features aid in reconstructing overall structural information, whereas frequency domain decomposition separates images into amplitude and phase components across frequencies. High-frequency components capture details, textures, and edges, whereas low-frequency components describe smooth structures. Compared to purely spatial modeling, frequency domain features improve the aggregation of dispersed high-frequency information, enhancing the representation of image textures and details. However, current frequency domain SR methods still show limitations in restoring high-frequency details, with blurring or distortion persisting in some scenarios. To address these challenges, this study proposes SR-FDN, an SR reconstruction network based on a frequency-domain diffusion model.  Methods  SR-FDN leverages the distribution mapping capability of diffusion models to improve image reconstruction. The proposed network integrates spatial and frequency domain features to enhance high-frequency detail restoration. Two constraints guide the model design: (1) The network must generate plausible HR images conditioned solely on LR inputs, which serve as the primary source of structural information, ensuring high-fidelity reconstruction. (2) The model should balance structural reconstruction with enhanced detail restoration. To achieve this, a dual-branch frequency domain attention mechanism is introduced. A portion of the features undergoes Fourier transform for frequency domain processing, where high-frequency information is emphasized through self-attention. The remaining features adjust frequency domain weights before being combined with spatial domain representations. Skip connections in the U-Net architecture preserve LR structural information while enhancing frequency domain details, improving both structural and textural reconstruction. Wavelet downsampling replaces conventional convolutional downsampling within the U-Net noise predictor, reducing spatial resolution while retaining more detailed information. In addition, a Fourier frequency domain loss function constrains amplitude and phase components of the reconstructed image, further enhancing high-frequency detail recovery. To guide the generative process, additional image priors are incorporated, enabling the diffusion model to restore textures consistent with semantic category features.  Results and Discussions  The results of SR-FDN on face datasets and general datasets for 4× and 8× SR (Table 1, Table 2, Table 3) demonstrate that the proposed method achieves strong performance across objective evaluation metrics. These results indicate that SR-FDN can effectively restore image detail information while better preserving structural and textural features. A comparison of iteration counts between SR-FDN and other diffusion-based methods (Fig. 2) shows that SR-FDN can reconstruct higher-quality images with fewer iterations. Despite the reduced number of iterations, SR-FDN maintains high-fidelity reconstruction, reflecting its ability to lower computational overhead without compromising image quality. To further verify the effectiveness of the proposed SR-FDN, visual comparisons on the FFHQ dataset (Fig. 3) and the DIV2K dataset (Fig. 4) are presented. The results show that SR-FDN offers clearer and more detailed image reconstruction, particularly in high-frequency regions such as facial features and hair textures. Ablation experiments (Table 5) and feature visualization results (Fig. 5) are also provided. These results confirm that the proposed dual-branch frequency domain design and the Fourier domain loss function significantly contribute to improved restoration of fine details.  Conclusions  This study proposes SR-FDN, a diffusion-based SR reconstruction model that integrates frequency domain information to enhance detail restoration. The SR-FDN model incorporates a dual-branch frequency domain attention mechanism, which adaptively reinforces high-frequency components, effectively addressing the limitations of conventional methods in recovering edge structures and texture details. In addition, SR-FDN employs wavelet downsampling to preserve informative features while reducing spatial resolution, and introduces a frequency domain loss function that constrains amplitude and phase information, enabling more effective fusion of frequency and spatial domain features. This design substantially enhances the model’s ability to recover high-frequency details. Extensive experiments on benchmark datasets demonstrate that SR-FDN reconstructs images with superior quality and richer details, exhibiting clear advantages in both qualitative and quantitative evaluations.
Low-Light Object Detection via Joint Image Enhancement and Feature Adaptation
QIAO Chengping, JIN Jiakun, ZHANG Junchao, ZHU Zhengliang, CAO Xiangxu
Available online  , doi: 10.11999/JEIT250302
Abstract:
  Objective  Object detection has advanced significantly under normal lighting conditions, supported by numerous high-accuracy, high-speed deep learning algorithms. However, in low-light environments, images exhibit reduced brightness, weak contrast, and severe noise interference, leading to blurred object edges and loss of color information, which substantially degrades detection accuracy. To address this challenge, this study proposes an end-to-end low-light object detection algorithm that balances detection accuracy with real-time performance. Specifically, an end-to-end network is designed to enhance feature quality and improve detection accuracy in real time under low-light conditions.  Methods  To improve object detection performance under low-light conditions while maintaining detection accuracy and real-time processing, this study proposes an end-to-end low-light image object detection method. Detection accuracy is enhanced through joint learning of image enhancement and feature adaptation, with the overall network structure shown in Fig. 1. First, a data augmentation module synthesizes low-light images from normal-light images. The paired normal-light and low-light images are mixed using the MixUp function provided by YOLOv5 to generate the final low-light images. These synthesized images are input into the low-light image enhancement module. In parallel, the matched normal-light images are provided as supervision to train the image enhancement network. Subsequently, both the enhanced low-light images and the corresponding normal-light images are fed into the object detection module. After processing through the YOLOv5 backbone, a matching loss is computed to guide feature adaptation.  Result and Discussions   The experiments are conducted primarily on the Polar3000 and LLVIP datasets. Fig. 3 presents the detection results obtained using YOLOv5 with different image enhancement methods applied to the Polar3000 dataset. Most existing methods tend to misclassify overexposed regions as bright Unmanned Aerial Vehicles (UAVs). In contrast, the proposed method demonstrates accurate object detection in low-light conditions without misidentifying overexposed areas as UAVs (Fig. 3). Furthermore, the detection performance of the proposed method, termed MAET, is compared with that of a standalone YOLOv5 model. Quantitative experiments show that the proposed method outperforms both image-enhancement-first detection pipelines and recent low-light object detection methods across both experimental groups A and B, irrespective of low-light fine-tuning. On the LLVIP dataset, the proposed method achieves a detection accuracy of 91.7% (Table 1), while on the Polar3000 dataset, it achieves 92.3% (Table 2). The model also demonstrates superior generalization performance on the ExDark and DarkFace datasets (Tables 4 and 3). Additionally, compared to the baseline YOLOv5 model, the proposed method increases parameter size by only 2.5% while maintaining real-time detection speed (Table 5).  Conclusions  This study proposes a low-light object detection method based on joint learning of image enhancement and feature adaptation. The approach simultaneously optimizes image enhancement loss, feature matching loss, and object detection loss within an end-to-end framework. It improves image illumination, preserves fine details, and aligns the features of enhanced images with those acquired under normal lighting conditions, enabling high-precision object detection in low-light environments. Comparative experiments on the LLVIP and Polar3000 datasets demonstrate that the proposed method achieves improved detection accuracy while maintaining real-time performance. Furthermore, the method achieves the best generalization results on the ExDark and DarkFace datasets. Future work will explore low-light object detection based on multimodal data fusion of visible and infrared images to further enhance detection performance in extremely dark conditions.
Dynamic Adaptive Partitioning of Deep Neural Networks Based on Early Exit Mechanism under Edge-End Collaboration
DING Nan, WANG Jiajia, JI Chenghui, HU Chuangye, XU Li
Available online  , doi: 10.11999/JEIT250291
Abstract:
  Objective  The deployment of Deep Neural Networks (DNNs) for inference tasks in industrial intelligence applications is constrained by the complexity of Directed Acyclic Graph (DAG) structures and dynamic resource limitations, making it challenging to simultaneously optimize both latency and accuracy. Existing methods are generally restricted to chain-structured DNNs and lack adaptive mechanisms to accommodate network variability and heterogeneous computational resources. To address these limitations, this paper proposes a Dynamic Adaptive Partitioning framework based on a Deep Early Exit mechanism (DAPDEE), designed to achieve low-latency, high-accuracy inference through edge-end collaborative computing. The significance of this work lies in its potential to provide a generalizable solution suitable for diverse network conditions and computing environments.  Methods  The proposed DAPDEE framework incorporates several technical innovations: First, it abstracts both chain and complex DNN architectures into a unified DAG representation, establishing a general topological foundation for partition optimization. Second, it employs offline optimization of early exit classifiers using a multi-task learning approach, requiring only the loading of pre-trained model parameters during deployment. Combined with real-time indicators, such as network bandwidth and terminal computational load, this enables dynamic selection of optimal exit points and partitioning strategies. Finally, an inverse search mechanism is applied to jointly optimize latency and accuracy, aiming to minimize single-frame end-to-end delay under light workloads and to maximize system throughput under heavy workloads. Through these strategies, the framework enables efficient inference suitable for time-sensitive scenarios, including smart manufacturing and autonomous driving.  Results and Discussions  Experimental results demonstrate that the DAPDEE framework substantially improves performance compared to conventional Device-Only methods under varying network conditions. Specifically, under CAT1, 3G, and 4G networks, DAPDEE achieves latency reductions of up to 7.7% under heavy loads and 7.5% under light loads, with throughput improvements reaching up to 9.9 times. Notably, the accuracy loss remains consistently below 1.2% (Fig. 6, Fig. 7), confirming the framework’s ability to maintain reliable inference performance. These results verify the effectiveness of DAPDEE in adapting to dynamic network environments and heterogeneous computational loads. For instance, when the bandwidth is fixed at 1.1 Mbps (3G), the optimal partition strategy adjusts in response to varying latency constraints, revealing a positive correlation between relaxed latency requirements and deeper exit points (Fig. 6). Conversely, with fixed latency constraints and increasing bandwidth, the partition point progressively shifts toward the terminal device, reflecting enhanced resource utilization on the end side (Fig. 7). Furthermore, practical deployments on a PC and a Raspberry Pi-based intelligent vehicle validate the theoretical performance gains, as demonstrated by the applied partitioning strategies (Algorithm 1, Algorithm 2).  Conclusions  In summary, the proposed DAPDEE framework effectively addresses the challenge of balancing inference efficiency and accuracy in edge-end collaborative scenarios involving complex DAG-structured DNNs. By integrating early exit mechanisms with dynamic partitioning strategies and multidimensional load evaluation, DAPDEE exhibits strong adaptability and robustness under diverse network conditions and resource constraints. These findings advance the current state of DNN partitioning methodologies and offer practical insights for optimizing cloud-edge-terminal architectures and reinforcement learning-based adaptive mechanisms. Nonetheless, areas for further improvement remain. These include incorporating multi-task concurrency, refining the energy consumption model, and enhancing real-time partitioning efficiency for complex DAG topologies. Future research will focus on extending the framework to support multi-task collaborative optimization and reducing the computational complexity of online partitioning algorithms for DAG-structured DNNs.
A Lightweight Semantic Visual Simultaneous Localization and Mapping Framework for Inspection Robots in Dynamic Environments
YU Haoyang, LI Yansheng, XIAO Lingli, ZHOU Jiyuan
Available online  , doi: 10.11999/JEIT250301
Abstract:
  Objective  In complex dynamic environments such as industrial parks and urban roads, inspection robots depend heavily on visual Simultaneous Localization And Mapping (SLAM) systems. However, the presence of moving objects often causes feature drift and map degradation, reducing SLAM performance. Furthermore, conventional semantic segmentation models typically require extensive computational resources, rendering them unsuitable for embedded platforms with limited processing capabilities, thereby constraining SLAM deployment in autonomous inspection tasks. To address these challenges, this study proposes a lightweight semantic visual SLAM framework designed for inspection robots operating in dynamic environments. The framework incorporates a semantic segmentation-based dynamic feature rejection method to achieve real-time identification of dynamic regions at low computational cost, thereby improving SLAM robustness and mapping accuracy.  Methods  Building upon the 11th generation lightweight YOLO segmentation model (YOLOv11n-seg), a systematic lightweight redesign is implemented. to enhance performance under constrained computational resources. First, the original neck is replaced with DyCANet, a lightweight multi-scale feature fusion module that integrates dynamic point sampling and channel attention to improve semantic representation and boundary segmentation. DyCANet combines DySample, a dynamic upsampling operator that performs content-aware spatial sampling with minimal overhead, and ChannelAttention_HSFPN, a hierarchical attention structure that strengthens multi-scale integration and highlights critical semantic cues, particularly for small or occluded objects in complex scenes. Second, a Dynamic Convolution module (DynamicConv) is embedded into all C3k2 modules to enhance the adaptability and efficiency of feature extraction. Inspired by the Mixture-of-Experts framework, DynamicConv applies a conditional computation mechanism that dynamically adjusts kernel weights based on the input feature characteristics. This design allows the network to extract features more effectively across varying object scales and motion patterns, improving robustness against dynamic disturbances with low computational cost. Third, the original segmentation head is replaced by the Reused and Shared Convolutional Segmentation Head (RSCS Head), which enables decoder structure sharing across multi-scale branches. RSCS reduces redundant computation by reusing convolutional layers and optimizing feature decoding paths, further improving overall model efficiency while maintaining segmentation accuracy. These architectural modifications result in DHSR-YOLOSeg, a lightweight semantic segmentation model that significantly reduces parameter count and computational cost while preserving performance. DHSR-YOLOSeg is integrated into the tracking thread of ORB-SLAM3. to provide real-time semantic information. This enables dynamic object detection and the removal of unstable feature points during localization, thereby enhancing the robustness and trajectory consistency of SLAM in complex dynamic environments.  Results and Discussions  Ablation experiments on the COCO dataset demonstrate that, compared with the baseline YOLOv11n-seg, the proposed DHSR-YOLOSeg achieves a 13.8% reduction in parameter count, a 23.1% decrease in Giga Floating Point Operations (GFLOPs), and an approximate 2% increase in mean Average Precision at IoU 0.5 (mAP50) (Table 1). On the KITTI dataset, DHSR-YOLOSeg reaches an inference speed of 60.19 frame/s, which is 2.14% faster than YOLOv11n-seg and 275% faster than the widely used Mask R-CNN (Table 2). For trajectory accuracy evaluation on KITTI sequences 00~10, DHSR-YOLOSeg outperforms ORB-SLAM3 in 8 out of 10 sequences, achieving a maximum Root Mean Square Error (RMSE) reduction of 16.76% and an average reduction of 8.78% (Table 3). Compared with DynaSLAM and DS-SLAM, the proposed framework exhibits more consistent error suppression across sequences, improving both trajectory accuracy and stability. In terms of runtime efficiency, DHSR-YOLOSeg achieves an average per-frame processing time of 48.86 ms on the KITTI dataset, 18.44% and 41.38% lower than DS-SLAM and DynaSLAM, respectively (Table 4). The per-sequence processing time ranges from 41 to 55 ms, which is comparable to the 35.64 ms of ORB-SLAM3, indicating that the integration of semantic segmentation introduces only a modest computational overhead.  Conclusions  This study addresses the challenge of achieving robust localization for inspection robots operating in dynamic environments, particularly in urban road settings characterized by frequent interference from pedestrians, vehicles, and other moving objects. To this end, a semantic-enhanced visual SLAM framework is proposed, in which a lightweight semantic segmentation model, DHSR-YOLOSeg, is integrated into the stereo-based ORB-SLAM3 pipeline. This integration enables real-time identification of dynamic objects and removal of their associated feature points, thereby improving localization robustness and trajectory consistency. The DHSR-YOLOSeg model incorporates three key architectural components—DyCANet for feature fusion, DynamicConv for adaptive convolution, and the RSCS Head for efficient multi-scale decoding. Together, these components reduce the model’s size and computational cost while preserving segmentation performance, providing an efficient and deployable perception solution for resource-constrained platforms. Experimental findings show that: (1) ablation tests on the COCO dataset confirm substantial reductions in complexity with preserved accuracy, supporting embedded deployment; (2) frame rate comparisons on the KITTI dataset demonstrate superior performance over both lightweight and standard semantic segmentation methods, meeting real-time SLAM requirements; (3) trajectory evaluations indicate that the dynamic feature rejection strategy effectively mitigates localization errors in dynamic scenes; and (4) the overall system maintains high runtime efficiency, ensuring a practical balance between semantic segmentation and real-time localization performance. However, current experiments are conducted under ideal conditions using standardized datasets, without fully reflecting real-world challenges such as multi-sensor interference or unstructured environments. Moreover, the trade-offs between model complexity and accuracy for each lightweight module have not been systematically assessed. Future work will focus on multimodal sensor fusion and adaptive dynamic perception strategies to enhance the robustness and applicability of the proposed system in real-world autonomous inspection scenarios.
Hyperspectral Image Denoising Algorithm via Joint Low-Rank Tensor Decomposition and Product Graph Modeling
MA Mou, CAI Mingjiao, SHEN Yu, ZHOU Fang, JIANG Junzheng
Available online  , doi: 10.11999/JEIT250130
Abstract:
Various types of noise in HyperSpectral Images (HSI) can severely degrade subsequent analysis. To address this problem, this study proposes a denoising framework based on Low-Rank Tensor Decomposition and Kronecker product Graph Laplacian Regularization (LRTDKGLR). Spatial and spectral graphs are constructed using Graph Signal Processing (GSP) and fused into a joint spatial-spectral product graph through the Kronecker product. Tucker decomposition is applied to this product graph to extract a unified low-dimensional representation, capturing the global spatial and spectral structures. A Kronecker Graph Laplacian Regularization (KGLR) term is introduced to enforce piecewise smoothness across both spatial and spectral dimensions, enhancing inter-band coherence. The denoising problem is formulated as an optimization task that integrates low-rank decomposition and graph-based regularization, solved efficiently using an Augmented Lagrangian Multiplier (ALM) approach. Experimental results on simulated and real HSI datasets demonstrate that LRTDKGLR achieves superior performance in edge preservation and noise suppression compared with existing HSI denoising methods.  Objective  Traditional denoising approaches often fail to simultaneously preserve spatial details and spectral fidelity due to the high dimensionality and complex spectral characteristics of HSI data. This study aims to develop a method that effectively suppresses mixed noise, such as Gaussian and stripe noise, while preserving critical spatial features and edge structures. The proposed approach leverages LRTDKGLR to enhance inter-dimensional correlations between spatial and spectral domains, ensuring improved noise reduction and detail preservation. Additionally, this research investigates an efficient ALM-based optimization strategy to address the HSI denoising problem, providing a more robust solution for noise reduction in high-dimensional, contaminated environments.  Methods  The proposed LRTDKGLR algorithm performs HSI denoising by effectively modeling spatial-spectral correlations. Separate spatial and spectral graphs are constructed and integrated through the Kronecker product to form a unified product graph representation. This joint graph considers both pixel intensity and the values of neighboring pixels across spatial and spectral dimensions, enabling enhanced denoising through spatial–spectral correlation modeling. Tucker decomposition is applied to extract a low-rank representation that captures global spatial-spectral relationships within the data. This decomposition facilitates effective modeling of the inherent structure of HSI while supporting the denoising process. The KGLR term models smoothness across both spatial and spectral domains, preserving essential spatial details and spectral continuity. The denoising task is formulated as an optimization problem that integrates low-rank tensor decomposition and KGLR. This problem is efficiently solved using the ALM method. The proposed approach achieves a balance between noise suppression and detail preservation, providing an effective solution for HSI denoising in high-dimensional, noisy conditions.  Results and Discussions  The proposed LRTDKGLR algorithm effectively exploits the inherent spatial and spectral correlations in HSI data. By incorporating a low-rank constraint and leveraging the similarity between adjacent spectral bands, the model enhances its capacity to capture inter-band spectral dependencies. Fig. 2 illustrates the denoising performance of various algorithms under Scenario 3, which combines Gaussian and impulse noise, applied to the 106th band of the Washington DC Mall dataset. A magnified view of a local region is also provided. The LRTV, LRTDTV, and LRTDGTV algorithms remove most of the noise but introduce over-smoothing, resulting in the loss of important edge structures. The NGmeet algorithm performs less effectively due to insufficient utilization of spatial neighborhood relationships. Although the LRMR, GLF, and TSLRLN algorithms reduce part of the noise, residual artifacts persist and over-smoothing remains evident. In contrast, the proposed LRTDKGLR algorithm demonstrates substantially improved denoising performance, supported by both visual comparisons and quantitative metrics. Fig. 3 presents the PSNR and SSIM values across multiple spectral bands under Scenario 3. The proposed method achieves higher performance in both metrics across most bands, indicating enhanced noise suppression and improved preservation of structural details. Numerical results summarized in Table 2 further confirm the effectiveness of the proposed approach, which consistently outperforms competing algorithms under identical conditions, providing superior denoising performance. Additionally, Table 3 shows that LRTDKGLR exceeds two representative deep learning-based methods, HSID-CNN and HSI-SDeCNN, across different noise levels. These findings demonstrate the robustness of the proposed method in the presence of Gaussian noise. Compared with deep learning-based approaches, LRTDKGLR offers practical advantages, including no requirement for pre-training, ease of deployment, and higher computational efficiency, which collectively improve its applicability to real-world HSI denoising tasks.  Conclusions  This study proposes an HSI denoising algorithm, LRTDKGLR, which integrates low-rank prior knowledge with a product graph model. The algorithm models HSI data as a product graph using GSP theory and applies Tucker decomposition to extract a core tensor and factor matrices, effectively capturing global spatial and spectral correlations. Simultaneously, KGLR is employed to model piecewise smoothness in both spatial and spectral dimensions, enhancing the representation of inter-dimensional relationships within the data. The denoising task is formulated as an optimization problem that combines low-rank tensor decomposition with KGLR constraints and is efficiently solved using the ALM method. Experimental results on simulated and real-world datasets demonstrate that the proposed LRTDKGLR algorithm achieves superior noise suppression while preserving structural details, validating its effectiveness for HSI denoising.
Noise-tolerant Terminal Zeroing Neural Networks for Solving Time-varying Quadratic Programming: A Triple Power-rate Speeding-up Strategy
ZHONG Guomin, XIAO Likun, WANG Liming, SUN Mingxuan
Available online  , doi: 10.11999/JEIT250128
Abstract:
  Objective  The computational performance of Zeroing Neural Networks (ZNNs) is enhanced by introducing additional power terms into the activation function. However, this strategy complicates the derivation of explicit settling time expressions. To address this issue, a triple power-rate activation function is designed, and a power-rate speeding-up noise-tolerant terminal ZNN is constructed, through which an exact expression for the settling time is derived. In previous studies, the optimization criterion parameter for repetitive motion planning was typically constant, which may reduce the operational efficiency of robotic manipulators. To overcome this limitation, a time-varying parameter optimization criterion is developed to satisfy task requirements at different stages of repetitive motion planning, thereby improving the operational efficiency of redundant robotic manipulators during task execution.  Methods  A triple power-rate activation function is proposed, extending the conventional bi-power activation function, and a power-rate speeding-up noise-tolerant terminal ZNN is constructed. The convergence process under different parameter settings is analyzed, and explicit settling time expressions are derived. Theoretical analysis confirms that the proposed neural network can effectively suppress vanishing noise. For the repetitive motion planning problem of redundant manipulators, the power-rate speeding-up noise-tolerant terminal ZNN is employed as a solver to ensure acquisition of the desired end-effector trajectory within fixed time. To address the limitations of constant-parameter optimization criteria in repetitive motion planning, a time-varying parameter optimization criterion is designed, which demonstrably improves the operational efficiency of redundant manipulators.  Results and Discussions  In this study, the power-rate speeding-up noise-tolerant terminal ZNN is employed together with bi-power-rate terminal ZNNs to solve time-varying quadratic programming problems. Simulation results show that the proposed power-rate speeding-up noise-tolerant terminal ZNN achieves a faster convergence rate (Fig.2(a), Fig.2(b)) and demonstrates improved capability in suppressing vanishing noise (Fig.2(c)). The convergence process of neural computational error under different parameter conditions is analyzed without noise (Fig 3). Furthermore, the power-rate speeding-up noise-tolerant terminal ZNN is applied to the repetitive motion planning problem of redundant manipulators. Its effectiveness in solving repetitive motion planning is validated (Fig. 4), and the integration of a time-varying parameter optimization criterion further enhances the operational efficiency of redundant manipulators (Fig. 5).  Conclusions  A power-rate speeding-up noise-tolerant terminal ZNN is proposed for solving time-varying quadratic programming problem with time-varying equality constraints, ensuring fixed time convergence of neural computing errors. Compared with conventional bi-power-rate terminal ZNNs, the proposed network achieves faster convergence and stronger noise-tolerance performance. To address the limitations of constant-parameter optimization criteria in repetitive motion planning, a time-varying parameter optimization criterion is designed and shown to improve the operational efficiency of redundant manipulators.
Cryption and Network Information Security
A Distributed KBB Index Tree Multi-Keyword Fuzzy Sorting Search Scheme
SUN Jin, SONG Nana, WANG Lu, KANG Mengna, YE Kexin
Available online  , doi: 10.11999/JEIT250151
Abstract:
  Objective  The rapid advancement of information technology has driven significant transformation in the medical domain. As a cornerstone of medical informatization, Electronic Health Records (EHRs) play a critical role in improving healthcare service efficiency and supporting medical research. Effective use of patient EHRs can enhance diagnostic and treatment processes and offer reliable data support for disease prevention and novel drug discovery. However, conventional EHR systems face several challenges, including data silos, secure sharing complexity, and heightened privacy risks. Data silos hinder the seamless exchange and integration of EHRs across institutions, limiting the potential for collaborative medical practices. A central challenge in secure data sharing lies in ensuring data interoperability without compromising patient privacy—a problem of urgent concern in modern healthcare systems. Moreover, privacy breaches not only jeopardize patient welfare but also damage the credibility and trustworthiness of healthcare institutions. To address these issues, this paper proposes a distributed multi-keyword fuzzy sorting search scheme based on a Keywords Balanced Binary (KBB) index tree. The scheme leverages the decentralized, tamper-proof, and traceable features of blockchain technology to create a secure and efficient framework for EHR sharing. By eliminating data silos, the blockchain facilitates cross-institutional interoperability, while smart contracts maintain strict privacy protections throughout the data sharing process.  Methods   The scheme first applies a K-means clustering algorithm to categorize EHRs, followed by hierarchical encrypted storage via the Interplanetary File System (IPFS). This dual-layer approach ensures distributed data management while protecting user privacy and enhancing storage robustness and fault tolerance. To improve search performance, the Porter stemming algorithm standardizes query keywords, reducing ambiguity from semantic variations and enabling consistent semantic matching. A KBB index tree is then constructed over the clustered EHRs to support fuzzy keyword matching and high retrieval precision. This structure incorporates adaptive sorting to reduce search latency. Access control is enforced using smart contracts, which implement fine-grained, role-based authentication to ensure that only authorized users can retrieve EHRs, thereby minimizing the risk of data leaks. During query processing, ranked searches are conducted within the KBB index tree. Once matching encrypted EHRs are located, data retrieval is performed via IPFS. Blockchain-based hashing ensures the integrity and immutability of the retrieved data, protecting it from tampering or corruption. Finally, users decrypt the data with private keys to access the original EHR content, completing the secure retrieval process.  Results and Discussions  Simulation experiments demonstrate that the proposed scheme offers superior implementation efficiency compared to related methods in terms of both time and storage overhead. The index tree construction time is shorter than that reported in[Ref. 7] and[Ref. 8], with an efficiency improvement of 80%–85% (Fig. 3). Although the trap generation time is longer than in[Ref. 7] and[Ref. 15] (Fig. 4), this increase stems from the scheme’s support for multiuser search, whereas[Ref. 7] only supports single-user search. The combined search time for indexing and trap generation is lower than in[Ref. 7],[Ref. 8], and[Ref. 15], with search efficiency improved by 50.5% and 75.8%, respectively (Fig. 5). The ciphertext generation and decryption times are also markedly better than those in the comparative literature, with improvements of 33.3% and 60%, respectively (Fig. 6, Fig. 7). Furthermore, the storage costs for keys and ciphertext in this scheme are lower than those of the other methods (Fig. 8, Fig. 9).  Conclusions  The proposed scheme substantially improves multi-keyword search efficiency, enabling users to rapidly locate target information within large-scale EHR datasets while enhancing the accuracy of search results. The fuzzy search functionality allows users to retrieve data even when the exact form of the keywords is unknown, increasing the flexibility and applicability of the search process. The encryption algorithm has been rigorously tested in experimental analyses, demonstrating that it maintains semantic security under the known ciphertext model, thereby ensuring EHR confidentiality. In addition, the algorithm performs well in terms of implementation efficiency, meeting the practical needs of EHR applications in medical contexts while maintaining strong security guarantees. By combining high-precision multi-keyword indexing with robust privacy protection, this scheme offers a scalable and secure framework for EHR sharing that meets the interoperability and confidentiality requirements of modern healthcare systems.
Decentralized Anti-quantum Internet of Vehicles Data Sharing Scheme
YU Huifang, DANG Ningze
Available online  , doi: 10.11999/JEIT250144
Abstract:
  Objective  Internet of Vehicles (IoV) faces emerging threats from quantum computing. NTRU lattice offers an effective approach to achieve anti-quantum security in data sharing. This study develops a decentralized anti-quantum IoV data sharing scheme. New scheme ensures the IND-CCA2 and UF-CMA security under the hardness of NTRU lattice computing problem; it uses short keys and small storage space to maintain low calculation complexity. Blockchain is employed to support the vehicle authentication, access control, key distribution and automated execution of smart contracts. New scheme provides a reliable anti-quantum solution for IoV data sharing and technical support for future development of intelligent transportation.  Methods  Based on unique identifiers and attribute set, the smart contract uses the Gaussian sampling algorithm to generate partial private keys and attribute private keys for users. Public-private key pairs are obtained during the key generation phase. IoV model integrates the blockchain, smart contracts, IPFS, RSUs, and OBUs. Smart contracts store and verify the data-sharing records, while IPFS stores encrypted IoV data to enhance the reliability and reduce the storage cost. RSUs collect the data uploaded by OBUs and forward it to the blockchain which broadcasts real-time traffic information to vehicles. Smart contracts dynamically evaluate the user permissions according to predefined access policies and trigger corresponding operations. Concurrent access experiments were conducted in large-scale and dynamic IoV scenario. Simulations under varying concurrent OBU access requests demonstrated that distributed blockchain architecture and smart contract permission verification enables the system to maintain stable performance under high concurrency.  Results and Discussions  The proposed scheme computes the data sharing signcryption strategy in accordance with current IoV security requirements. Table 2 compares its security characteristics and communication overhead. Table 3 reports four instances of private key, ciphertext and signature sizes, showing that the proposed scheme requires less overhead. Table 4 demonstrates that, as the number of concurrent users increases from 100 to 1000, the proposed scheme achieves lower total latency and more linear Throughput-Per-Second (TPS) growth compared with scheme [9]. Figure 2 illustrates that although the time consumption of all schemes increases with higher parameter values, the proposed scheme consistently exhibits the lowest time cost.  Conclusions  In the proposed scheme, the user private keys are bound to attribute sets and the data access is controlled by predefined access policies. This design ensures that only users with specific attributes can perform unsigncryption, thereby enabling fine-grained access control. The trapdoor generation and Gaussian sampling algorithms within the NTRU lattice reduce key size while ensuring anti-quantum security for IoV data transmission and storage. The integration of IPFS, blockchain, and smart contracts enables efficient index management and permission control, which strengthens the reliability of IoV system. The scheme supports concurrent access by large numbers of IoV users and has the potential to mitigate traffic accidents.
Information Entropy-Driven Black-box Transferable Adversarial Attack Method for Graph Neural Networks
WU Tao, JI Qionghui, XIAN Xingping, QIAO Shaojie, WANG Chao, CUI Canyixing
Available online  , doi: 10.11999/JEIT250303
Abstract:
  Objective  Graph Neural Networks (GNNs) achieve state-of-the-art performance in modeling complex graph-structured data and are increasingly applied in diverse domains. However, their vulnerability to adversarial attacks raises significant concerns for deployment in security-critical applications. Understanding and improving GNN robustness under adversarial conditions is therefore crucial to ensuring safe and reliable use. Among adversarial strategies, transfer-based black-box attacks have attracted considerable attention. Yet existing approaches face inherent limitations. First, they rely heavily on gradient information derived from surrogate models, while insufficiently exploiting critical structural cues embedded in graphs. This reliance often leads to overfitting to the surrogate, thereby reducing the transferability of adversarial samples. Second, most methods adopt a global perspective in perturbation selection, which hinders their ability to identify local substructures that decisively influence model predictions, ultimately resulting in suboptimal attack efficiency.  Methods  Motivated by the intrinsic structural characteristics of graph data, the latent association between information entropy and node vulnerability is investigated, and an entropy-guided adversarial attack framework is proposed. For homogeneous GNNs, a transferable black-box attack method, termed NEAttack, is designed. This method exploits node entropy to capture the structural complexity of node-level neighborhood subgraphs. By measuring neighborhood entropy, reliance on surrogate model gradients is reduced and perturbation selection is made more efficient. Based on this framework, the approach is further extended to heterogeneous graphs, leading to the development of GEHAttack, an entropy-based adversarial method that employs graph-level entropy to account for the semantic and relational diversity inherent in heterogeneous graph data.  Results and Discussions  The effectiveness and generalizability of the proposed methods are evaluated through extensive experiments on multiple datasets and model architectures. For homogeneous GNNs, NEAttack is assessed against six representative baselines on four datasets (Cora, CoraML, Citeseer, and PubMed) and three GNN models (Graph Convolutional Network (GCN), Graph Attention Network (GAT), Simplified Graph Convolution (SGC)). As reported in (Table 1) and (Table 2), NEAttack consistently outperforms existing approaches. In terms of accuracy, average improvements of 10.25%, 17.89%, 6.68%, and 12.6% are achieved on Cora, CoraML, Citeseer, and PubMed, respectively. For the F1-score, the corresponding gains are 9.41%, 16.83%, 6.21%, and 17.24%. Random Attack and Delete Internal, Connect External (DICE), which rely on randomness, exhibit stable but weak transferability, leading to only minor reductions in model performance. Meta-Self and Projected Gradient Descent (PGD) generate effective adversarial samples in white-box scenarios but show poor transfer performance due to overfitting to surrogate models. AtkSE and GraD perform better but remain affected by overfitting, while their computational cost increases sharply with data scale. For heterogeneous GNNs, GEHAttack is compared with four baselines on three datasets (ACM, IMDB, and DBLP) and six Heterogeneous Graph Neural Network (HGNN) models (Heterogeneous Graph Attention Network (HAN), Heterogeneous Graph Transformer (HGT), Simple Heterogeneous Graph Neural Network (SimpleHGN), Relational Graph Convolutional Network (RGCN), Robust Heterogeneous (RoHe), and Fast Robust Heterogeneous Graph Convolutional Network (FastRo-HGCN)). As shown in (Table 3 and Table 4), GEHAttack exhibits clear advantages. On the ACM dataset, compared with the HG Baseline, GEHAttack improves the average Micro-F1 and Macro-F1 scores of HAN, HGT, SimpleHGN, and RGCN by 3.93% and 3.46%, respectively. On the more robust RoHe and FastRo models, the corresponding improvements are 2.75% and 1.65%. Similar improvements are also observed on the IMDB and DBLP datasets, confirming the robustness and transferability of GEHAttack.  Conclusions  This study presents a unified entropy-oriented adversarial attack framework for both homogeneous and heterogeneous GNNs in black-box transfer settings. By leveraging the relationship between entropy and structural vulnerability, the proposed NEAttack and GEHAttack methods address the key limitations of gradient-dependent approaches and enable more efficient perturbation generation. Extensive evaluations across diverse datasets and models demonstrate their superiority in both performance and adaptability, providing new insights into advancing adversarial robustness research on graph-structured data.
Radar, Navigation and Array Signal Processing
Double Deep Q-Network for Non-Uniform Position Optimization in Sparse Circular Arrays
CHEN Tao, LIANG Yaopeng, CHEN Xu, ZHAN Lei
Available online  , doi: 10.11999/JEIT250125
Abstract:
  Objective  To address sparse circular array deployment in practical engineering scenarios, where the number and positions of array elements are constrained, this study proposes an optimization algorithm based on Double Deep Q-Networks (DDQN) to maintain Direction-of-Arrival (DOA) estimation performance under limited channel conditions. This method enables flexible and efficient array design strategies and overcomes challenges that conventional optimization approaches are unable to resolve effectively. The sparse circular array design problem is formulated by minimizing the two-dimensional DOA estimation Ziv–Zakai Bound (ZZB) and Peak Sidelobe Level (PSL) as joint objectives to ensure both angular resolution and estimation accuracy. The state space, action space, and reward function are constructed accordingly, and the DDQN algorithm is employed to solve the optimization task. Experimental results demonstrate that the proposed method achieves stable convergence and robust DOA estimation performance under deployment constraints, and confirm its practical effectiveness.  Methods  To optimize sparse circular arrays under structural and channel limitations, a DDQN-based design approach is proposed. The method selects a subset of elements from a uniform circular array to maximize DOA estimation accuracy and angular resolution while satisfying constraints on the number of antennas and inter-element spacing. The array design task is cast as a constrained optimization problem with the two-dimensional DOA ZZB and PSL as the performance metrics. Within the reinforcement learning framework, the state space reflects potential array configurations, the action space corresponds to candidate element selections, and the reward function is derived from the optimization objectives. Once trained, the DDQN model outputs an optimized sparse array configuration that balances resolution and sidelobe suppression under the given constraints.  Results and Discussions  Simulation results show that the reward function of the proposed algorithm converges as the number of training episodes increases (Fig. 8). In contrast, traditional reinforcement learning algorithms exhibit slower convergence and yield suboptimal solutions, while genetic algorithms tend to suffer from premature convergence. The designed sparse circular array satisfies the optimization constraints, including the maximum inter-element spacing requirement (Fig. 7(a)). Under a three-source scenario, the array demonstrates robust DOA estimation capability, effectively resolving multiple incident signals DOA estimation problem (Fig. 9). In evaluations of DOA estimation with Root Mean Squared Error (RMSE) under varying Signal-to-Noise Ratio (SNR) conditions (Fig. 10), the proposed array achieves an estimation error below 0.5° when SNR is ≥ 0 dB. Compared with other sparse circular arrays, it achieves the lowest RMSE, indicating superior estimation performance. In angular resolution tests, the proposed array also exhibits lower PSL values (Table 2) and a higher angle estimation success rate. When the angular separation is ≥ 3°, the success rate exceeds 95% (Fig. 10), confirming the array’s high DOA estimation accuracy and strong angular resolution.  Conclusions  This study formulates sparse circular array optimization as a constrained problem with the maximum inter-element spacing as a design constraint. To enhance both DOA estimation accuracy and angular resolution, the two-dimensional DOA estimation ZZB and PSL are minimized as joint objectives function. A DDQN algorithm with a dual-network structure is employed to solve the optimization problem and generate the array configuration. Simulation experiments verify that, under channel limitations, the proposed array satisfies the imposed constraints and achieves the intended optimization goals. Compared with other sparse circular arrays, the design demonstrates superior overall DOA estimation performance.
Features Extraction and Correlation Analysis of Multi-Source Data for Maritime Targets
LIU Ningbo, ZHANG Zihao, CHEN Baoxin, DONG Yunlong, LI Jia
Available online  , doi: 10.11999/JEIT250200
Abstract:
  Objective  The growing demand for maritime target detection and recognition has made multi-source information fusion a critical area of research. Different sensing modalities including radar, visible-light imaging, and infrared imaging offer complementary information that can improve detection and classification. However, the correlation among heterogeneous features extracted from these sources remains insufficiently understood. In addition, the effects of observational conditions on feature stability and discriminability needs further investigation. This study investigates the intrinsic relationships among multi-source features, evaluates their robustness under varying conditions, and provides theoretical support for effective multi-source feature fusion in maritime target detection.  Methods  Nine physically interpretable features are extracted across three categories: radar features (radial dimension, number of peaks, position distribution entropy, cross-range dimension, and relative average amplitude), visible image features (number of edges and horizontal projected width), and infrared image features (high-temperature connected component count and thermal texture energy). To ensure accurate feature extraction, data preprocessing consists of cleaning each compressed dataset. Radar data with excessively low signal-to-noise ratios and images with extensive occlusion are excluded. For dataset construction, radar echo data are visualized frame by frame, and a subset of radar, visible-light, and infrared images is manually annotated using LabelImg. Targets are classified into five types: passenger ships, dry cargo ships, container ships, tugboats, and search-and-rescue vessels. Based on these annotations, deep learning models are independently trained for each modality to automate annotation, and the results are manually validated to ensure quality. A standardized target dataset is then constructed by integrating the manually and automatically labeled data. Radar echo data are temporally aligned with visible-light and infrared images using prior time synchronization information. Features are extracted from each modality, and inter-feature correlations are analyzed. Spearman correlation coefficients are used to quantify relationships, and hypothesis testing is conducted to assess significance, revealing intrinsic associations among multi-source features.  Results and Discussions  Comparative analysis indicates that the correlation between radar echo and image features is strongly affected by feature attributes. Size-related features show stronger correlations, reflecting greater stability for multi-source data fusion, whereas structural features are more sensitive to observation conditions and exhibit weaker correlations. External factors including target motion state, ship type, and radar polarization mode also influence feature extraction and intermodal correlation. When targets are anchored, posture remains stable and motion blur is reduced, strengthening the correlation between radar structural features and image features. Larger vessels, such as container ships and passenger ships, benefit from multiple scattering centers and larger physical dimensions, which enhance feature extraction and intermodal correlation. In contrast, smaller vessels, such as tugboats and search-and-rescue boats, exhibit irregular structures and weaker radar backscatter, leading to lower correlations. The HH polarization mode, being less sensitive to background clutter, improves radar feature stability across various target types and enhances alignment with image features. Image feature stability also varies by modality: visible-light imaging is effective for extracting structural features, such as contours and edges, but is sensitive to illumination and occlusion; infrared imaging offers more stable size-related features and performs robustly in complex and low-visibility environments. These results highlight the complementary properties of multi-source features and establish a theoretical basis for their integration, supporting improved maritime target detection and classification.  Conclusions  This study demonstrates the complementary characteristics of multi-source features and their potential to improve maritime target detection and recognition. By analyzing feature correlations and stability across varying observational conditions, the results provide a theoretical foundation for refining multi-source fusion strategies. These insights support the development of more robust and reliable detection algorithms, contributing to enhanced situational awareness in maritime surveillance and defense.
Constant-Modulus Waveform Design for SLP-Based DFRC Systems
ZHAI Yihao, ZHAO Hongyu, NIU Fengliang, LI Jingyan
Available online  , doi: 10.11999/JEIT250288
Abstract:
  Objective  Dual-Function Radar-Communication (DFRC) Integration is a promising technology for addressing spectrum congestion. The key challenge in DFRC lies in designing a dual-functional waveform that simultaneously supports both radar sensing and communication. Previous research has primarily focused on designing transmit waveforms under Multi-User Interference (MUI) energy constraints. However, not all MUI energy is detrimental in DFRC systems. Symbol-Level Precoding (SLP) can exploit symbol information to transform harmful MUI into constructive components, thereby enhancing communication Quality of Service (QoS). To further leverage the advantages of SLP in mitigating MUI energy and to improve target detection in cluttered environments, this work studies the joint design of transmit waveforms and receive filters for DFRC systems.  Methods  The analytical expression of the radar output Signal-to-Interference-plus-Noise Ratio (SINR) is first derived. To guarantee communication performance, Constructive Interference (CI) constraints are formulated using SLP. Based on these, a joint optimization problem is established, where maximizing the radar output SINR serves as the objective function and the communication CI constraints are imposed. A constant-modulus condition is also enforced to prevent nonlinear distortion of the transmitted waveform. To address the non-convex nature of the joint optimization problem, an iterative algorithm based on cyclic optimization is proposed. For a fixed transmitted waveform, the subproblem reduces to a Minimum Variance Distortionless Response (MVDR) problem. For a fixed filter, the communication CI constraints are incorporated into the objective function of the subproblem as a penalty term, leading to an unconstrained problem whose feasible region is a Riemannian complex manifold. This problem is then solved efficiently using the Riemannian Conjugate Gradient (RCG) method.  Results and Discussions  As shown in (Fig 3): (a) the proposed method converges within a short time, and (b) the designed waveform satisfies the constant-modulus constraint. (Figs. 4 and 5) analyze the effect of the number of communication users and the communication QoS threshold on radar output SINR and communication SER, confirming the trade-off between radar sensing and communication performance. (Figs. 6 and 7) illustrate the effect of interference power and communication QoS threshold on radar output SINR and communication SER, demonstrating the robustness of the proposed method under different interference power levels. When communication QoS requirements are fixed, the radar output SINR and communication SER performance of the three compared cases remain similar. (Fig. 8(a)) shows the radar output SINR versus the number of communication users. The results indicate that the proposed CI-based waveform consistently outperforms the conventional MUI-based waveform in radar output SINR. Moreover, the two CI-based methods achieve almost identical SINR performance. (Fig. 8(b)) presents the radar received beampatterns, where it can be observed that the deep nulls formed in the interference directions are deeper than those obtained with the conventional MUI-based method. This confirms that the proposed method effectively suppresses signal-dependent interference. (Fig. 9(a)) presents the received communication constellations. The synthesized communication signals of the CI-based methods are more sparsely distributed and exhibit higher power compared with the conventional MUI-based method. (Fig. 9(b)) further evaluates the communication SER performance, showing that the proposed method achieves a lower SER. This improvement arises because the communication CI constraints transform harmful MUI into useful signal energy, thereby enhancing communication QoS. (Fig. 10) presents the relationship between CPU time and the number of communication users. The results show that CPU time increases for all methods as the number of communication users grows, since a larger number of users reduces the feasibility region. Nevertheless, the proposed method demonstrates significantly higher computational efficiency compared with existing approaches.  Conclusions  This paper investigates the joint design of transmit waveforms and receive filters for MIMO-DFRC systems. A joint optimization problem is formulated to maximize radar output SINR under constant-modulus and communication CI constraints. To solve the resulting non-convex problem, an alternating optimization algorithm based on the RCG method is developed. Simulation results demonstrate that CI-based methods outperform the MUI-based ADMM method in both radar sensing performance and communication QoS. In addition, the proposed method achieves faster convergence compared with the MM-neADMM method, with only a minor loss in radar sensing performance.
A Successive Convex Approximation Optimization based Prototype Filter Design Method for Universal Filtered Multi-Carrier Systems
HUA Jingyu, YANG Le, WEN Jiangang, ZOU Yuanping, SHENG Bin
Available online  , doi: 10.11999/JEIT250278
Abstract:
  Objective  In response to the extensive demands of sixth-generation (6G) communications, new waveform designs are expected to play a critical role. Conventional Orthogonal Frequency Division Multiplexing (OFDM) relies on strict orthogonality among subcarriers; however, this orthogonality is highly vulnerable to synchronization errors, which lead to severe Inter-Carrier Interference (ICI). To address this issue, filtered multicarrier modulation techniques apply high-performance filters to each subcarrier, thereby confining spectral leakage and mitigating ICI caused by non-ideal frequency synchronization. Among these techniques, Universal Filtered Multi-Carrier (UFMC) has shown particular promise, offering enhanced spectral flexibility and reduced out-of-band emissions compared with traditional OFDM. Despite these advantages, most existing studies recommend Dolph-Chebyshev (DC) filters as UFMC prototype filters. Nevertheless, DC filters suffer from limited controllability over design parameters and insufficient robustness against interference. Recent research has sought to improve system performance by applying constrained optimization techniques in filter design, typically optimizing metrics such as Signal-to-Interference Ratio (SIR) and Signal-to-Interference-plus-Noise Ratio (SINR). Nevertheless, the Symbol Error Rate (SER) has not achieved an optimal level, indicating room for further improvement. To bridge this gap, this paper proposes a novel prototype filter design method that directly targets the average SER in interference-limited UFMC systems. This approach improves the anti-interference capability of UFMC systems and contributes to the development of robust waveform solutions for 6G communications.  Methods  This study first derives the SINR of the UFMC system under non-zero Carrier Frequency Offset (CFO) and formulates the SER expression under interference-limited conditions. A mathematical model is then established for prototype filter optimization, with SER defined as the objective function. Because the nonlinear coupling between SINR and the filter coefficients introduces strong non-convexity, the Successive Convex Approximation (SCA) framework is employed to locally linearize the non-convex components. Furthermore, a quadratic upper-bound technique is applied to guarantee both convexity and convergence of the approximated problem. Finally, an iterative algorithm is developed to solve the optimization model and determine the optimal prototype filter.  Results and Discussions  The interference suppression capability of the proposed SCA filter is comprehensively evaluated, as shown in Figs. 2 and 3. The simulation results in Fig. 2 reveal several important findings. (1) The deviation between the theoretical SINR and Monte Carlo simulation results is less than 0.1 dB (Fig. 2), confirming the accuracy of the derived closed-form expressions. (2) CFO is shown to have a strong association with system interference. As the residual CFO increases from 0 to 0.05, the SINR with conventional DC filters decreases by 3.6 dB, whereas the SCA filter achieves an SINR gain of approximately 1 dB compared with the DC filter. (3) Under a CFO of 0.025, the UFMC waveform demonstrates clear superiority over the ideal OFDM system. At a Signal-to-Noise Ratio (SNR) of 18 dB, the UFMC system with the SCA filter attains an SINR of 18.4 dB, outperforming OFDM by 0.3 dB. Fig. 3 further highlights the robustness of the SCA filter in dynamic interference environments. Although the SER increases with both larger CFO and higher modulation orders, the SCA filter consistently yields the lowest SER across all interference scenarios. Under severe interference conditions (CFO = 0.05, 16QAM modulation, SNR = 17 dB), the SCA filter achieves an SER of 7.4×10–3, markedly outperforming the DC filter, which exhibits an SER of 2.9×10–2. These results demonstrate that the proposed SCA filter substantially enhances the anti-interference capability of UFMC systems.  Conclusions  This study first derives analytical expressions for the SINR and SER of the UFMC system under CFO. On this basis, an optimization model is established to design the prototype filter with the objective of minimizing the average SER. To address the non-convexity arising from the nonlinear coupling between SINR and filter coefficients, the SCA method is employed to reformulate the problem into a series of convex subproblems. An iterative algorithm is then proposed to obtain the optimal prototype filter. Simulation results demonstrate that, compared with conventional filters, the proposed SCA-based optimization algorithm provides flexible control over key filter parameters, achieving a narrower transition band and higher stopband attenuation under the same filter length. This improvement translates into significantly enhanced anti-interference performance under various system conditions. In summary, the main contributions of this work are: (1) Proposing a novel SCA-based optimization method for UFMC prototype filter design, which overcomes the parameter control limitations of traditional DC filters; (2) Systematically analyzing the performance advantages of the SCA filter under different modulation schemes and CFO conditions, and quantitatively demonstrating its contributions to SINR and SER improvements.
Low Elevation Angle Estimation Method for MIMO Radar in Complex Terrain
WANG Jiajia, GUO Rui, LIU Qi, ZHANG Yue, CHEN Zengping
Available online  , doi: 10.11999/JEIT250236
Abstract:
  Objective  Conventional low-elevation angle estimation algorithms for Multiple-Input Multiple-Output (MIMO) radar generally assume a single-path propagation model, which limits their applicability in complex terrain where multipath effects are time-varying. Compressive Sensing (CS) algorithms exploit the sparsity of direct and multipath signals in the spatial domain and remain effective in such environments. Nonetheless, CS-based approaches for MIMO radar require the construction of a two-dimensional grid dictionary, and their computational complexity increases sharply as the number of multipath components grows. Existing complexity-reduction methods sacrifice array aperture, leading to degraded estimation accuracy. To resolve the trade-off between aperture utilization and computational complexity in low-elevation angle estimation for MIMO radar under complex terrain conditions, a tensor-based two-step estimation algorithm is proposed.  Methods  A three-dimensional tensor observation model is first established to fully preserve the multi-dimensional structure of the received signal, and the tensor signal subspace is extracted using High-Order Singular Value Decomposition (HOSVD). After eliminating redundancy in the tensor subspace, Sparse Bayesian Learning (SBL) is applied to rapidly obtain initial estimates of the low-elevation and multipath angles. These initial results are then refined by an alternating iterative Generalized Multiple Signal Classification (GMUSIC) algorithm, which leverages the complete tensor subspace. The proposed method maintains full array aperture, adapts to scenarios with unknown numbers of multipaths, and achieves a favorable balance between estimation accuracy and computational efficiency.  Results and Discussions  Simulation results demonstrate that the proposed algorithm achieves high estimation accuracy under both single- and double-reflection paths (Fig. 25) compared with other benchmark algorithms, while maintaining lower computational complexity (Table 1). Relative to the sub-optimal Alternative Projection Maximum Likelihood (APML) algorithm, the running speed is improved by 92.16%. In addition, the method remains robust under time-varying multipath conditions (Fig. 6) without requiring prior knowledge of the spatial distribution of reflection paths. Validation with real measured data (Fig. 810) further confirms its practical applicability: 86.95% of estimates fall within the 0–0.4° error range, and the error remains consistently below 0.6° across the observation window. These findings highlight the superior estimation accuracy and reliability of the proposed method, supporting its suitability for real-world engineering applications.  Conclusions  By integrating tensor modeling, sparse preliminary estimation, and alternating iterative optimization, the proposed algorithm fully exploits the multi-dimensional structure of the received signal and the complete array aperture of MIMO radar. It demonstrates high estimation accuracy while maintaining low computational complexity. Simulation results confirm its effectiveness and robustness in complex terrain, and validation with measured data further verifies its feasibility and engineering applicability. Nonetheless, this study is limited to a single-target scenario with a relatively simple motion trajectory. Future research should extend the method to address complex motion patterns with multiple targets.
Global Navigation Satellite System Interference Source Localization Algorithm via Single-antenna Synthetic Aperture on the Unmanned Aerial Vehicle Platform
WANG Lu, WANG Xuan, WU Renbiao
Available online  , doi: 10.11999/JEIT250169
Abstract:
  Objective  As Radio Frequency (RF) spectrum congestion intensifies and electromagnetic interference incidents occur more frequently, the reliable operation of the Global Navigation Satellite System (GNSS) is increasingly threatened. To ensure GNSS functionality, interference sources must be accurately monitored and localized. Conventional localization methods are constrained by the limited aperture of physical antenna arrays and often incur high deployment costs. With the gradual opening of low-altitude airspace, small Unmanned Aerial Vehicles (UAVs) are increasingly used as platforms for remote sensing, electronic reconnaissance, and communication tasks, offering a practical alternative for GNSS interference localization. Passive Synthetic Aperture (PSA) techniques have emerged as a solution to overcome the aperture limitations of physical arrays and are well suited for integration with UAV platforms. In this study, PSA technology is combined with a UAV-based system to locate GNSS interference sources. A unified signal model is established to describe various interference types. Based on this model, a slant-range-accurate Range-Doppler (RD) algorithm is derived and further integrated with the Back Projection (BP) algorithm to perform joint localization. This combined RD–BP approach enhances positioning accuracy while reducing computational complexity, demonstrating the feasibility and efficiency of the proposed UAV-based PSA localization framework.  Methods  This approach integrates a RD positioning algorithm based on a precise slant range expression with a BP algorithm. First, an accurate slant range expression is derived using the interference source direction and Doppler frequency. Applying this expression and the principle of stationary phase, the signal is transformed into the RD domain. Multiple matched filters are then constructed in the Doppler domain using the range gate. The outputs of these filters are converted into the range-azimuth domain for peak detection, yielding a coarse estimate of the interference source location. However, this initial estimate is susceptible to errors due to factors such as platform sampling rate and azimuth sampling interval. To refine the estimate, the BP algorithm is employed for fine positioning. A grid is defined around the coarse estimate, and each grid point is used to compute the projected signal. These projected signals are coherently accumulated with the received signals and integrated along the azimuth direction. Peak searching is then performed on the accumulated results to obtain the final, high-precision localization result.  Results and Discussions  Simulation results validate the effectiveness and excellent performance of the proposed RD-BP algorithm. As shown in (Fig. 7), (Fig. 10), and (Fig. 13), the algorithm successfully localizes interference sources after preprocessing for narrowband interference, Linear Frequency Modulated (LFM) interference, and spoofing interference. These results indicate the algorithm’s broad applicability across various interference types. (Fig. 14) illustrates the Root Mean Square Error (RMSE) under two scenarios in which the interference source is either close or distant. In both cases, RMSE decreases as the Signal-to-Noise Ratio (SNR) increases. When the source is nearby, RMSE changes slowly, indicating stable performance. When the source is distant, the algorithm maintains strong performance at higher SNRs. (Fig. 15) shows that under SNR conditions of –10 dB and 15 dB, RMSE increases with target distance. However, the RD-BP algorithm consistently outperforms other methods in accuracy, demonstrating its robustness and adaptability across varying distances. (Fig. 16) further confirms that RMSE stabilizes as the synthetic aperture length increases under both –10 dB and 15 dB SNR conditions. When the aperture length exceeds 100 m, the RD-BP algorithm achieves superior performance relative to other methods and approaches the Cramér–Rao Lower Bound (CRLB), indicating near-optimal estimation accuracy. (Table 1) compares the computational and localization performance of the proposed algorithm with that of the RD algorithm alone. While the RD algorithm is sensitive to platform sampling frequency, flight velocity, and other operational parameters, the RD-BP algorithm mitigates these limitations. It improves localization accuracy while maintaining manageable computational complexity, offering an effective trade-off between precision and efficiency.  Conclusions  This study addresses the problem of locating GNSS interference sources using low-altitude, low-speed UAV platforms. A localization method based on single-antenna synthetic aperture technology is proposed. The study begins by categorizing different types of GNSS interference and their respective preprocessing strategies. An RD algorithm is then derived based on the precise slant range representation, leveraging the relative direction of the interference source with respect to the UAV to achieve coarse localization. To address the accuracy limitations inherent in the RD algorithm due to initial condition constraints, an improved approach is proposed by incorporating the BP algorithm. This enhancement enables refined localization within a narrowed search region. Simulation results confirm that the proposed algorithm can reliably localize common GNSS interference sources under various interference conditions. Compared to conventional methods such as array antenna direction finding, the proposed technique requires only a single antenna mounted on the UAV platform, thereby increasing system flexibility and reducing both cost and design complexity. This demonstrates the feasibility of the method for rapid and efficient localization of GNSS interference sources.
Data Enhancement for Few-shot Radar Countermeasure Reconnaissance via Temporal-Conditional Generative Adversarial Networks
HUANG Linxuan, HE Minghao, YU Chunlai, FENG Mingyue, ZHANG Fuqun, ZHANG Yinan
Available online  , doi: 10.11999/JEIT250280
Abstract:
  Objective  Radar electronic warfare systems rely on Pulse Descriptor Words (PDWs) to represent radar signals, capturing key parameters such as Time of Arrival (TOA), Carrier Frequency (CF), Pulse Width (PW), and Pulse Amplitude (PA). However, in complex electromagnetic environments, the scarcity of PDW data limits the effectiveness of data-driven models in radar pattern recognition. Conventional augmentation methods (e.g., geometric transformations, SMOTE) fall short in addressing three core issues: (1) failure to capture temporal physical laws (e.g., gradual frequency shifts and pulse modulation patterns); (2) distributional inconsistencies (e.g., frequency band overflows and PW discontinuities); and (3) weak coupling between PDWs and modulation types, leading to reduced classification accuracy. This study proposes Time-CondGAN, a temporal-conditional generative adversarial network designed to generate physically consistent and modulation-specific PDW sequences, enhancing few-shot radar reconnaissance performance.  Methods   Time-CondGAN integrates three core innovations: (1) Multimodal Conditional Generation Framework: A label encoder maps discrete radar modulation types (e.g., VS, HRWS, TWS in Table 2) into 128-dimensional feature vectors. These vectors are temporally expanded and concatenated with latent noise to enable category-controllable generation. The generator architecture (Fig. 4) employs bidirectional Gated Recurrent Units (GRUs) to capture long-range temporal dependencies. (2) Multi-task Discriminator Design: The discriminator (Fig. 5) is designed for joint adversarial discrimination and signal classification. A shared bidirectional GRU with an attention mechanism extracts temporal features, while classification loss constrains the generator to maintain category-specific distributions. (3) Temporal-Statistical Joint Optimization: The training process incorporates multiple loss components: Supervisor Loss (Eq. 7): A bidirectional GRU-based supervisor enforces temporal consistency. Feature Matching Loss (Eq. 9): Encourages alignment between high-level features of real and synthetic signals. Adversarial and Classification Losses (Eqs. (8)–(9)): Promote distribution realism and accurate category separation. A two-stage training strategy is adopted to improve stability, consisting of pre-training (Fig. 2) followed by adversarial training (Fig. 3).  Results and Discussions  Time-CondGAN demonstrates strong performance across three core dimensions. (1) Physical plausibility is achieved through accurate modeling of radar signal dynamics. Compared with TimeGAN, Time-CondGAN reduces the Kullback-Leibler (KL) divergence by 28.25% on average. Specifically, the PW distribution error decreases by 50.20% (KL = 4.68), the TOA interval error decreases by 20.5% (KL = 0.636), and the CF deviation is reduced by 13.29% (KL = 7.93), confirming the suppression of non-physical signal discontinuities. (2) Downstream task enhancement highlights the model’s few-shot generation capability. With only 10 real samples, classification accuracy improves markedly: VGG16 accuracy increases by 37.2% (from 43.0% to 59.0%) and LSTM accuracy by 28.6% (from 49.0% to 63.0%), both substantially outperforming conventional data augmentation methods. (3) Ablation studies validate the contribution of key modules. Removing the conditional encoder increases the PW KL divergence by 107.4%. Excluding the supervisor loss degrades CF continuity by 76.2%, and omitting feature matching results in a 44.6% misalignment in amplitude distribution.  Conclusions  This study establishes Time-CondGAN as an effective solution for radar PDW generation under few-shot conditions, addressing three key limitations: temporal fragmentation is resolved through bidirectional GRU supervision, mode collapse is alleviated via multimodal conditioning, and modulation specificity is maintained by classification constraints. The proposed framework offers operational value, enabling over 35% improvement in recognition accuracy under low-intercept conditions. Future work will incorporate radar propagation physics to refine amplitude modeling (current KL = 8.51), adopt meta-learning approaches for real-time adaptation in dynamic battlefield environments, and expand the model to multi-radar cooperative scenarios to support heterogeneous electromagnetic contexts.
Signal Sorting Method Based on Multi-station Time Difference and Dirichlet Process Mixture Model
CHEN Jinli, WANG Yanjie, FAN Yu, LI Jiaqiang
Available online  , doi: 10.11999/JEIT250191
Abstract:
  Objective  Signal sorting is a crucial technology in electronic reconnaissance that enables the deinterleaving of mixed pulse sequences emitted by multiple radar radiation sources, thereby supporting military decision-making. With the rapid advancement of electronic technology, multi-station cooperative signal sorting has received increasing attention. However, existing multi-station signal sorting methods depend heavily on manually selected parameters, which limits adaptability. Moreover, in complex environments with pulse loss and noise interference, conventional methods struggle to process unpaired pulses effectively, reducing the accuracy and stability of sorting. To address these challenges, this study applies the Dirichlet Process Mixture Model (DPMM) to multi-station cooperative signal sorting. The proposed approach enables adaptive sorting even when the number of radiation sources is unknown or measurement errors exist, thereby improving flexibility and adaptability. Furthermore, it can effectively classify unpaired pulses caused by pulse loss or noise, enhancing the robustness and reliability of sorting. This research provides a novel strategy for signal sorting in complex electromagnetic environments and holds promising application value in radar signal processing.  Methods  In multi-station cooperative signal sorting, the spatial distribution of multiple receiving stations detecting the same radar signal makes efficient and accurate signal pairing and classification a core challenge. To address this issue, a multi-station cooperative signal sorting method based on the DPMM is proposed. The process comprises three stages: pulse pairing, time-difference clustering and sorting, and mismatched pulse classification. In the pulse pairing stage, identical pulses originating from the same radiation source are identified from the sequences intercepted by each receiving station. To ensure accurate pairing, a dual-constraint strategy is adopted, combining a time-difference window with multi-parameter matching. Successfully paired pulses are then constructed into a time-difference vector set, which provides the data foundation for the subsequent clustering and sorting stage. In the time-difference clustering and sorting stage, DPMM is employed to cluster the time-difference vector set. DPMM adaptively determines the number of clusters to model the data structure, enabling the system to infer the optimal cluster count. Gibbs sampling is used to optimize model parameters, further enhancing clustering robustness. Based on the clustering results, radar pulse sets are constructed, achieving signal sorting across multiple radiation sources. In the mismatched pulse classification stage, unpaired pulses caused by noise interference or pulse loss during transmission are further processed. DPMM is applied to fit radar pulse parameter vectors, including pulse width, radio frequency, and bandwidth. The affiliation degree of each mismatched pulse relative to the radar pulse sets is then calculated. Pulses with affiliation degrees exceeding a predefined threshold are merged into the corresponding pulse set, whereas those below the threshold are classified as anomalous pulses, likely due to interference or noise, and are discarded. This method enhances the adaptability and robustness of multi-station cooperative signal sorting and provides an effective solution for complex electromagnetic environments.  Results and Discussions  In the experimental validation, radar pulse data are generated through simulation to evaluate the effectiveness of the proposed method. Compared with traditional multi-station cooperative signal sorting approaches, the method achieves high-precision sorting without requiring prior knowledge of the number of radiation sources or parameter measurement errors, thereby demonstrating strong adaptability and practicality. To comprehensively assess performance in complex environments, simulations are conducted to analyze sorting capability under varying measurement errors, pulse loss rates, and interference rates. The final sorting results are summarized in (Table. 3). The results indicate that even in the presence of noise interference and data loss, most radar pulses are accurately identified, with only a small fraction misclassified as interference signals. The final sorting accuracy reaches 98.8%, confirming the robustness and stability of the method against pulse loss, noise, and other uncertainties. To further validate its superiority, the method is compared with other algorithms under different conditions. Sorting accuracy under different Time of Arrival (TOA) measurement errors (Fig. 6) shows that stable performance is maintained even under severe noise interference, reflecting strong noise resistance. Further analyses of sorting accuracy under different pulse loss rates and interference rates (Figs. 7 and 8) demonstrate that higher efficiency and stability are achieved in handling unpaired pulses, and pulses that fail to be paired are more accurately classified. The sorting accuracy of different algorithms in various scenarios (Fig. 9) further confirms that the method performs more consistently in complex environments, indicating higher adaptability. Overall, the method adapts well to diverse application scenarios and provides efficient, stable, and reliable signal sorting for multi-station cooperative electronic reconnaissance tasks.  Conclusions  This study proposes a multi-station cooperative signal sorting method based on the DPMM to address the limitations of traditional approaches, which rely heavily on prior information and perform poorly in processing unpaired pulses. By applying DPMM for adaptive clustering of time-difference information, the proposed method avoids sorting errors caused by improper manual parameter settings and effectively classifies unpaired pulses based on radar pulse parameter characteristics. Simulation results show that this method not only improves the accuracy and stability of multi-station cooperative signal sorting but also maintains high sorting performance even when the number of radiation sources is unknown or measurement errors are present, highlighting its engineering application value. Future research may extend this approach to dynamic electromagnetic environments and adaptive real-time processing to meet the demands of more complex electronic reconnaissance tasks.
Multi-Dimensional Resource Management Optimization Strategy for Multi-Group Target Tracking
LIN Mushen, CAO Bingxia, YAN Fenggang, MENG Xiangtian, SHA Minghui, LI Zhanguo, JIN Ming
Available online  , doi: 10.11999/JEIT250152
Abstract:
  Objective  Although existing resource management strategies can improve the performance of Multiple Target Tracking (MTT), they generally assume that targets are well separated. With the development of swarm intelligence, multiple maneuvering targets can adopt group merging and splitting tactics to enhance task efficiency. This often results in the presence of several closely spaced targets within a single tracking beam, making them difficult to distinguish. In such cases, networked radars must coordinate the allocation of signal bandwidth and pulse width to increase both the resolution of adjacent targets and overall MTT performance. However, previous studies have not accounted for the effect of signal resolution capability on MTT performance. When adjacent targets emerge in the surveillance region, resource allocation schemes derived from existing methods become insufficient. To address this limitation, this study proposes a joint Beam Selection and Multi-dimensional Resource Management (BSMRM) strategy. By jointly optimizing the transmit beam, power, signal bandwidth, and pulse width, the proposed approach markedly improves MTT performance for multiple groups of clustered maneuvering targets and enhances the resolution of adjacent targets within the surveillance area.  Methods  This study proposes a BSMRM strategy for netted radar systems to improve target resolution capability and MTT accuracy. First, the Bayesian Cramér-Rao Lower Bound (BCRLB) is derived under the Probabilistic Data Association (PDA) fusion rule and the Interacting Multiple Model-Extended Kalman Filter (IMM-EKF) framework. Resolution performance is quantified using the normalized magnitude of the ambiguity function for both observed and unobserved targets located within the beam. A utility function is then formulated using the logarithmic barrier method to quantify global performance. Multi-dimensional resource management is modeled as a constrained optimization problem, with the objective of maximizing global performance under system resource constraints. A three-stage solver is developed to address the optimization problem. Simulation results show that the proposed BSMRM strategy achieves comparable target resolution and MTT accuracy to state-of-the-art algorithms, while demonstrating superior efficiency in spectrum usage and power allocation.  Results and Discussions  As shown in (Table 1) and (Fig. 1), this study evaluates the effectiveness and robustness of the proposed BSMRM strategy in a complex scenario where four groups of swarm targets maneuver within the surveillance region of three radar nodes. The performance of the BSMRM strategy is assessed through four simulation experiments. Experiment 1 presents the allocation of transmission resources, including beam selection and power distribution. The results are shown in (Fig. 2), where color-coded regions represent normalized transmit power levels determined by the BSMRM strategy. Experiment 2 reports the allocation of waveform resources—specifically, signal bandwidth and pulse width. In this experiment, resolution thresholds corresponding to various target states are first established. The BSMRM strategy is then used to adjust the signal bandwidth and pulse width of each transmit beam to enable effective discrimination of swarm targets. Experiment 3 analyzes tracking performance through numerical simulations, comparing the proposed strategy with benchmark algorithms. A sequence of 60 tracking frames with a time interval of T = 2.5 s is simulated, with 300 Monte Carlo runs. Root Mean Square Error (RMSE) is used as the evaluation metric. (Fig. 4) displays the BCRLB and RMSE curves for each target group under the BSMRM strategy. The RMSE converges toward the corresponding BCRLBs, indicating that the proposed method meets tracking accuracy requirements across all target groups. When sufficient transmit power is available, the BSMRM strategy drives the accuracy to the desired thresholds, thereby maximizing utility. (Fig. 5) compares the normalized average power consumption of three strategies. (Figs. 6 and 7) show resolution performance curves for the BSMRM and JBPAWLS strategies, respectively. (Fig. 8) compares spectrum occupancy across two operational phases to assess spectral efficiency under resolution constraints. Experiment 4 evaluates the computational complexity of the BSMRM strategy. Results show that, with parallel computing, the strategy satisfies the real-time requirements of resource scheduling.  Conclusions  This study proposes a BSMRM strategy for MTT in a C-MIMO radar network. The strategy jointly optimizes beam selection, transmit power, signal bandwidth, and pulse length to improve both target resolvability and tracking accuracy. An efficient three-stage solver is developed to address the resulting optimization problem. Simulation results show that the proposed strategy achieves high-resolution MTT while reducing spectral occupancy. Under sufficient power conditions, tracking accuracy is comparable to that of state-of-the-art algorithms. In the same simulation environment, the average CPU runtime per frame is 2.8617 s, which is reduced to 0.3573 s with GPU-based parallel processing—meeting the real-time requirements for resource allocation. Moreover, the BSMRM strategy demonstrates flexibility and robustness under limited power constraints.
Wireless Communication and Internet of Things
Study on Time Slot Allocation and Monitoring Performance of UAT2 Data Link Status Bitmap
TANG Xinmin, TANG Shengjia, WEN Jingyu, GU Junwei
Available online  , doi: 10.11999/JEIT250251
Abstract:
  Objective  With the advancement of Urban Air Mobility (UAM), the rapid growth in aircraft numbers under such scenarios requires improved time slot allocation algorithms for the Universal Access Transceiver 2 (UAT2) data link to enhance surveillance capacity. This study analyzes the original time slot allocation algorithm for UAT2, identifying limitations related to downlink message time slot management. An improved allocation algorithm based on a state bitmap with a random drift mechanism is proposed considering message characteristics. Additionally, three transmission schemes to expand the number of time slots are proposed, and a surveillance capacity calculation model is established. The performance of each transmission scheme under different slot allocation strategies is simulated and evaluated. The research addresses the challenge of insufficient surveillance capacity in UAT2 data link under high-density UAM scenarios and provides an optimized approach for time slot allocation and surveillance performance improvement.  Methods  The study begins with an analysis of the original UAT2 time slot allocation algorithm, which is limited in high-density aircraft environments due to its pseudo-random mechanism based on geographic coordinates. The proposed algorithm introduces a state bitmap, implemented as a bit vector table where each bit indicates whether the corresponding time slot is occupied. When an aircraft selects a time slot, it first generates a pseudo-random number using the algorithm specified in DO-282C MOPS. The state bitmap is then checked: if the corresponding bit is 0, the time slot is selected; if 1, a piecewise random drift mechanism adjusts the slot selection, as shown in Formula (17). The drift mechanism segments the time slots to distribute selections more evenly and reduce collision probability. The algorithm also applies a slot retention period T, allowing an aircraft to occupy the same time slot for T consecutive UAT frames before reselecting, as defined by Formulas (19)–(21). To further expand time slot availability, three transmission schemes are proposed: Multi-ary Continuous Phase Frequency Shift Keying (MCPFSK) with modulation orders M = 4 and 8, constant modulation index with increased symbol rate, and constant carrier frequency difference with reduced modulation index. The available number of time slots for each scheme is calculated using Formula (26). A surveillance capacity model incorporating bit error rate and collision probability is established, as expressed by Formulas (27)–(30).  Results and Discussions  Simulation results demonstrate that the improved algorithm substantially outperforms both the original algorithm and the fixed time-window dynamic slot allocation algorithm. Under the original transmission scheme, the improved algorithm reduces slot collision probability by 16.78% and increases slot utilization by 16.12% compared to the original algorithm (Table 3). Relative to the algorithm described in [Ref. 19], the collision probability decreases by 10.80%, and slot utilization increases by 10.48%. For the expanded time slot schemes, when maximum surveillance capacity is reached, the improved algorithm reduces collision probability by 16.14% and increases slot utilization by 16.13% relative to the original algorithm (Table 3). Among these schemes, the 8CPFSK expansion achieves the highest surveillance capacity of 3913, with a slot utilization rate of 79.37% (Table 3). Real-time performance testing indicates that even in high-density scenarios, the improved algorithm maintains scheduling times within 120 ms, meeting the real-time operational requirements of UAT2 (Fig. 10). Bit Error Rate (BER) simulations reveal that the MCPFSK scheme provides superior anti-interference performance, whereas the constant carrier frequency difference scheme exhibits the highest BER (Fig. 11).  Conclusions  A slot allocation algorithm for the UAT2 data link based on state bitmaps and a random drift mechanism is proposed in this study. Compared to the original algorithm, the proposed method reduces slot collision probability by 16.78% and improves slot utilization by 16.12% under the original transmission scheme. When compared to the fixed time-window dynamic slot allocation algorithm described in [Ref. 19], collision probability decreases by 10.80%, and slot utilization increases by 10.48%. Three transmission schemes designed to expand slot availability are also proposed. Simulation results show that as the number of available slots increases, the performance advantage of the improved algorithm becomes more pronounced. Across all transmission schemes, when maximum surveillance capacity is reached, the improved algorithm reduces slot collision probability by approximately 16% and increases slot utilization by approximately 17%. In addition, a surveillance capacity calculation model for the UAT2 data link is established. Quantitative simulation results based on slot collision probability and BER performance confirm that the effective surveillance capacity of the improved algorithm is significantly higher than that of both the original algorithm and the fixed time-window dynamic slot allocation algorithm. It is also demonstrated that the proposed algorithm achieves optimal surveillance performance when all aircraft are equipped with both transmission and reception capabilities. Future research will focus on optimizing slot allocation algorithms for scenarios where aircraft possess only transmission capability.
Security and Reliability-Optimal Offloading for Mobile Edge Computing in Low-latency Industrial IoT
HU Yulin, YU Xinlan, GAO Wei, ZHU Yao
Available online  , doi: 10.11999/JEIT250262
Abstract:
  Objective  With the advancement of Industry 4.0, the increasing demand for mission-critical and latency-sensitive applications is driving the evolution of Industrial Internet of Things (IIoT) networks. Mobile Edge Computing (MEC) has emerged as a key solution to alleviate the computational constraints of User Equipment (UE), thus enhancing IIoT system performance. However, ensuring high reliability and security remains a critical challenge in multi-access MEC networks. To support Ultra-Reliable and Low-Latency Communication (URLLC), short packet transmission is employed, introducing Finite BlockLength (FBL) effects that impacts communication performance. Additionally, queuing delays at the servers present a significant bottleneck in meeting latency requirement. While existing studies predominantly focus on secure transmission under the Infinite BlockLength (IBL) assumption, comprehensive investigations into Physical Layer Security (PLS) and reliability in the FBL regime remain scarce. To address this gap, this paper aims to jointly optimize the security and reliability of MEC networks. Specifically, an analytical expression for end-to-end Leakage Failure Probability (LFP) is derived to assess the security and reliability performance. Furthermore, an efficient two-phase time length allocation algorithm is proposed to enhance overall system performance.  Methods  This study examines the uplink URLLC scenario in a multi-access MEC network. First, leveraging FBL theory, the communication phase within frame structure is modeled, and the decoding error probabilities of both the server and the eavesdropper are characterized. Subsequently, Extreme Value Theory (EVT) is employed to model the delay violation probability in the computation phase. Based on these formulations, the end-to-end LFP is derived to comprehensively assess the system’s PLS and reliability. Given the stringent end-to-end latency constraints in URLLC, a fundamental trade-off between security and reliability arises. To investigate this trade-off, an optimization problem is formulated to minimize LFP while satisfying the latency constraint, with the time length of the communication and computation phases as the optimization variables. However, due to the presence of multiplicative non-convex Q-functions in the objective function, solving this optimization problem poses significant challenges. To overcome this, a local point expansion technique is employed to approximate the non-convex terms as summations, facilitating problem reformulation. Additionally, auxiliary variables are introduced to enable convex relaxation of the Q-functions, transforming the problem into a convex optimization framework. The reformulated problem is then efficiently solved using the Successive Convex Approximation (SCA) algorithm. In response to the issue of performance degradation in algorithm convergence when the task scale increases, an Accelerated-SCA (A-SCA) algorithm is introduced. By leveraging the problem structure, a closed-form expression for the time allocation strategy corresponding to the system’s optimal reliability is derived, which improves the initial point selection mechanism. Finally, simulation results validate the accuracy, convergence, and adaptability of the proposed algorithm, while also revealing the trade-off between security and reliability.  Results and Discussions  The proposed FBL coding-based time allocation scheme is evaluated through numerical simulations to validate its effectiveness. A comparative analysis with the IBL coding-based scheme is conducted, demonstrating the performance gains achieved through the joint optimization of communication and computation phases, while highlighting the limitations of IBL coding. Furthermore, the quality of the sub-optimal solution obtained via the SCA algorithm is assessed by benchmarking against the exhaustive search method. Additionally, both constrained and unconstrained resource scenarios are considered to explore the trade-offs between security and reliability, as well as between time resource and LFP. First, (Fig. 2) analyzes the convergence behavior of the proposed SCA-based time allocation scheme, demonstrating that the sub-optimal solution obtained closely approximates the global optimum. Furthermore, by optimizing the initial point selection with the A-SCA algorithm, the convergence speed and scalability are significantly enhanced. (Fig. 3) evaluates the variation of LFP with the eavesdropper’s channel gain under different task data sizes, revealing the fundamental trade-off between security and reliability, along with the associated performance compensation mechanism. (Fig. 4) illustrates the LFP trend as a function of the server’s channel gain, comparing the proposed scheme with IBL coding-based scheme. The results indicate that even in scenarios with low secure channel capacity, the proposed scheme achieves a 21.4% performance improvement, thereby validating its superiority in joint optimization across both phases. (Fig. 5) investigates the impact of task workload on LFP, while (Fig. 6) explores an unconstrained resource scenario, analyzing the influence of transmit power on LFP and the corresponding service time requirements.  Conclusions  This study explores the joint optimization of security and reliability in multi-access MEC networks supporting URLLC. In the FBL regime, the decoding error probability is modeled in the communication phase, while the delay violation probability is characterized using EVT in the computation phase. Based on these models, the LFP is derived as a key metric to quantify the system’s end-to-end PLS and reliability. Given the stringent latency constraints, a two-phase time length allocation optimization problem is formulated to minimize LFP, balancing security and reliability. To address the problem’s non-convexity, this paper first decouples the coupled variables through local linearization and reconstructs the problem. Next, convex upper and lower bounds of the complex Q-function are derived, enabling convex relaxation of the objective function. Subsequently, the SCA algorithm is applied iteratively to obtain a sub-optimal solution for the original problem. To tackle the issue of deteriorating convergence performance and increased complexity as the task scale grows, an A-SCA algorithm is proposed, which optimizes the initial point selection to address this challenge. Finally, simulation results validate the convergence, accuracy, and scalability of the proposed algorithm, demonstrating its performance advantages through comparison with existing approaches. Additionally, the impact of key system parameters on performance is systematically analyzed. The proposed LFP metric exhibits significant extensibility, being applicable to collaborative optimization of physical layer security and network layer security in URLLC systems, reliable task offloading in MEC, and throughput optimization in MIMO systems. Furthermore, the solution framework for LFP optimization problem, including problem decoupling and reconstruction, introduction of auxiliary variables for Q-function convex relaxation, iterative solving via SCA, and optimized initial point selection, provides valuable insights for solving multi-variable non-convex optimization problems in the FBL regime.
Joint Beamforming and Antenna Position Optimization in Movable Antenna Empowered ISAC Systems
LI Zhendong, BA Jianle, SU Zhou, ZHAO Weichun, CHEN Wen, ZHU Zhengyu
Available online  , doi: 10.11999/JEIT250146
Abstract:
  Objective  With the rapid advancement of information technology, mobile communication is transitioning from the fifth generation (5G) to the sixth generation (6G), with a central goal of evolving from the Internet of Things to the Internet of Intelligence. Integrated Sensing and Communication (ISAC) has emerged as a key technology for next-generation wireless systems. By jointly designing sensing and communication functionalities, ISAC substantially improves spectral efficiency and overall system performance, and is regarded as a core technology for future intelligent networks. However, conventional fixed antenna arrays are limited by insufficient spatial degrees of freedom, making them inadequate for meeting the stringent sensing and communication requirements in dynamic and complex environments. To address this challenge, Movable Antenna (MA) is proposed as a novel architecture. MA systems enable antenna elements to move within designated transmit and receive regions, allowing real-time adjustment of their positions according to instantaneous channel states and system demands. This study proposes an MA-empowered ISAC framework that dynamically reconfigures antenna positions to exploit the spatial characteristics of wireless channels, with the objective of minimizing transmit power. The proposed design contributes new insights into energy-efficient ISAC system development and offers theoretical and practical relevance for future wireless communication networks.  Methods  This study formulates a joint optimization problem that considers the discrete positions of MA elements, beamforming vectors, and sensing signal covariance matrices. To address the inherent coupling among optimization variables and the presence of binary discrete variables, a Discrete Binary Particle Swarm Optimization (BPSO) framework is adopted. The algorithm iteratively determines the discrete antenna positions through a fitness-based search. Based on the obtained positions, Semi-Definite Relaxation (SDR) and Successive Convex Approximation (SCA) techniques are applied to manage non-convex constraints and solve for the beamforming vectors and sensing covariance matrices. This approach yields suboptimal yet effective solutions in complex non-convex optimization scenarios, thereby enhancing system performance. In terms of system modeling, a Dual-Function Radar and Communication Base Station (DFRC BS) equipped with MA is considered. The DFRC BS transmits downlink ISAC signals, communicates with multiple single-antenna downlink users, and senses a radar target. Antenna elements are restricted to a discrete set of candidate positions, facilitating practical system deployment. The radar sensing channel is modeled as a Line-of-Sight (LoS) channel, whereas the communication channels follow a field-response-based model. The DFRC BS transmits narrowband ISAC signals through beamforming, which are used concurrently for radar target detection and downlink communication.  Results and Discussions  Simulation results demonstrate that the proposed MA-empowered ISAC system achieves superior performance in transmission power optimization compared with conventional fixed antenna array systems. Under the constraint of sensing Signal-to-Interference-plus-Noise Ratio (SINR), transmit power increases with more stringent sensing requirements. However, the MA-empowered ISAC system substantially lowers the required transmit power, with a maximum observed reduction of 101.1 W. Under the communication SINR constraint and fixed sensing requirements, transmit power also rises as communication demands grow. In this setting, the MA-empowered ISAC system again shows a clear advantage, reducing transmit power by up to 134.6 W compared with traditional systems. Furthermore, as the number of downlink users increases—while maintaining consistent sensing and communication requirements—transmit power increases accordingly. Even under these conditions, the MA-empowered system continues to outperform its fixed-array counterpart. Beamforming pattern simulations further confirm that the BPSO-based optimization framework achieves multi-beam alignment and provides a degree of interference suppression, highlighting its effectiveness in managing dual-function transmission within MA-enabled ISAC systems. Overall, the introduction of MA significantly enhances ISAC system performance in terms of energy efficiency and spatial adaptability.  Conclusions  This paper proposes a joint optimization framework based on Discrete BPSO for MA-empowered ISAC systems, which effectively reduces transmit power while maintaining sensing and communication performance. The optimization process begins by iteratively solving a fitness function to determine the discrete positions of the MA elements. SDR and SCA techniques are employed during this process to address non-convex constraints. Once the MA element positions are established, the associated beamforming vectors and sensing signal covariance matrices are computed or further optimized. By dynamically adjusting antenna positions, the system gains additional spatial degrees of freedom, enabling more efficient utilization of the spatial characteristics of wireless channels. This work provides a new approach for energy-efficient design in ISAC systems and offers meaningful guidance for the development of next-generation wireless networks. Future research will continue to explore advanced optimization algorithms for MA positioning and system performance enhancement, aiming to meet sensing and communication requirements in highly dynamic environments. Additionally, the applicability of MA in other wireless scenarios—such as vehicular networks and aerial communication platforms, will be studied to facilitate broader adoption of MA-based technologies.
Cross-Layer Collaborative Resource Allocation in Maritime Wireless Communications: QoS-Aware Power Control and Knowledge-Enhanced Service Scheduling
ZHANG Zhilin, MAO Zhongyang, LU Faping, PAN Yaozong, LIU Xiguo, KANG Jiafang, YOU Yang, JIN Yin
Available online  , doi: 10.11999/JEIT250252
Abstract:
  Objective  Maritime wireless communication networks face significant challenges, including dynamic topology drift, large-scale channel fading, and cross-layer resource competition. These factors hinder the effectiveness of traditional single-layer resource allocation methods, which struggle to maintain the balance between high-quality communications and heterogeneous service demands under limited network resources. This results in degraded Quality of Service (QoS) and uneven service guarantees. To address these challenges, this study proposes a cross-layer collaborative resource allocation framework that achieves balanced enhancement of system throughput and QoS assurance through closed-loop optimization, integrating physical-layer power control with network-layer service scheduling. First, a cross-layer wireless network transmission model is established based on the coupling mechanism between physical-layer channel capacity and transport-layer TCP throughput. Second, a dual-threshold water-level adjustment mechanism, incorporating both Signal-to-Noise Ratio (SNR) and QoS metrics, is introduced into the classical water-filling framework, yielding a QoS-aware dual-threshold water-filling algorithm. This approach strategically trades controlled throughput loss for improved QoS of high-priority services. Furthermore, a conflict resolution strategy optimization filter with dual-channel feature decoupling is designed within a twin deep reinforcement learning framework to enable real-time, adaptive node-service dynamic matching. Simulation results demonstrate that the proposed framework improves average QoS scores by 9.51% and increases critical service completion by 1.3%, while maintaining system throughput degradation within 10%.  Methods  This study advances through three main components: theoretical modeling, algorithm design, and system implementation, forming a comprehensive technical system. First, leveraging the coupling relationship between physical-layer channel capacity and transport-layer Transmission Control Protocol (TCP) throughput, a cross-layer joint optimization model integrating power allocation and service scheduling is established. Through mathematical derivation, the model reveals the nonlinear mapping between wireless resources and service demands, unifying traditionally independent power control and service scheduling within a non-convex optimization structure, thus providing a theoretical foundation for algorithm development. Second, the proposed dynamic dual-threshold water-filling algorithm incorporates a dual-regulation mechanism based on SNR and QoS levels. A joint mapping function is designed to enable flexible, demand-driven power allocation, enhancing system adaptability. Finally, a twin deep reinforcement learning framework is constructed, which achieves independent modeling of node mobility patterns and service demand characteristics through a dual-channel feature decoupling mechanism. A dynamic adjustment mechanism is embedded within the strategy optimization filter, improving critical service allocation success rates while controlling system throughput loss. This approach strengthens system resilience to the dynamic, complex maritime environment.  Results and Discussions  Comparative ablation experiments demonstrate that the dynamic dual-threshold water-filling algorithm within the proposed framework achieves a 9.51% improvement in QoS score relative to conventional water-filling methods. Furthermore, the Domain Knowledge-Enhanced Siamese DRL (DKES-DRL) method exceeds the Siamese DRL approach by 3.25% (Fig. 6), albeit at the expense of a 9.3% reduction in the system’s maximum throughput (Fig. 7). The average number of completed transactions exceeds that achieved by the traditional water-filling algorithm by 1.3% (Fig. 8, Fig. 9). In addition, analysis of the effect of node density on system performance reveals that lower node density corresponds to a higher average QoS score (Fig. 10), indicating that the proposed framework maintains service quality more effectively under sparse network conditions.  Conclusions  To address the complex challenges of dynamic topology drift, multi-scale channel fading, and cross-layer resource contention in maritime wireless communication networks, this paper proposes a cross-layer collaborative joint resource allocation framework. By incorporating a closed-loop cross-layer optimization mechanism spanning the physical and network layers, the framework mitigates the imbalance between system throughput and QoS assurance that constrains traditional single-layer optimization approaches. The primary innovations of this work are reflected in three aspects: (1) Cross-layer modeling is applied to overcome the limitations of conventional hierarchical optimization, establishing a theoretical foundation for integrated power control and service scheduling. (2) A dual-dimensional water-level adjustment mechanism is proposed, extending the classical water-filling algorithm to accommodate QoS-driven resource allocation. (3) A knowledge-enhanced intelligent decision-making system is developed by integrating model-driven and data-driven methodologies within a deep reinforcement learning framework. Simulation results confirm that the proposed framework delivers robust performance in dynamic maritime channel conditions and heterogeneous traffic scenarios, demonstrating particular suitability for maritime emergency communication environments with stringent QoS requirements. Future research will focus on resolving engineering challenges associated with the practical deployment of the proposed framework.
Graph-structured Data-driven Topology Inference for Non-cooperative Clustered Wireless Communication Networks
HOU Changbo, FU Dingyi, SONG Zhen, WANG Bin, ZHOU Zhichao
Available online  , doi: 10.11999/JEIT250084
Abstract:
  Objective  The emergence of clustered target communication networks complicates electromagnetic environment detection in non-cooperative scenarios, creating challenges for electromagnetic situation awareness and electronic countermeasures. Existing research seldom addresses topology prediction under conditions with no prior knowledge, where the absence of explicit structural information and the dynamic nature of the networks hinder accurate inference. This study investigates topology prediction for non-cooperative clustered wireless communication networks using graph-structured data-driven approaches. Specifically, it evaluates the performance of multiple topology inference methods, including the Multivariate Hawkes Process (MHP), Peter-Clark Momentary Conditional Independence (PCMCI), Graph Encoder-Decoder (GED), and Graph Convolutional Networks (GCN). The effects of network properties such as node count and edge probability on inference accuracy are analyzed. Additionally, a hybrid framework that integrates statistical models with graph-based learning is explored to improve inference accuracy and computational efficiency.  Methods  The proposed methodology combines causal inference with Graph Neural Network (GNN)-based learning. Adjacency matrices are first generated through causal discovery, using time-domain matrices derived from simulated wireless communication events. These matrices are constructed by thresholding power spectra to yield binary communication states. The GNN module subsequently refines the causal discovery output by suppressing false positives and optimizing global topology through encoder-decoder operations with multi-head attention mechanisms. To assess robustness, synthetic datasets are generated with NS-3 simulations under varying conditions: edge probabilities (0.15~0.60), node densities (8~13 nodes), sampling durations (0.05~0.30 ms), and node feature completeness (partial, 50%; full, 100%). Connectivity patterns are modeled by incorporating distance-adjusted edge probabilities. Performance evaluation uses F1-score, accuracy, recall, and inference time, with systematic comparison across baseline models (MHP, PCMCI, GCN, GED) and hybrid variants (PCMCI+GED, MHP+GED).  Results and Discussions  The PCMCI+GED hybrid framework consistently achieves superior topology prediction across diverse network configurations. At an edge probability of 0.45, PCMCI+GED with full node features attains an F1-score of 0.808, exceeding the performance of standalone PCMCI and GED by 31.1% and 4.9%, respectively (Fig. 7). This improvement arises from the synergy between causal priors and graph neural networks: PCMCI establishes preliminary causal relationships, while GED refines inference through global attention mechanisms that reduce false positives. Comparative analysis reveals that richer node features enhance topology inference in causal inference methods (Fig. 7). For example, MHP+GED with full features exceeds its 50% feature counterpart by 2.10%, and PCMCI+GED with full features improves by 3.04%. Yet, the most substantial gains come from combining causal inference with GED. Relative to standalone MHP with full features, MHP+GED improves by 30.65% with 50% features and 33.40% with full features. Similarly, PCMCI+GED improves by 34.43% and 38.51% under the same conditions. In contrast, relying solely on GNNs proves insufficient for modeling causal relationships. GED alone performs similarly to GCN, with Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE) values of 0.0149 and 0.0206, respectively (Fig. 8). Without causal priors, GED offers no significant advantage over GCN; however, when priors are incorporated, GED outperforms GCN in inference accuracy (Fig. 9). Sampling duration analysis shows that 0.1 ms is optimal, balancing redundancy and information loss (Fig. 10, Table 2). Shorter intervals (0.05 ms) inflate computational costs through high-dimensional operations, whereas longer intervals (0.30 ms) obscure temporal dependencies, reducing the F1-score of PCMCI+GED with full features to 40.57% of its value at 0.1 ms. Efficiency evaluations highlight trade-offs between accuracy and runtime. With 50% node features, PCMCI+GED reduces inference time by 88.63% while retaining 96.96% of its F1-score. Under the same conditions, MHP+GED reduces inference time by 85.48% with only a 2.07% drop in performance (Fig. 11). PCMCI’s exponential complexity makes it computationally prohibitive in high-dimensional settings, whereas MHP’s quadratic scaling with node count and linear scaling with event frequency yield more modest efficiency gains. In low-dimensional settings, however, MHP’s event-driven computation leads to longer runtimes than PCMCI. Heatmap analysis further confirms the precision of the hybrid models. Adjacency matrices generated by PCMCI+GED and MHP+GED with full features closely align with the ground truth, demonstrating high predictive accuracy (Fig. 9). In sparse networks, standalone PCMCI introduces noise by linking non-interacting nodes, while GCN generates fragmented predictions due to the absence of causal priors. The hybrid framework alleviates these limitations by combining PCMCI’s local causal inference with GED’s global optimization. Overall, the hybrid framework addresses key shortcomings of individual methods: the high computational cost of PCMCI and MHP, and the limited interpretability of GNNs. By integrating causal discovery with graph-based deep learning, the model achieves state-of-the-art predictive accuracy while maintaining scalability. Its performance highlights the potential for real-time applications in resource-constrained environments, emphasizing the importance of balancing causal priors and data-driven learning for advancing non-cooperative wireless communication network analysis.  Conclusions  This study demonstrates the effectiveness of combining PCMCI-based causal inference with GED-enhanced GNN for topology prediction in non-cooperative clustered wireless communication networks. The hybrid model achieves state-of-the-art accuracy, particularly in dense networks, while partial node feature inputs substantially reduce computational overhead. Although the exponential complexity of PCMCI constrains scalability in high-dimensional settings, integration with GED alleviates this limitation through feature reduction and global optimization. The findings highlight the need to balance accuracy and efficiency in practical applications, where GCN offers a viable option for real-time inference. Future research will explore attention mechanisms and self-supervised learning to further enhance robustness. These advancements hold promise for improving electromagnetic situation awareness and electronic countermeasure strategies in dynamic adversarial environments.
Pareto Optimization of Sensing and Communication Performance of Near-field Integrated Sensing and Communication System
ZHANG Guangchi, XIE Zhili, CUI Miao, WU Qingqing
Available online  , doi: 10.11999/JEIT250231
Abstract:
  Objective  With the rapid development of Sixth-Generation (6G) communication technology, Integrated Sensing And Communication (ISAC) systems are regarded as key enablers of emerging applications such as the Internet of Things, smart cities, and autonomous driving. High-precision communication and sensing are required under limited spectrum resources. However, most existing studies concentrate on the far-field region, where incomplete derivation of the sensing mutual information metric, neglect of scatterer interference, and insufficient consideration of communication-sensing trade-offs limit the flexibility of beamforming design and reduce practical effectiveness. As application scenarios expand, the demand for efficient integration of communication and sensing becomes more pronounced, particularly in near-field environments where scatterer interference strongly affects system performance In this work, beamforming design for near-field ISAC systems under scatterer interference is investigated. A general expression for sensing mutual information is derived, a multi-objective optimization problem is formulated, and auxiliary variables, the Schur complement, and the Dinkelbach algorithm are employed to obtain Pareto optimal solutions. The proposed method provides a flexible and effective approach for balancing communication and sensing performance, thereby enhancing overall system performance and resource utilization in diverse application scenarios. The findings serve as a valuable reference for the optimal trade-off design of communication and sensing in near-field ISAC systems.  Methods  The proposed beamforming design method first derives a general expression for sensing mutual information in near-field scenarios, explicitly accounting for and quantifying the effect of scatterer interference on sensing targets. A multi-objective optimization problem is then formulated, with the Signal-to-Interference-plus-Noise Ratio (SINR) of communication users and sensing mutual information as objectives. Within this multi-objective framework, communication and sensing performance can be flexibly balanced to satisfy the requirements of different application scenarios. To enable tractable optimization, the sensing mutual information expression is transformed into a Semi-Definite Programming (SDP) problem using auxiliary variables and the Schur complement. Multi-user SINR expressions are reformulated with the Dinkelbach algorithm to convert them into convex functions, facilitating efficient optimization. The multi-objective problem is subsequently reduced to a single-objective one by constructing a system utility function, and the Pareto optimal solution is obtained to achieve the optimal balance between communication and sensing performance. This method provides a flexible and effective design strategy for near-field ISAC systems, substantially enhancing overall system performance and resource utilization.  Results and Discussions  This study presents a beamforming design method that balances communication and sensing performance through innovative optimization strategies. The method derives the general expression of sensing mutual information under scatterer interference, formulates a multi-objective optimization problem with the SINR of communication users and sensing mutual information as objectives, and transforms the problem into a convex form using auxiliary variables, the Schur complement, and the Dinkelbach algorithm. The Pareto optimal solution is then obtained via a system utility function, enabling the optimal balance between communication and sensing performance. Simulation results demonstrate that adjusting the weight parameter ρ flexibly balances user communication and target sensing performance (Fig. 2). As ρ increases from 0 to 1, sensing mutual information rises while user rate decreases, showing that a controllable trade-off can be achieved by tuning weights. In multi-user scenarios, near-field ISAC systems exhibit superior performance compared with far-field systems (Fig. 3). Under near-field conditions, the proposed method achieves more flexible and adjustable trade-offs than the classic Zero-Forcing (ZF) algorithm and single-objective optimization algorithms (Fig. 4, Fig. 5), confirming its effectiveness and superiority in practical applications. Furthermore, the study reveals the interference pattern of scatterers on sensing targets with respect to distance (Fig. 6). The results indicate that the greater the distance difference between a scatterer and a sensing target, the weaker the interference on the target, with sensing mutual information gradually increasing and eventually converging. This finding provides a valuable reference for the design of near-field ISAC systems.  Conclusions  This paper proposes a beamforming design method for balancing communication and sensing performance by jointly optimizing sensing mutual information and communication rate. The method derives the general form of sensing mutual information, reformulates it as an SDP problem, and applies the Dinkelbach algorithm to process multi-user SINR expressions, thereby establishing a multi-objective optimization framework that can flexibly adapt to diverse application requirements. The results demonstrate three key findings: (1) The method enables flexible adjustment of communication and sensing performance, achieving an optimal trade-off through weight tuning, and allowing dynamic adaptation of system performance to specific application needs. (2) It reveals the interference pattern of scatterers on sensing targets with respect to distance, providing critical insights for near-field ISAC system design and supporting optimized system layout and parameter selection in complex environments. (3) In multi-user scenarios, the proposed approach outperforms traditional single-objective optimization methods in both communication rate and sensing mutual information, highlighting its competitiveness and practical value.
Unmanned Aircraft Vehicle-assisted Multi Cluster Concurrent Authentication Scheme for Internet of Things Devices
MA Ruhui, HE Shiyang, CAO Jin, LIU Kui, LI Hui, QIU Yuan
Available online  , doi: 10.11999/JEIT250279
Abstract:
  Objective  With the rapid expansion of Internet of Things (IoT) devices, application scenarios such as smart cities and industrial intelligent manufacturing demand wider coverage and higher connection density from communication systems. Traditional terrestrial base stations have limited capacity to serve IoT devices in remote or complex environments. Unmanned Aircraft Vehicles (UAVs), owing to their flexible deployment and high mobility, can function as aerial base stations that effectively complement terrestrial networks, providing reliable and energy-efficient access for remote IoT terminals. Additionally, the expected 6G connectivity of tens of billions of devices may give rise to signaling conflicts and congestion at key nodes. To address these challenges, multi-cluster access schemes based on cluster division have been proposed. In these schemes, different clusters connect simultaneously to orthogonal subchannels, enabling UAVs to assist multiple IoT device clusters in accessing terrestrial networks concurrently. However, UAV-assisted multi-cluster communication faces pressing security and performance issues, including the susceptibility of air interface channels to attacks, the limited computational and storage capacities of IoT devices, signaling conflicts arising from massive concurrent access, and the requirement for seamless handover mechanisms due to the restricted endurance of UAVs. Therefore, the development of a secure and efficient UAV-assisted multi-cluster concurrent access and handover authentication scheme is essential.  Methods  This study proposes a secure authentication scheme for the UAV-assisted multi-cluster IoT device communication model, comprising four main components. First, UAV access authentication is achieved through a traditional pre-shared key mechanism, enabling mutual authentication and key agreement between the UAV and the ground network. Second, concurrent access authentication for multi-cluster IoT devices is realized using multi-layer aggregated signaling and aggregated message authentication code technologies, which effectively mitigate signaling conflicts and node congestion during massive concurrent access. Meanwhile, a Physically Unclonable Function (PUF) mechanism is incorporated to strengthen device-level security, protecting IoT devices against physical attacks while maintaining low storage and computational requirements. Finally, the UAV-assisted concurrent handover authentication integrates multi-layer aggregated signaling, aggregated message authentication code, and a pre-distribution key mechanism to enable fast and secure handovers between multi-cluster IoT devices and new UAVs, thereby ensuring the continuous security of network services.  Results and Discussions  The security of the proposed scheme is validated through formal analysis with the Tamarin tool, complemented by informal security analysis. The results show that the scheme satisfies mutual authentication and data security, and resists replay and man-in-the-middle attacks. The signaling overhead, as well as the computational and storage requirements of IoT devices during concurrent access and handover in multi-cluster communication, are also evaluated. The findings indicate that the scheme generates minimal signaling overhead (Fig. 3), thereby preventing signaling conflicts and node congestion. Moreover, the computational cost on devices remains low (Fig. 4), and the storage demand is minimal (Fig. 5), demonstrating that the scheme is well suited for resource-constrained IoT devices.  Conclusions  This paper proposes a UAV-assisted authentication scheme for concurrent access and handover of multi-cluster IoT devices. In this scheme, UAVs can securely and efficiently access the ground network, while multi-cluster IoT devices achieve concurrent and secure access through UAVs and perform rapid authentication and key agreement during handover to a new UAV. Security and performance analyses demonstrate that the scheme ensures multiple security properties, including mutual authentication, data security, and resistance to replay, man-in-the-middle, and physical attacks, while maintaining low computational and storage overhead on IoT devices. In addition, the scheme features low signaling overhead, effectively preventing signaling conflicts and key node congestion during large-scale concurrent access. Nevertheless, some limitations remain. Future work will explore more comprehensive and practical authentication mechanisms. Specifically, lightweight dynamic key update mechanisms tailored to UAV communication scenarios will be investigated to enhance security with minimal overhead. To address design complexity and environmental adaptability issues caused by PUF hardware dependence, more robust hardware security mechanisms will be considered to improve system stability in complex environments. Moreover, to mitigate the computational and energy burden on UAVs resulting from aggregation and forwarding tasks, approaches such as edge computing offloading will be examined to enable dynamic task allocation and load balancing, ensuring efficient and sustainable operation. Finally, a prototype system will be developed, and field experiments will be conducted to validate the feasibility and performance of the proposed solution in real-world scenarios.
Dual-Reconfigurable Intelligent Surface Phase Shift Optimization and Unmanned Aerial Vehicle Trajectory Control for Vehicle Communication
CHANG Kuan, ZHANG Lei, WANG Yu, SHANG Yulong, CHEN Weicong, MA Junchao
Available online  , doi: 10.11999/JEIT250274
Abstract:
This study considers a scenario in which an Unmanned Aerial Vehicle (UAV) equipped with a Reconfigurable Intelligent Surface (RIS) cooperates with a fixed RIS to enhance communication with a mobile User Equipment (UE) vehicle. A joint optimization problem is formulated to maximize the UE’s communication rate by controlling the UAV’s flight trajectory and the phase shifts of both RISs. Given the system complexity and environmental dynamics, a solution is proposed that integrates a Deep Deterministic Policy Gradient (DDPG) algorithm with phase-shift alignment to optimize continuous UAV trajectories and RIS configurations. Simulation results confirm that the proposed method achieves stable reward convergence within 1,000 training episodes. Compared with benchmark approaches, the algorithm improves communication rates by at least 3 dB over random trajectory and phase-shift strategies in dual-RIS deployments. The study further presents optimal UAV trajectories under varying base station and RIS placements and evaluates algorithm performance across different vehicle speeds.  Objective   This study investigates a vehicular communication scenario in which a UAV-mounted RIS cooperates with a fixed RIS to assist a mobile UE device. A joint optimization framework is established to maximize UE communication rates during movement by simultaneously optimizing the UAV trajectory and the phase shifts of both RISs. To address system complexity and environmental dynamics, a DDPG algorithm is employed for continuous trajectory control, while a low-complexity phase-shift alignment method configures the RISs. Simulation results show that the proposed algorithm achieves stable reward convergence within 1,000 training episodes and improves communication rates by at least 3 dB compared with randomized trajectory and phase-shift baselines. It also outperforms alternative reinforcement learning approaches, including Twin Delayed Deep Deterministic policy gradient (TD3) and Soft Actor-Critic (SAC). Optimal UAV trajectories are derived for various base station and RIS deployment scenarios, with additional simulations confirming robustness across a range of vehicle speeds.  Methods   This study establishes a Multiple-Input Single-Output (MISO) system in which a UAV-mounted RIS cooperates with a fixed RIS to support mobile vehicular communication, with the objective of maximizing user information rates. To address the complexity of continuous trajectory control under dynamic environmental conditions, a DDPG-based algorithm is developed. The phase shifts of RIS elements are optimized using a low-complexity alignment method. A reward function based on the achievable information rates of vehicular users is designed to guide the agent’s actions and facilitate policy learning. The proposed framework enhances adaptability by dynamically optimizing UAV trajectories and RIS configurations under time-varying channel conditions.  Results and Discussions   (1) The convergence behavior of the DDPG algorithm is verified in Fig. 3, where the reward values progressively converge as the number of training episodes increases. (2) Fig. 4 shows the effect of varying the number of RIS elements on system performance, indicating that additional elements lead to a steady increase in reward values, confirming the channel gain enhancement provided by RIS deployment. (3) As shown in Fig. 5, the DDPG algorithm outperforms baseline methods and demonstrates greater adaptability to target scenarios; concurrently, optimized RIS phase shifts yield significantly higher rewards than random configurations, validating the proposed phase-alignment strategy. (4) Figs. 67 highlight notable variations in UAV trajectories and system performance across different base station and RIS deployments, demonstrating the adaptability of the trajectory optimization strategy. Fig. 8 further compares performance across scenarios with optimized UAV trajectories, highlighting the algorithm’s versatility. (5) System performance under different UE mobility speeds is evaluated in Fig. 9, showing a performance decline at higher speeds, indicating strong efficacy in low-speed environments but reduced effectiveness under high-speed conditions. These results collectively illustrate the operational strengths and limitations of the proposed framework in dynamic vehicular communication systems.  Conclusions   This paper investigates a vehicular communication scenario assisted by both fixed and UAV-mounted mobile RISs, aiming to maximize UE information rates under dynamic mobility conditions. A joint optimization framework is developed, combining dual-RIS phase shift alignment based on channel state information with UAV trajectory planning using a DDPG algorithm. The proposed method features a low-complexity design that addresses both network architecture and RIS configuration challenges. Extensive simulations under varying vehicular speeds, RIS element counts, and base station deployments demonstrate the algorithm’s superiority over SAC, TD3, and randomized phase shift strategies. Results further highlight the framework’s adaptability to heterogeneous base station–RIS topologies and reveal performance degradation at higher vehicle speeds, indicating the need for future research into real-time adaptive mechanisms.
Resilient Semantic Communication for Space-Air-Ground Networks
WANG Wenyuan, ZHOU Mingyu, WANG Chaowei, XU Jisong, ZHANG Yunze, PANG Mingliang, JIANG Fan, XU Lexi, ZHANG Zhi
Available online  , doi: 10.11999/JEIT250077
Abstract:
  Objective  Space-air-ground networks represent a fundamental evolution in wireless communication infrastructure by extending conventional terrestrial networks into airspace and outer space through the integration of satellites, unmanned aerial vehicles, and ground-based nodes. These networks have become a cornerstone technology for future sixth-generation wireless communications, providing wide-area coverage and flexible networking capabilities that overcome the limitations of geographically constrained terrestrial systems. However, image transmission over space-air-ground networks faces considerable challenges due to inherent bandwidth constraints, severe channel impairments, and highly dynamic propagation environments. Traditional Separate Source-Channel Coding (SSCC) methods, although theoretically sound, exhibit poor performance under these adverse conditions, particularly near information-theoretic limits. The rapid expansion of multimedia applications and service demands requires new approaches to information extraction and transmission that meet ultra-reliable low-latency communication requirements. Conventional image coding algorithms, such as JPEG and JPEG2000, experience significant performance degradation in satellite communication systems due to channel impairments. These limitations highlight the need for advanced semantic communication frameworks that move beyond conventional bit-level transmission and prioritize the preservation of semantic content integrity under complex wireless conditions.  Methods  A resilient semantic communication framework is proposed for image transmission in space-air-ground networks, based on enhanced information bottleneck theory. The key innovation is the development of an augmented rate-distortion function that jointly optimizes system capacity, pixel-level reconstruction fidelity, semantic feature preservation, and perceptual quality. The framework combines deep learning with information-theoretic principles to ensure reliable performance under varying channel conditions. A Gumbel-Softmax approach is integrated with variable rate networks to enable dynamic rate adaptation in response to fluctuating channel states. The system employs a weighted multi-asymmetric Gaussian distribution model to characterize the probability density functions of heterogeneous semantic features, providing accurate entropy estimation in the latent space. The architecture incorporates attention mechanisms and residual learning structures to enhance feature extraction and semantic preservation. A reconfigurable neural network architecture is introduced, featuring adaptive module selection driven by real-time signal-to-noise ratio assessments and service quality requirements. The encoder subsystem consists of semantic extractor networks that identify and isolate critical image features, joint source-channel encoders that perform integrated compression and error protection, and adaptive networks that generate binary mask vectors for intelligent symbol selection. The semantic extractor combines convolutional neural networks with fully connected layers to capture hierarchical feature representations. The joint source-channel coding architecture integrates residual convolution blocks, attention feature blocks, and residual transpose convolution blocks to optimize rate-distortion performance. The adaptive network produces dynamic masks through Gumbel-Softmax sampling, controlling the transmission of semantic symbols based on their relevance and channel state. A two-stage training strategy is implemented. First, end-to-end optimization of the joint source-channel encoders and decoders is performed using mean squared error loss. This is followed by full system training based on a composite loss function that jointly considers transmission rate, pixel-level distortion, semantic distortion, and perceptual quality.  Results and Discussions  Comprehensive experimental validation is conducted using the CIFAR-10, CIFAR-100, and Kodak24 datasets to assess the effectiveness of the proposed framework. Performance is evaluated using multiple metrics, including Peak Signal-to-Noise Ratio (PSNR) for objective image quality assessment and USC metric for overall system efficiency. Comparative analysis with conventional JPEG combined with LDPC and QAM schemes shows substantial performance gains, particularly under low Signal-to-Noise Ratio (SNR) conditions (Fig. 3, Fig. 4). The proposed resilient semantic communication framework consistently outperforms conventional methods across different compression ratios, demonstrating robust resistance to channel impairments and effectively mitigating the cliff effect observed in traditional SSCC systems. When compared with advanced deep learning-based approaches such as ADJSCC and DeepJSCC-V, the proposed method achieves significant improvements in both PSNR and transmission efficiency (Fig. 5, Fig. 6). Efficiency evaluation using USC metrics shows that the proposed framework achieves higher utility values across various compression settings, with performance advantages becoming more evident as SNR decreases (Fig. 7). Further analysis of different network configurations demonstrates the adaptability of the architecture in balancing computational complexity and transmission performance (Fig. 8, Fig. 9). Configurations with five network units consistently provide higher PSNR values compared to three- and four-unit designs. However, four-unit configurations achieve optimal efficiency under high SNR conditions, indicating effective resource allocation in response to varying channel states. Visual quality assessment using the Kodak dataset confirms improved reconstruction performance with reduced channel bandwidth requirements, supporting the practical feasibility of the proposed approach (Fig. 10). Computational complexity analysis shows that the five-unit configuration maintains complexity comparable to existing benchmarks, while three- and four-unit designs significantly reduce computational demands, supporting deployment on resource-constrained platforms (Table 4). These experimental results demonstrate that the proposed framework provides superior reconstruction quality, transmission efficiency, and adaptive capability across diverse and challenging wireless environments.  Conclusions  This study presents a resilient semantic communication framework that addresses key challenges in image transmission for space–air–ground networks. Experimental validation confirms that the proposed method achieves substantial improvements over conventional approaches in reconstruction quality, transmission efficiency, and adaptability.
SINR Adaptive Symbol Level Precoding and Position Joint Optimization Strategy for Multiple Unmanned Aerial Vehicles Anti-Jamming Communication
WEI Haoran, YAO Rugui, FAN Ye, MA Weixin, ZUO Xiaoya
Available online  , doi: 10.11999/JEIT250221
Abstract:
  Objective  Unmanned Aerial Vehicles (UAVs) are widely applied in areas such as traffic monitoring, wireless coverage, and precision agriculture due to their high mobility and deployment flexibility. In air–ground communication scenarios requiring flexible deployment, UAV mobility can be leveraged to counteract external malicious jamming. Further, the collaborative operation of multiple UAVs enables improved system performance. However, the broadcast nature of wireless communication renders multiple UAV communication systems vulnerable to jamming attacks that disrupt legitimate communication services. Addressing this challenge, this study proposes a Signal-to-Interference-plus-Noise Ratio (SINR) adaptive Symbol-Level Precoding (SLP) and position joint optimization strategy for anti-jamming communication in multi-UAV systems. The approach fully exploits UAV mobility to enhance communication robustness under different user requirements. By integrating Coordinated Multi-Point (CoMP) transmission with SLP in a Multi-User Multiple-Input Single-Output (MU-MISO) system, the strategy improves interference utilization, enhances system energy efficiency, and reduces computational complexity.  Methods  An SINR adaptive SLP and position joint anti-jamming optimization strategy for multiple UAVs is proposed by integrating CoMP and SLP technologies. To address the challenges of three-dimensional operational space and the overlap of nodes assigned to multiple sets, a ground-to-air multi-node matching mechanism based on three-dimensional K-means++ collaborative set partitioning is designed. To reduce the computational complexity of the joint optimization process, an iterative optimization algorithm based on particle reconstruction is developed. This algorithm simultaneously solves for both the UAV precoding matrix and spatial positions with low computational overhead. Additionally, an SINR adaptive SLP approach is introduced to enable optimized power allocation for multiple UAVs, considering the varying power characteristics of jamming and noise experienced by users. Simulation results demonstrate that the integration of CoMP and SLP technologies effectively enhances the communication performance of jammed users, while stable communication performance is maintained for ordinary users.  Results and Discussions  In the proposed UAV anti-jamming communication strategy, the SINR of jammed users is improved without compromising the normal communication performance of ordinary users. In the simulation results, UAV positions are marked with five-pointed stars (Figs. 6 and 7), UAV coverage areas are represented by circles, and users are indicated by “*”. A comparison of SINR variations under four schemes (Fig. 8) shows that the received SINR of jammed users increases by approximately 12~13 dB, while the SINR of ordinary users remains above the required threshold. When different SINR thresholds are applied, the received SINR of each user type varies accordingly (Fig. 9). By setting appropriate thresholds based on actual scenario requirements, different energy allocation effects can be achieved. Following optimization, the Bit Error Rate (BER) of jammed users is significantly reduced (Fig. 10). The constellation diagrams comparing the received signals under two precoding schemes (Fig. 11) indicate that the proposed SINR adaptive SLP strategy for multiple UAVs effectively improves the SINR of jammed users, while maintaining the communication quality of ordinary users. Moreover, the fitness evolution curve of the iterative optimization algorithm based on particle reconstruction (Fig. 12) shows that the algorithm approaches the global optimal solution at an early stage of iteration.  Conclusions  To address the challenge of anti-jamming communication for multi-user services supported by multiple UAVs, this study integrates CoMP transmission with SLP technology and proposes an SINR adaptive SLP and position joint optimization strategy for multi-UAV anti-jamming communication. The strategy is implemented in two stages. First, to solve the clustering problem in three-dimensional space and allow nodes to belong to multiple groups, a ground-to-air multi-node matching mechanism based on three-dimensional K-means++ collaborative set partitioning is designed. Second, an SINR adaptive SLP method is proposed to optimize UAV power allocation based on the SINR requirements of different user types. To reduce the computational complexity of jointly optimizing the precoding matrix and UAV positions, an iterative optimization algorithm based on particle reconstruction is developed. Simulation results demonstrate that the proposed strategy, by combining CoMP and SLP, effectively improves the communication performance of jammed users while maintaining reliable communication for ordinary users.
Low-Complexity Spectrum-efficiency Optimization Algorithm for Cell-Free Massive MIMO-NOMA Systems
ZHOU Wei, YANG Yu, XIANG Bo, ZHANG Yi, HUANG Hua
Available online  , doi: 10.11999/JEIT250189
Abstract:
  Objective  With the evolution of wireless communication toward ultra-dense networks, optimizing spectrum efficiency in cell-free massive Multiple-Input Multiple-Output (MIMO) Non-Orthogonal Multiple Access (NOMA) systems faces the challenge of balancing algorithmic complexity and performance. Traditional user clustering methods, such as random clustering, exhibit high randomness and easily converge to suboptimal solutions, whereas exhaustive search is computationally prohibitive. Similarly, power allocation schemes with rigid fixed thresholds often fail to accommodate dynamic user demands, resulting in imbalanced resource utilization and reduced fairness. To address these limitations, this study proposes a low-complexity joint optimization algorithm. By collaboratively designing user clustering and power allocation, the algorithm maximizes the system’s sum spectrum efficiency while guaranteeing the quality of service for low-rate users, thereby offering an efficient resource allocation strategy for cell-free massive MIMO-NOMA systems.  Methods  A downlink sum spectrum efficiency maximization model is first constructed and decomposed into two sub-problems: user clustering and power allocation. A clustering algorithm based on cluster head selection and channel difference maximization is then proposed, which reduces the complexity of pairing searches by optimizing the selection of cluster heads. Based on the clustering results, a minimum-rate enhancement constraint mechanism is incorporated. The resulting non-convex power allocation problem is subsequently transformed into a convex optimization form using the Successive Convex Approximation (SCA) method.  Results and Discussions  The relationship between sum spectrum efficiency and the number of users under different clustering algorithms is shown in (Fig. 2). Under perfect Successive Interference Cancellation (SIC) conditions, the proposed algorithm achieves performance comparable to that of the statistical clustering algorithm, whereas the combined clustering algorithm yields the lowest efficiency. Under imperfect SIC, the proposed algorithm maintains the highest spectrum efficiency, while the performance of the statistical algorithm decreases. The sum spectrum efficiency is markedly reduced by imperfect SIC, particularly when the number of users is small. This reduction arises from insufficient elimination of intra-cluster interference, leading to increased residual interference. Notably, the NOMA system supports twice the user capacity of Orthogonal Multiple Access (OMA), confirming its superior spectrum resource utilization. The relationship between sum spectrum efficiency and the number of Access Points (APs) is shown in (Fig. 3). Spectrum efficiency improves substantially as the number of APs increases, owing to enhanced channel hardening and interference suppression. Under perfect SIC, both the proposed algorithm and the statistical clustering algorithm achieve similar performance, exceeding the combined algorithm. The advantage of the proposed algorithm is further demonstrated under imperfect SIC. Increasing the number of deployed APs strengthens anti-interference capability and expands user capacity, verifying the efficiency and robustness of the proposed clustering algorithm under both perfect and imperfect SIC conditions. The effect of intra-cluster user numbers on spectrum efficiency is evaluated in (Fig. 4). Spectrum efficiency increases with the number of APs, largely due to improved coverage and channel estimation accuracy. However, efficiency decreases significantly as intra-cluster user numbers increase, which is attributed to aggravated intra-cluster interference and constraints on power resource allocation. The convergence of the proposed algorithm is demonstrated in (Fig. 5), where the optimal solution is reached within approximately seven iterations. After power allocation optimization, the efficiency gap between imperfect and perfect SIC remains below 0.2 bit/(s·Hz). Compared with Full Power Control (FPC), the Proposed Power Control (PPC) scheme effectively mitigates residual interference and achieves performance close to perfect SIC under practical conditions. The relationship between sum spectrum efficiency and AP numbers with different antenna configurations is presented in (Fig. 6). Efficiency continuously improves with increasing AP numbers and per-AP antenna numbers, owing to stronger channel hardening effects. Across all AP configurations, the PPC scheme demonstrates clear advantages over FPC, with the benefits becoming more pronounced in densely deployed networks. Power allocation optimization further enhances efficiency as antenna numbers increase. Finally, the cumulative distribution of per-user spectrum efficiency under different power control schemes is illustrated in (Fig. 7). The PPC scheme substantially reduces the proportion of inefficient users (0~0.2 bit/(s·Hz)), thereby improving system fairness. By contrast, the FPC scheme performs slightly better in the high-efficiency region, but this advantage comes at the expense of user fairness. In user-intensive scenarios, the PPC scheme effectively balances system stability and fairness by ensuring minimum rates for weak users.  Conclusions  A low-complexity joint optimization algorithm is presented to address the challenge of spectrum efficiency optimization in cell-free massive MIMO-NOMA systems. Through theoretical analysis and simulation, the spectrum efficiency and computational complexity of the system are compared under different user clustering algorithms and power allocation schemes. The results show that the proposed clustering algorithm significantly enhances system performance across various AP deployments and antenna configurations, while reducing computational complexity by 47.5% compared with the statistical clustering algorithm. Furthermore, the joint power allocation scheme demonstrates clear advantages over FPC in terms of spectrum efficiency and user fairness, verifying the effectiveness and practicality of the proposed algorithm.
Resource Allocation in Reconfigurable Intelligent Surfaces Assisted NOMA Based Space-Air-Ground Integrated Network
LIANG Wei, LI Aoying, LUO Wei, LI Lixin, LIN Wensheng, LI Xu, WEI Baoguo
Available online  , doi: 10.11999/JEIT250078
Abstract:
  Objective  The exponential growth of 6G wireless communication demands has positioned the Space-Air-Ground Integrated Network (SAGIN) as a promising architecture, aiming to achieve broad coverage and adaptive networking. However, complex geographic environments, including building obstructions, frequently hinder direct communication between ground users and base stations, thereby requiring effective relay strategies to maintain reliability. Reconfigurable Intelligent Surfaces (RIS) have attracted considerable attention for their capacity to improve signal coverage through passive beamforming. This study develops an RIS-assisted SAGIN architecture incorporating aerial RIS clusters and High-Altitude Platforms (HAPS) to enable communication between ground users and a Low Earth Orbit (LEO) satellite. To enhance energy efficiency, the system further optimizes user relay selection, power allocation, and beamforming for both LEO and RIS components.  Methods  The proposed system integrates LEO satellites, HAPS, Unmanned Aerial Vehicles (UAVs) equipped with RIS, and ground users within a three-dimensional communication space. Due to environmental obstructions, ground users are unable to maintain direct links with base stations; instead, RIS functions as a passive relay node. To improve relay efficiency, users are grouped and associated with specific RIS units. The total system bandwidth is partitioned into sub-channels assigned to different user groups. A matching algorithm is designed for user selection, followed by user group association with each RIS. For LEO communications, HAPS serve as active relay nodes that decode and forward signals to ground base stations. The system considers both direct and RIS-assisted communication links. An optimization problem is formulated to maximize energy efficiency under constraints related to user Quality of Service (QoS), power allocation, and beamforming for both LEO and RIS. To solve this, the proposed Alternating Pragmatic Iterative Algorithm in SAGIN (APIA-SAGIN) decomposes the problem into three sub-tasks: user relay selection, LEO beamforming, and RIS beamforming. The Successive Convex Approximation (SCA) and SemiDefinite Relaxation (SDR) methods are employed to transform the original non-convex problem into tractable convex forms for efficient solution.g.  Results and Discussions  Simulation results confirm the effectiveness of the proposed APIA-SAGIN algorithm in optimizing the energy efficiency of the RIS-assisted SAGIN. As shown in (Fig. 5), increasing the number of RIS elements and LEO antennas markedly improves energy efficiency compared with the random phase shift algorithm. This demonstrates that the proposed algorithm enables channel-aware control by aligning RIS beamforming with the ground transmission channel and jointly optimizing LEO beamforming, RIS beamforming, and LEO-channel alignment. As illustrated in (Fig. 6), both energy efficiency and the achievable rate of the LEO link increase with transmission power. However, beyond a certain power threshold, energy consumption rises faster than the achievable rate, leading to diminishing or even negative returns in energy efficiency. (Fig. 7) shows that higher power in ground user groups leads to increased achievable rates. Nonetheless, expanding the number of RIS elements proves more effective than increasing transmission power for enhancing user throughput. As shown in (Fig. 8), a higher number of RIS elements leads to simultaneous improvements in energy efficiency and achievable rate in the ground segment. Moreover, increasing the number of ground users does not degrade energy efficiency; instead, it results in a gradual increase, suggesting efficient resource allocation. Compared with the random phase shift algorithm, the proposed approach achieves superior performance in both energy efficiency and achievable rate. These findings support its potential for practical deployment in SAGIN systems.  Conclusions  This study proposes an RIS-assisted SAGIN architecture that utilizes aerial RIS clusters and HAPS to support communication between ground users and LEO satellites. The APIA-SAGIN algorithm is developed to jointly optimize user relay selection, LEO beamforming, and RIS beamforming with the objective of maximizing system energy efficiency. Simulation results demonstrate the effectiveness and robustness of the algorithm under complex conditions. The proposed approach offers a promising direction for improving the energy efficiency and overall performance of SAGIN, providing a foundation for future research and practical implementation.
Covert Communication Transmission Scheme in Low Earth Orbit Satellite Integrated Sensing and Communication Systems
ZHU Zhengyu, OUYANG Zebin, PAN Gaofeng, WANG Shuai, SUN Gangcan, CHU Zheng, HAO Fengyu
Available online  , doi: 10.11999/JEIT250208
Abstract:
  Objective  Globalization has driven growing demand for ubiquitous communication. However, traditional terrestrial networks face limitations in areas with poor infrastructure, such as deserts, oceans, and airspace. Satellite communication, with its flexibility, wide frequency coverage, and high capacity, offers a viable alternative for these environments. Yet, due to the open nature of satellite channels and long-distance transmission, signals remain vulnerable to interception, posing serious security challenges. Conventional wireless security measures, including physical-layer techniques and encryption, are limited by assumptions about computational hardness and the statistical structure of intercepted ciphertexts. Covert communication aims to reduce the probability of signal detection by eavesdroppers, providing an alternative layer of security. In Low Earth Orbit (LEO) satellite systems, high mobility leads to dynamic channel variations and misalignment of transmission beams, further complicating covert communication. To mitigate these challenges, Integrated Sensing And Communication (ISAC) technology can be embedded in satellite systems. By detecting the approximate location of potential eavesdroppers and directing radar beams to degrade their detection capability, ISAC improves communication covertness. This study proposes a covert communication strategy tailored for LEO satellite ISAC systems, offering enhanced security in satellite-based infrastructure.  Methods  This paper investigates a covert communication scheme in LEO satellite ISAC systems. First, the system model of the LEO satellite ISAC system is proposed, and the covertness constraints within the system are analyzed. Based on this, an optimization problem is formulated with the objective of maximizing the total covert communication rate for multiple users, subject to constraints including satellite power limits, radar power limits, sensing performance thresholds, and covertness requirements. This optimization problem is non-convex and highly coupled in variables, making it intractable to solve directly. To address this, relaxation variables are introduced to simplify the objective function into a solvable form. Subsequently, an alternating optimization algorithm is employed to decompose the original problem into two subproblems. One subproblem is transformed into an equivalent convex problem using methods like SemiDefinite Relaxation (SDR) and Successive Convex Approximation (SCA), which is then solved iteratively. The obtained solution from this subproblem is then used to solve the other subproblem. Once the solutions to both subproblems are determined, the original problem is effectively solved.  Results and Discussions  This paper validates the performance of the proposed covert communication transmission scheme through numerical simulations. To demonstrate the effectiveness of the proposed scheme, a comparison with a system lacking ISAC technology is introduced. First, the convergence of the proposed algorithm is verified (Fig. 2). The algorithm achieves rapid convergence under different resource allocation configurations in the system (converging at the 7th iteration). Additionally, higher covertness requirements lead to lower covert communication rates, as stricter covertness constraints reduce the allowable satellite transmit power. Increasing radar power causes a slight decline in the covert rate because Bob is subjected to minor interference from the radar beam. However, since the primary radar beam is focused on Eve, the interference level remains limited. Second, a higher number of satellite antennas improves the system’s covert communication rate (Fig. 3). This is attributed to the increased spatial dimensions available for optimizing signal transmission and reception. More antennas enable additional independent signal paths, enhancing transmission performance and flexibility. When the satellite is equipped with 81 antennas and the covertness requirement is 0.1, the total covert rate reaches 17.18 bit/(s·Hz). Compared to the system without ISAC technology, the proposed system demonstrates superior resistance to detection (Fig. 4). The radar beam introduces interference to eavesdroppers, and higher radar power degrades the eavesdropper’s detection performance. Finally, increasing the satellite transmit power improves the covert communication rate (Fig. 5). In the system without ISAC, the satellite power cannot be arbitrarily increased due to covertness constraints. However, the ISAC-enabled system leverages radar integration to grant the satellite greater flexibility in power allocation, thereby enhancing the covert rate. Notably, raising the lower bound of sensing performance slightly reduces the covert rate, as higher sensing demands require increased radar power, which introduces additional interference to Bob’s signal reception.  Conclusions  This paper investigates a covert communication transmission scheme in an LEO satellite-enabled ISAC system. An optimization problem is formulated to maximize the sum of multi-user covert communication rates, subject to constraints including covertness requirements, minimum sensing performance, radar power budget, and satellite power budget. The alternating optimization algorithm and SCA algorithm are jointly employed to design the radar and satellite beamforming, as well as the radar receiver filter. Simulation results demonstrate that, compared to a system without radar beam interference, the proposed system with radar beam interference significantly reduces the detectability of communication between Alice and Bob. Furthermore, the integration of radar enhances the satellite’s flexibility to increase transmit power, thereby improving the overall covert communication rate of the system.
Efficient Blind Detection Clustering Algorithm for Reconfigurable Intelligent Surface-aided Spatial Modulation Systems
ZHANG Lijuan, SHA Sha, ZHONG Huaqian
Available online  , doi: 10.11999/JEIT250162
Abstract:
  Objective  Reconfigurable Intelligent Surface (RIS)-aided Spatial Modulation (SM) systems (RIS-SM) represent an advanced integration of two promising technologies, offering considerable potential to enhance spectral and energy efficiency in future Sixth-Generation (6G) networks. However, the passive nature of RIS poses a significant challenge to acquiring accurate Channel State Information (CSI), which is essential for conventional signal detection methods. Most traditional algorithms assume perfect CSI, an unrealistic assumption in practical deployments that constrains both performance and scalability. To overcome this limitation, this study proposes an enhanced K-means clustering detection algorithm that reframes the blind detection problem as a clustering task. By exploiting the statistical channel distribution properties inherent to RIS-SM systems, the proposed algorithm achieves efficient blind signal detection without prior CSI and with low computational complexity. These findings contribute to the optimization of RIS-SM system performance and support the development of robust blind detection strategies in complex wireless communication scenarios.  Methods  This study considers a typical RIS-SM architecture in which the RIS comprises multiple passive reflecting elements and the receiver is equipped with multiple antennas. The wireless channel is modeled as an equivalent channel derived from the unique propagation characteristics of the RIS-SM system. By exploiting the statistical distribution of the diagonal elements of this equivalent channel, an improved K-means clustering detection algorithm is proposed, eliminating the requirement for prior CSI. The algorithm’s performance is assessed through extensive simulations. The communication channel is modeled as a Rayleigh fading environment with Additive White Gaussian Noise (AWGN). Phase Shift Keying (PSK) and Quadrature Amplitude Modulation (QAM) schemes are adopted. Bit Error Rate (BER) is used as the primary performance metric to compare the proposed approach against conventional detection methods. Simulations are carried out on the MATLAB platform. Results demonstrate that the proposed algorithm significantly lowers computational complexity while achieving near-optimal detection accuracy, highlighting its suitability for practical deployment in RIS-SM systems.  Results and Discussions  Simulation results indicate that the proposed improved K-means clustering detection algorithm markedly enhances the performance of RIS-SM systems under unknown CSI conditions. As illustrated in Fig. 4 and Fig. 5, the proposed method achieves BER performance that closely approaches that of ideal KMC (KMC(ideal)) and Maximum Likelihood (ML) detectors. It outperforms several benchmark algorithms, including traditional KMC, multi-initialization KMC (KMC(\begin{document}$ P $\end{document})), improved KMC (IKMC), K-means++ (KMC++), and Constrained Clustering-Based Detection (CCBD). While repeated initialization in KMC (\begin{document}$ P $\end{document}) can alleviate poor local minima, it substantially increases computational complexity. IKMC improves centroid initialization but yields limited enhancement in detection accuracy. KMC++ exhibits performance degradation at high Signal-to-Noise Ratio (SNR) due to persistent error floors. CCBD utilizes structural priors to improve clustering but shows performance deterioration when the number of received signals is small. Furthermore, Fig. 6 demonstrates that the proposed algorithm converges more rapidly than other clustering-based methods, particularly under low SNR conditions, highlighting its robustness and efficiency in noisy environments. Table 1 provides additional evidence that the proposed method achieves a favorable balance between detection accuracy and computational cost, supporting its practical value for blind detection in RIS-SM systems.  Conclusions  This study proposes an improved K-means clustering detection algorithm for RIS-SM systems, reformulating the blind detection task as an unsupervised clustering problem. By exploiting the distribution properties of the diagonal elements of the equivalent channel, the algorithm enhances the initialization of clustering centroids, enabling efficient blind signal detection without prior CSI. Simulation results confirm that the proposed method achieves near-optimal detection performance across a wide range of SNR levels and system configurations. These findings offer meaningful contributions to both the theoretical development and practical deployment of RIS-SM systems. The proposed approach is particularly relevant for the design of low-complexity and high-efficiency detection schemes in next-generation wireless networks, including Beyond-5G (B5G) and 6G systems. This work also provides a foundation for future research on blind detection algorithms and their integration with emerging technologies such as RIS, contributing to the advancement of intelligent and adaptive wireless communication systems.
Overviews
Integrating Intelligent Sensing, Transmission, and Control for Industrial IoT Networks: Key Technologies and Future Directions
ZHANG Mingqiang, MA Xiaocong, YANG Yajuan, LI Dongyang, Li Tiantian, WANG Leiyu, ZHANG Haixia, YUAN Dongfeng
Available online  , doi: 10.11999/JEIT250305
Abstract:
  Significance   The edge intelligence-enhanced sixth-generation (6G) mobile networks aim to build an integrated architecture that combines sensing, communication, and computation, continuing the trend of 5G’s rapid expansion into vertical industries. Looking ahead, Industry 5.0—defined by human-centric design and large-scale personalized customization—requires 6G-enabled industrial networks to simultaneously meet the demands of sensing, transmission, and control. The efficient interconnection, communication, and intelligent management of large-scale Industrial Internet of Things (IIoT) devices remains fundamental to the digital, networked, and intelligent transformation of the manufacturing sector and its high-quality development. However, limited device resources, complex industrial environments, and the fragmented design of sensing, transmission, and control systems present major challenges. These include limited capability for comprehensive and accurate information sensing, inefficient interaction among heterogeneous devices and systems, and difficulties in achieving intelligent closed-loop collaboration across sensing, transmission, and control. Integrating Intelligent Sensing, Transmission, and Control (ISTC) is essential to enabling intelligent communications in industrial scenarios, facilitating the intelligent interconnection of humans, machines, objects, and environments, and enhancing intelligent management and control across production lines.  Progress   Achieving semantic interoperability across heterogeneous industrial systems is the core barrier to the integrated design of sensing, transmission, and control, and is also critical to enabling agile interaction between diverse systems, reducing subsystem development and deployment costs, and building autonomous, self-managing industrial networks. Modern IIoT systems typically integrate parallel subsystems across Information Technology (IT) and Operational Technology (OT) domains, each with independent data models and semantic specifications, resulting in natural interoperability barriers. These barriers restrict efficient interaction and expected collaborative operation across vendors and platforms, significantly limiting large-scale interconnection and data sharing. Therefore, comprehensive and accurate information sensing, reliable and efficient transmission, and responsive feedback control have become key requirements for future IIoT networks. Specifically: (1) Intelligent Sensing: Overcoming the limitations of the Nyquist sampling theorem through interpretable intelligent sensing is a prerequisite for ISTC. (2) Semantic Transmission: The effective extraction and unified representation of industrial semantics, combined with intelligent semantic-level interaction, are critical to ensuring interoperability in heterogeneous systems while maintaining operational efficiency and sustainable performance. (3) Integrated ISTC: Joint design of edge-intelligent sensing, efficient transmission, and optimal control enables streamlined workflows in industrial scenarios, reducing system response time, improving control accuracy, and optimizing energy efficiency.  Conclusions   This paper proposes an intelligent collaborative architecture for IIoT networks comprising edge nodes or terminals, intelligent gateways, and industrial cloud platforms. The focus is placed on three key technologies within Integrating ISTC: (1) Intelligent sensing methods for IIoT networks: These methods enhance sensing efficiency and accuracy by applying interpretable, physics-informed deep compressed sensing approaches to IIoT devices and systems. (2) Robust Industrial Semantic Communications (ISC) driven by cognitive intelligence: This technology combines industrial knowledge graphs with semantic communication mechanisms to improve semantic interoperability and transmission efficiency across heterogeneous industrial systems. (3) Joint design of edge-intelligent sensing, efficient semantic transmission, and optimal control: By clarifying the intrinsic coupling among sensing, transmission, and control processes, this approach optimizes the collaborative service capability of heterogeneous industrial networks and systems.  Prospects   Despite progress, ISTC still faces considerable challenges. Future research may focus on the following directions: (1) Industrial large models and intelligent agents: The development of specialized AI models remains essential, particularly in core industrial domains where implicit knowledge is concentrated. (2) Industry 5.0: Achieving efficient, semantic-level human-machine collaborative interaction will be a key breakthrough for future industrial scenarios. (3) Industrial cross-modal collaborative interaction: Integrating data across modalities and mining knowledge from diverse sources present significant challenges but are essential for enabling advanced collaborative interaction in IIoT networks. (4) Industrial digital twins: For complex industrial environments and physical systems, continued advances in digital twin technology—particularly in high-precision semantic perception, real-time efficient interaction, and adaptive fault-tolerant control, will play a critical role in accelerating ISTC development.
Network Protocol Fuzzing: Method Classification and Research Progress
LAI Guoqing, ZHU Yuefei, CHEN Di, LU Bin, LIU Long, ZHANG Zihao
Available online  , doi: 10.11999/JEIT250188
Abstract:
  Significance   Network security vulnerabilities arising from flaws in network protocol implementations cause substantial losses and adverse societal effects, exemplified by the Heartbleed vulnerability in OpenSSL (CVE-2014-0160). This flaw allows attackers to extract encryption keys from public servers, enabling decryption of traffic or unauthorized authentication on less secure systems. Approximately 500,000 internet servers are affected when the vulnerability is publicly disclosed. Against this backdrop of escalating network security risks, ensuring the security of network protocol software becomes a critical research area. Fuzzing, a software testing technique, emerges as one of the most widely used approaches for identifying vulnerabilities in network protocol implementations due to its ease of deployment and high efficiency. The core concept of fuzzing is to improve software security by generating and sending crafted test cases to the target program. Despite significant progress in this field, Network Protocol Fuzzing (NPF) still faces technical challenges. Currently, no systematic and up-to-date review of NPF research exists, limiting researchers’ ability to grasp recent advances. This paper conducts a comprehensive review of NPF techniques, aiming to provide researchers in the field of network protocol security with valuable references for tool selection and optimization.  Progress   Since the proposal of AFLNET in 2020, considerable progress occurs in the field of NPF, addressing key challenges such as protocol packet construction, protocol state awareness, and network communication efficiency optimization. Specifically, for protocol packet construction, researchers propose machine learning-based packet generation methods that integrate both generation and mutation strategies. Mutation operators and guidance techniques are designed to target specific protocol characteristics. In terms of protocol state awareness, state tracking capabilities are enhanced through state variable capture and state model construction. Furthermore, state fuzzing has been widely employed to detect state machine bugs. For protocol stack efficiency optimization, researchers improve testing efficiency by refining communication mechanisms and applying snapshot techniques.  Conclusions  To comprehensively summarize the research progress in NPF, this paper first clarifies the unique characteristics of network protocol software compared with other fuzzing targets. These characteristics include strict protocol message formats, asynchronous network interactions, and complex protocol state maintenance. A problem-oriented classification framework for NPF is proposed, structured around three core challenges: protocol packet construction, protocol state awareness, and protocol stack efficiency optimization. Based on this framework, research advancements in NPF over the past eight years are systematically reviewed, with a technical analysis of the capabilities, and limitations of existing approaches. This review highlights several key challenges in the field. For input construction, major limitations include weak validity of generated inputs, input space explosion, and restrictions imposed by encryption and authentication mechanisms. In terms of state awareness, the field faces insufficient protocol state space exploration and low levels of test intelligence and automation. Regarding performance optimization, technical challenges include slow network communication speed, limited scalability across different protocol implementations, and inadequate adaptability to complex network environments. This study provides both theoretical foundations and practical references to guide future research and technological development in the NPF domain.  Prospects   Future research in NPF can integrate emerging technologies such as Artificial Intelligence (AI) to enhance the intelligence and automation of testing processes. For example, combining AI methods with program analysis techniques may enable a deeper understanding of protocol behavior and more efficient generation of test packets. Developing state representations tailored to different protocol characteristics, implementing real-time protocol state mapping, and applying state prediction and reasoning based on LLM can further improve the efficiency and applicability of NPF tools. In addition, introducing technologies such as parallelization, distributed computing, modular test architectures, and integrated network simulation with virtualization can significantly enhance testing scalability and adaptability. Through the integration of emerging technologies and interdisciplinary research, NPF is expected to play an increasingly critical role in network protocol security, providing robust technical support for building secure and reliable network environments.
Human-Machine Fusion Intelligent Decision-Making: Concepts, Framework, and Applications
LI Zhe, WANG Ke, WANG Biao, ZHAO Ziqi, LI Yafei, GUO Yibo, HU Yazhou, WANG Hua, LV Pei, XU Mingliang
Available online  , doi: 10.11999/JEIT250260
Abstract:
  Significance  The exponential growth of data volume, advances in computational power, and progress in algorithmic theory have accelerated the development of Artificial Intelligence (AI). Although AI offers unprecedented opportunities across industries, it continues to face limitations such as dependence on large datasets, poor interpretability of learning and decision-making mechanisms, limited robustness, and susceptibility to hallucinations. To overcome these challenges, integrating human cognitive decision-making capabilities and human-like cognitive models into AI systems is essential. This integration gives rise to a new form of intelligence—Human-Machine Fusion Intelligence—which combines physiological and physical characteristics. The core concept is to harness the complementary strengths of humans and machines in information processing and decision-making: humans provide intuitive judgment and contextual understanding, whereas machines are capable of high-speed computation and large-scale data analysis. By establishing a synergistic, collaborative “partnership,” Human-Machine Fusion Intelligent Decision-Making seeks to optimize decision quality through coordinated organic and probabilistic integration. This paradigm holds significant potential to improve decision reliability in mission-critical contexts, such as military operations, medical procedures, and autonomous driving, thus offering both theoretical research value and practical application relevance.  Conclusions  This review adopts a systematic research approach to examine Human-Machine Fusion Intelligence in decision-making across four core dimensions. First, it presents a theoretical analysis of the fundamental concepts underpinning Human-Machine Fusion Intelligence and highlights its unique advantages in complex decision-making contexts. Second, it proposes a general framework for Human-Machine Fusion Intelligent Decision-Making systems, emphasizing two key components: situational awareness and collaborative decision-making. Based on this framework, decision-making approaches are categorized into three types according to task characteristics and the nature of human-machine interaction: human-led, machine-led, and human-machine collaborative decision-making. Third, the review synthesizes recent practical advancements in representative application domains. Finally, it examines emerging trends in the development of Human-Machine Fusion Intelligent Decision-Making.  Progress  Unlike prior reviews that focus primarily on specific application domains, this article presents a comprehensive overview of Human-Machine Fusion Intelligence across four key dimensions: conceptual foundations, system framework, practical applications, and current challenges and future prospects. The core contributions of this review are summarized in the following four areas: First, it elucidates the advantages of Human-Machine Fusion Intelligent Decision-Making systems: (1) Improved decision-making accuracy—By combining machines’ strengths in data processing and logical reasoning with human capabilities in handling unstructured problems and ethically complex decisions, the system enables dynamic adjustment through a human-in-the-loop mechanism. (2) Enhanced interpretability of decision outcomes—The decision-making process bridges the cognitive gap between humans and machines, providing a transparent, traceable decision path and clarifying accountability boundaries. (3) Greater system robustness—By integrating machines’ risk monitoring and adaptive capabilities with human experiential judgment in complex or uncertain environments, the system establishes a closed-loop collaboration that balances technological rationality with human cognition. Second, the article highlights that Human-Machine Fusion systems cannot operate independently in safety-critical contexts due to imperfect trust mechanisms and ethical constraints. In response, it proposes a hierarchical architecture comprising two key layers: (1) Situational awareness layer, including three core processes: multimodal data perception, cross-modal information fusion, and situational analysis. (2) Collaborative decision-making layer, which distinguishes three decision-making paradigms based on task characteristics and human-machine interaction mode: (a) Human-led decision-making, suited for tasks with high uncertainty and open-ended conditions, where an enhanced intelligence model with a human-in-the-loop is adopted. (b) Machine-led decision-making, appropriate for tasks with lower uncertainty, emphasizing hybrid intelligence through cognitive model integration in automated workflows. (c) Human-machine collaborative decision-making, applicable when human and machine strengths are complementary, allowing for equal, synergistic cooperation to optimize decision efficiency. Third, the article synthesizes recent technological progress, summarizing representative applications of Human-Machine Fusion Intelligent Decision-Making in mission-critical domains such as the military, healthcare, and autonomous driving. Finally, it identifies six key directions for future development: optimization of multimodal perception, fusion of semantic and feature spaces, construction of deep collaborative feedback loops, dynamic task allocation mechanisms, enhancement of system reliability, and development of ethical guidelines. These directions aim to advance efficient collaboration and sustainable evolution of human-machine intelligence.  Prospects  Human-Machine Fusion Intelligent Decision-Making offers substantial research value and strong application potential for advancing emerging industries and enabling new intelligent paradigms. Although several exploratory efforts have been made, the field remains in its infancy, lacking a unified and mature theoretical or technological foundation. Key scientific and engineering challenges persist, including the optimization of multimodal perception and data fusion, bridging the semantic gap between human cognition and machine-represented feature spaces, and achieving deep integration of human and machine intelligence. Continued interdisciplinary collaboration will be essential to drive theoretical progress and technological innovation, further unlocking the potential of Human-Machine Fusion Intelligent Decision-Making.
Overview of the Research on Key Technologies for AI-powered Integrated Sensing, Communication and Computing
ZHU Zhengyu, YIN Menglin, YAO Xinwei, XU Yongjun, SUN Gangcan, XU Mingliang
Available online  , doi: 10.11999/JEIT250242
Abstract:
The Integration of Sensing, Communication and Computing (ISCC) combined with Artificial Intelligence(AI) algorithms has emerged as a critical enabler of Sixth-Generation (6G) networks due to its high spectral efficiency and low hardware cost. AI-powered ISCC systems, which combine sensing, communication, computing, and intelligent algorithms, support fast data processing, real-time resource allocation, and adaptive decision-making in complex and dynamic environments. These systems are increasingly applied in intelligent vehicular networks—including Unmanned Aerial Vehicles (UAVs) and autonomous driving—as well as in radar, positioning, tracking, and beamforming. This overview outlines the development and advantages of AI-enabled ISCC systems, focusing on performance benefits, application potential, evaluation metrics, and enabling technologies. It concludes by discussing future research directions. Future 6G networks are expected to evolve beyond data transmission to form an integrated platform that unifies sensing, communication, computing, and intelligence, enabling pervasive AI services.  Significance   AI-powered ISCC marks a transformative shift in wireless communication, enabling more efficient spectrum utilization, reduced hardware cost, and improved adaptability in complex environments. This integration is central to the development of 6G networks, which aim to deliver intelligent and efficient services across applications such as autonomous vehicles, UAVs, and smart cities. The significance of this research lies in its potential to reshape the management and optimization of communication, sensing, and computing resources, advancing the realization of a ubiquitously connected and intelligent infrastructure.  Progress   Recent advances in AI—particularly in machine learning, deep learning, and reinforcement learning—have substantially improved the performance of ISCC systems. These methods enable real-time data processing, intelligent resource management, and adaptive decision-making, which are critical for future 6G requirements. Notable progress includes AI-driven waveform design, beamforming, channel estimation, and dynamic spectrum allocation, all of which enhance ISCC efficiency and reliability. Additionally, the integration of edge computing and federated learning has mitigated challenges related to latency, data privacy, and scalability, facilitating broader deployment of AI-enabled ISCC systems.  Conclusions  Research on AI-powered ISCC systems highlights the benefits of integrating AI with sensing, communication, and computing. AI algorithms improve resource efficiency, sensing precision, and real-time adaptability, making ISCC systems well suited for dynamic and complex environments. The adoption of lightweight models and distributed learning has broadened applicability to resource-limited platforms such as drones and IoT sensors. Overall, AI-enabled ISCC systems advance the realization of 6G networks, where sensing, communication, and computing are unified to support intelligent and ubiquitous services.  Prospects   The advancement of AI-powered ISCC systems depends on addressing key challenges, including data quality, model complexity, security, and real-time performance. Future research should focus on developing robust AI models capable of generalizing across diverse wireless environments. Progress in lightweight AI and edge computing will be critical for deployment in resource-constrained devices. The integration of multi-modal data and the design of secure, privacy-preserving algorithms will be essential to ensure system reliability and safety. As 6G networks evolve, AI-powered ISCC systems are expected to underpin intelligent, efficient, and secure communication infrastructures, reshaping human-technology interaction in the digital era.
Excellence Action Plan Leading Column
HNN4RP: Heterogeneous Graph Neural Network for Rug Pull Detection on Ethereum
LI Chenchen, JIN Hai, WU Minrui, XIAO Jiang
Available online  , doi: 10.11999/JEIT250160
Abstract:
  Objective  The rapid expansion of Decentralized Finance (DeFi) has been accompanied by a rise in fraudulent activities, with Rug Pull scams posing a significant threat to the security, stability, and credibility of blockchain ecosystems. Due to the permissionless and decentralized nature of DeFi platforms, malicious actors can exploit smart contracts to deceive investors, artificially inflate token prices, and withdraw liquidity abruptly, leaving investors with worthless assets. The anonymous and pseudonymous characteristics of blockchain transactions further hinder fraud detection and regulatory oversight, making Rug Pull scams one of the most challenging forms of financial fraud in the DeFi sector. Although traditional fraud detection approaches, such as rule-based heuristics and statistical analysis, provide basic risk assessment, they often fail to capture the complex transaction patterns, dynamic interactions, and heterogeneous relationships within blockchain networks. To address these limitations, this study proposes Heterogeneous graph Neural Network for Rug Pull detection (HNN4RP), a fraud detection framework based on Heterogeneous Graph Neural Networks (GNNs), specifically designed for the early identification of Rug Pull scams. By leveraging the graph structure of blockchain transactions, HNN4RP models Ethereum transaction data as a heterogeneous graph to capture the relationships among smart contracts, liquidity pools, and trader addresses. The framework integrates advanced representation learning techniques to reveal hidden dependencies, detect anomalous patterns, and differentiate between legitimate and fraudulent behaviors in a scalable, data-driven manner. Furthermore, HNN4RP incorporates temporal dynamics and contextual information to adapt to evolving fraud tactics.  Methods  HNN4RP models Ethereum transactions as a heterogeneous graph, distinguishing between regular nodes and smart contract nodes while preserving both structured and unstructured transactional data. By leveraging a graph-based representation, the model captures complex interactions within the Ethereum network, supporting comprehensive analysis of transactional behavior. This approach explicitly represents key relationships between entities, including token transfers, contract interactions, and liquidity movements, thereby facilitating more accurate fraud detection.The core components of HNN4RP include a Random Walk with Restart (RWR) sampling mechanism, which mitigates biases caused by skewed transaction distributions and enables balanced exploration of node neighborhoods. This technique captures long-range dependencies in the transaction network while preserving the local structural context, which is essential for detecting anomalous behaviors associated with fraud. By dynamically adjusting the probability of revisiting specific nodes, RWR improves the model’s focus on relevant transactional pathways, reducing the effects of data sparsity and imbalance.In addition to RWR-based sampling, HNN4RP incorporates multi-modal feature encoding to capture both structured transactional relationships and unstructured node attributes, such as smart contract code and token metadata. This integration enables the model to leverage a diverse set of informative features, improving its ability to differentiate between legitimate and fraudulent activities. By encoding both numerical and textual representations of transaction data, the framework enhances the robustness of fraud detection against evasive tactics employed by malicious actors.Furthermore, HNN4RP employs a BiLSTM-based temporal fusion mechanism to retain the sequential dependencies within transaction histories, effectively modeling temporal behavioral patterns. This component is essential for capturing the evolution of fraudulent activities over time, as Rug Pull scams often exhibit time-dependent characteristics. By learning the progression of transactional patterns, the model enhances fraud classification accuracy and enables early detection of fraudulent schemes before they escalate. Through this comprehensive approach, HNN4RP provides a scalable, adaptive, and high-precision solution for fraud detection in decentralized financial ecosystems.  Results and Discussions  The performance of the proposed HNN4RP model is evaluated against three categories of baseline methods: attribute-based approaches, random walk-based algorithms, and deep learning models. Across all dataset scales, HNN4RP consistently outperforms these baselines across three predefined evaluation metrics, demonstrating superior fraud detection capability. Notably, on the D2 and D3 datasets, HNN4RP achieves an accuracy of 96% and an F1 score of 0.95 (Table 2). Ablation experiments indicate that both attribute-based features and textual representations contribute significantly to Rug Pull detection performance (Fig. 3). Parameter sensitivity analysis shows that the optimal neighborhood size is 16, and the best embedding dimension is 27 (Figs. 4 and 5). Further experiments incorporating different types of noise interference confirm that HNN4RP exhibits strong robustness to data perturbations, with the F1 score declining by no more than 7% under moderate noise levels (Table 3). Early fraud detection tests demonstrate that HNN4RP effectively identifies fraudulent activities at an early stage, providing reliable warning signals before scams fully develop (Table 4).  Conclusions  This study addresses the emerging challenge of Rug Pull scams in the DeFi and Web3 domains and proposes a deep learning-based detection model, HNN4RP, with significant implications for the stability and security of decentralized financial ecosystems. By formalizing the Rug Pull detection task using heterogeneous graph modeling, this work tackles three major challenges in Ethereum-based fraud detection: biased transaction exploration, incomplete feature representation, and insufficient temporal pattern recognition. These challenges are addressed through the integration of RWR-based random walks, heterogeneous node feature extraction, and BiLSTM-based temporal feature fusion. Experimental results on real Ethereum transaction datasets demonstrate that HNN4RP achieves 96% accuracy and provides reliable early warnings within four hours before fraudulent activities occur, confirming its practical applicability for early fraud detection in decentralized environments.
Satellite Navigation
Research on GRI Combination Design of eLORAN System
LIU Shiyao, ZHANG Shougang, HUA Yu
Available online  , doi: 10.11999/JEIT201066
Abstract:
To solve the problem of Group Repetition Interval (GRI) selection in the construction of the enhanced LORAN (eLORAN) system supplementary transmission station, a screening algorithm based on cross interference rate is proposed mainly from the mathematical point of view. Firstly, this method considers the requirement of second information, and on this basis, conducts a first screening by comparing the mutual Cross Rate Interference (CRI) with the adjacent Loran-C stations in the neighboring countries. Secondly, a second screening is conducted through permutation and pairwise comparison. Finally, the optimal GRI combination scheme is given by considering the requirements of data rate and system specification. Then, in view of the high-precision timing requirements for the new eLORAN system, an optimized selection is made in multiple optimal combinations. The analysis results show that the average interference rate of the optimal combination scheme obtained by this algorithm is comparable to that between the current navigation chains and can take into account the timing requirements, which can provide referential suggestions and theoretical basis for the construction of high-precision ground-based timing system.