Latest Articles

Articles in press have been peer-reviewed and accepted, which are not yet assigned to volumes/issues, but are citable by Digital Object Identifier (DOI).
Display Method:
BIRD1445: Large-scale Multimodal Bird Dataset for Ecological Monitoring
WANG Hongchang, XIAN Fengyu, XIE Zihui, DONG Miaomiao, JIAN Haifang
Available online  , doi: 10.11999/JEIT250647
Abstract:
  Objective  With the rapid advancement of Artificial Intelligence (AI) and growing demands in ecological monitoring, high-quality multimodal datasets have become essential for training and deploying AI models in specialized domains. Existing bird datasets, however, face notable limitations, including challenges in field data acquisition, high costs of expert annotation, limited representation of rare species, and reliance on single-modal data. To overcome these constraints, this study proposes an efficient framework for constructing large-scale multimodal datasets tailored to ecological monitoring. By integrating heterogeneous data sources, employing intelligent semi-automatic annotation pipelines, and adopting multi-model collaborative validation based on heterogeneous attention fusion, the proposed approach markedly reduces the cost of expert annotation while maintaining high data quality and extensive modality coverage. This work offers a scalable and intelligent strategy for dataset development in professional settings and provides a robust data foundation for advancing AI applications in ecological conservation and biodiversity monitoring.  Methods  The proposed multimodal dataset construction framework integrates multi-source heterogeneous data acquisition, intelligent semi-automatic annotation, and multi-model collaborative verification to enable efficient large-scale dataset development. The data acquisition system comprises distributed sensing networks deployed across natural reserves, incorporating high-definition intelligent cameras, custom-built acoustic monitoring devices, and infrared imaging systems, supplemented by standardizedpublic data to enhance species coverage and modality diversity. The intelligent annotation pipeline is built upon four core automated tools: (1) spatial localization annotation leverages object detection algorithms to generate bounding boxes; (2) fine-grained classification employs Vision Transformer models for hierarchical species identification; (3) pixel-level segmentation combines detection outputs with SegGPT models to produce instance-level masks; and (4) multimodal semantic annotation uses Qwen large language models to generate structured textual descriptions. To ensure annotation quality and minimize manual verification costs, a multi-scale attention fusion verification mechanism is introduced. This mechanism integrates seven heterogeneous deep learning models, each with different feature perception capacities across local detail, mid-level semantic, and global contextual scales. A global weighted voting module dynamically assigns fusion weights based on model performance, while a prior knowledge-guided fine-grained decision module applies category-specific accuracy metrics and Top-K model selection to enhance verification precision and computational efficiency.  Results and Discussions  The proposed multi-scale attention fusion verification method dynamically assesses data quality based on heterogeneous model predictions, forming the basis for automated annotation validation. Through optimized weight allocation and category-specific verification strategies, the collaborative verification framework evaluates the effect of different model combinations on annotation accuracy. Experimental results demonstrate that the optimal verification strategy—achieved by integrating seven specialized models—outperforms all baseline configurations across evaluation metrics. Specifically, the method attains a Top-1 accuracy of 95.39% on the CUB-200-2011 dataset, exceeding the best-performing single-model baseline, which achieves 91.79%, thereby yielding a 3.60% improvement in recognition precision. The constructed BIRD1445 dataset, comprising 3.54 million samples spanning 1,445 bird species and four modalities, outperforms existing datasets in terms of coverage, quality, and annotation accuracy. It serves as a robust benchmark for fine-grained classification, density estimation, and multimodal learning tasks in ecological monitoring.  Conclusions  This study addresses the challenge of constructing large-scale multimodal datasets for ecological monitoring by integrating multi-source data acquisition, intelligent semi-automatic annotation, and multi-model collaborative verification. The proposed approach advances beyond traditional manual annotation workflows by incorporating automated labeling pipelines and heterogeneous attention fusion mechanisms as the core quality control strategy. Comprehensive evaluations on benchmark datasets and real-world scenarios demonstrate the effectiveness of the method: (1) the verification strategy improves annotation accuracy by 3.60% compared to single-model baselines on the CUB-200-2011 dataset; (2) optimal trade-offs between precision and computational efficiency are achieved using Top-K = 3 model selection, based on performance–complexity alignment; and (3) in large-scale annotation scenarios, the system ensures high reliability across 1,445 species categories. Despite its effectiveness, the current approach primarily targets species with sufficient data. Future work should address the representation of rare and endangered species by incorporating advanced data augmentation and few-shot learning techniques to mitigate the limitations posed by long-tail distributions.
Multi-modal Joint Distillation Optimization for Source Code Vulnerability Detection
Available online  , doi: 10.11999/JEIT250453
Abstract:
  Objective  As software systems increase in scale and complexity, the frequency of security vulnerabilities in source code rises accordingly, threatening system reliability, data integrity, and user privacy. Conventional automated vulnerability detection methods typically depend on a narrow set of shallow features—such as API call sequences, opcode patterns, or syntactic heuristics—rendering them susceptible to learning spurious correlations and unable to capture the rich semantic and structural information essential for accurate detection. Moreover, most existing approaches either rely on single-modal representations or weakly integrate multiple modalities without explicitly addressing distribution mismatches across them. This often results in overfitting to seen datasets and limited generalization to unseen codebases, particularly across different projects or programming languages. Although recent advances in machine learning and deep learning have improved source code analysis, effectively modeling the complex interactions between code semantics and execution structures remains a major challenge. To overcome these limitations, this paper proposes a multi-modal joint Distillation Optimization for Vulnerability Detection (mVulD-DO), a multimodal framework that combines deep feature distillation with dynamic global feature alignment. The proposed method aims to enhance semantic comprehension, structural reasoning, and cross-modal consistency, which are critical for robust vulnerability detection. By enforcing both intra-modality refinement and inter-modality alignment, mVulD-DO addresses the semantic-structural gap that constrains traditional methods.  Methods  The mVulD-DO framework begins by extracting multiple semantic modalities from raw source code—function names, variable names, token_type attributes, and local code slices—using program slicing and syntactic parsing techniques. In parallel, a Program Dependency Graph (PDG) is constructed to capture both control-flow and data-flow relationships, generating a heterogeneous graph that represents explicit and implicit program behaviors. Each semantic modality is embedded using pretrained encoders and subsequently refined via a deep feature distillation module, which integrates multi-head self-attention and multi-scale convolutional layers to emphasize salient patterns and suppress redundant noise. To model the sequential dependencies intrinsic to program execution, a Bidirectional Long Short-Term Memory (BLSTM) network captures long-range contextual interactions. For structural representation, a Graph Attention Network (GAT) processes the PDG-C to produce topology-aware embeddings. To bridge the gap between modalities, adaptive dynamic Sinkhorn regularization is applied to globally align the distributions of semantic and structural embeddings. This approach mitigates modality mismatches while preserving flexibility by avoiding rigid one-to-one correspondences. Finally, the distilled and aligned multimodal features are fused and passed through a lightweight fully connected classifier for binary vulnerability prediction. The model is jointly optimized using both classification and alignment objectives, improving its ability to generalize across unseen codebases.  Results and Discussions  Comprehensive comparisons conducted on the mixed CVEFixes+SARD dataset—covering 25 common CWE vulnerability types with an 8:1:1 train:validation:test split—demonstrate that traditional source code detectors, which directly map code to labels, often rely on superficial patterns and show limited generalization, particularly for out-of-distribution samples. These methods achieve accuracies ranging from 55.41% to 85.84%, with recall typically below 80% (Table 1). In contrast, mVulD-DO leverages multi-layer feature distillation to purify and enhance core representations, while dynamic Sinkhorn alignment mitigates cross-modal inconsistencies. This results in accuracy of 87.11% and recall of 83.59%, representing absolute improvements of 1.27% and 6.26%, respectively, over the strongest baseline method (ReGVD). Although mVulD-DO reports a slightly higher False Positive Rate (FPR) of 3.54%—2.92% above that of ReGVD—it remains lower than that of most traditional detectors. This modest increase is considered acceptable in practice, given that failing to detect a critical vulnerability typically incurs greater cost than issuing additional alerts. Compared with instruction-tuned large language models (e.g., VulLLM), which maintain low FPRs below 10% but suffer from recall below 75%, mVulD-DO offers a more favorable trade-off between false alarms and coverage of true vulnerabilities. Ablation studies (Table 2) further validate the contribution of each component. Removing function name embeddings (unfunc) results in a 1.3% decrease in F1 score; removing variable name embeddings (unvar) causes a 1.3% drop; and omitting token_type attributes (untype) leads to a 3.35% reduction. The most substantial performance degradation—9.11% in F1—occurs when the deep feature distillation module is disabled (undis), highlighting the critical role of multi-scale semantic refinement and noise suppression. Additional evaluations on vulnerability-sensitive subsets—Call, OPS, Array, and PTR—demonstrate consistent benefits from Sinkhorn alignment. F1 score improvements over unaligned variants are observed as follows: 1.45% for Call, 4.22% for OPS, 1.38% for Array, and 0.36% for PTR (Table 3), confirming the generalization advantage across a broad spectrum of vulnerability types.  Conclusions  Experimental results demonstrate that the proposed mVulD-DO framework consistently outperforms existing vulnerability detection methods in recall, F1-score, and accuracy, while maintaining a low FPR The effectiveness of deep feature distillation, multi-scale semantic extraction, and dynamic Sinkhorn alignment is validated through extensive ablation and visualization analyses. Despite these improvements, the model incurs additional computational overhead due to multimodal distillation and Sinkhorn alignment, and shows sensitivity to hyperparameter settings, which may limit its suitability for real-time applications. Moreover, while strong performance is achieved on the mixed dataset, the model's generalization across unseen projects and programming languages remains an open challenge. Future work will focus on developing lightweight training strategies—such as knowledge distillation and model pruning—to reduce computational costs. Additionally, incorporating unsupervised domain adaptation and incremental alignment mechanisms will be critical to support dynamic code evolution and enhance cross-domain robustness. These directions aim to improve the scalability, adaptability, and practical deployment of multimodal vulnerability detection systems in diverse software environments.
Mutualistic Backscatter NOMA Method for Coordinated Direct and Relay Transmission System
XU Yao, HU Rongfei, JIA Shaobo, LI Bo, WANG Gang, ZHANG Zhizhong
Available online  , doi: 10.11999/JEIT250405
Abstract:
  Objective  The exponential growth in data traffic necessitates that cellular Internet of Things (IoT) systems achieve both ultra-high spectral efficiency and wide-area coverage to meet the stringent service requirements of vertical applications such as industrial automation and smart cities. Non-Orthogonal Multiple Access-based Coordinated Direct and Relay Transmission (NOMA-CDRT) method can enhance both spectral efficiency and coverage by leveraging power-domain multiplexing and cooperative relaying, making it a promising approach to address these challenges. However, existing NOMA-CDRT frameworks are primarily designed for cellular communications and do not effectively support spectrum sharing or the deep integration of cellular and IoT transmissions. To overcome these limitations, this study proposes a Mutualistic Backscatter NOMA-CDRT (MB-NOMA-CDRT) method. This approach facilitates spectrum sharing and mutualistic coexistence between cellular users and IoT devices, while improving the system’s Ergodic Sum Rate (ESR).  Methods  The proposed MB-NOMA-CDRT method integrates backscatter modulation and power-domain superposition coding to develop a bidirectional communication strategy that unifies information transmission and cooperative assistance, enabling spectrum sharing and mutualistic coexistence between cellular users and IoT devices. Specifically, the base station uses downlink NOMA to serve the cellular center user directly and the cellular edge user via a relaying user. Simultaneously, IoT devices utilize cellular radio frequency signals and backscatter modulation to transmit their data to the base station, thereby achieving spectrum sharing. The backscattered IoT signals act as multipath gains, contributing to improved cellular communication quality. To rigorously characterize the system performance, the squared generalized-K distribution and Meijer-G functions are adopted to derive closed-form expressions for the ESR under both perfect and imperfect Successive Interference Cancellation (SIC). Building on this analytical foundation, a power allocation optimization scheme is developed using an enhanced Particle Swarm Optimization (PSO) algorithm to maximize system ESR. Finally, extensive Monte Carlo simulations are conducted to verify the ESR gains of the proposed method, confirm the theoretical analysis, and demonstrate the efficacy of the optimization scheme.  Results and Discussions  The performance advantage of the proposed MB-NOMA-CDRT method is demonstrated through comparisons of ESR with conventional NOMA-CDRT and Orthogonal Multiple Access (OMA) schemes (Fig. 2 and Fig. 3). The theoretical ESR results closely match the simulation data, confirming the validity of the analytical derivations. Under both perfect and imperfect SIC, the proposed method consistently achieves the highest ESR. This improvement arises from spectrum sharing between cellular users and IoT devices, where the IoT link contributes multipath gain to the cellular link, thereby enhancing overall system performance. To investigate the influence of power allocation, simulation results illustrate the three-dimensional relationship between ESR and power allocation coefficients (Fig. 4). A maximum ESR is observed under specific coefficient combinations, indicating that optimized power allocation can significantly improve system throughput. Furthermore, the proposed optimization scheme demonstrates rapid convergence, with ESR values stabilizing within a few iterations (Fig. 5), supporting its computational efficiency. Finally, ESR performance is compared among the proposed optimization scheme, exhaustive search, and fixed power allocation strategies (Fig. 6). The proposed scheme consistently yields higher ESR across both perfect and imperfect SIC scenarios, demonstrating its superiority in enhancing spectral efficiency while maintaining low computational complexity.  Conclusions  This study proposes a MB-NOMA-CDRT method that enables spectrum sharing between IoT devices and cellular users while improving cellular communication quality through the backscatter-assisted reflection link. To evaluate system performance, closed-form expressions for the ESR are derived under both perfect and imperfect SIC. Building on this analytical foundation, a power allocation optimization scheme based on PSO is developed to maximize the system ESR. Simulation results demonstrate that the proposed method consistently outperforms conventional NOMA-CDRT and OMA schemes in terms of ESR, under both perfect and imperfect SIC conditions. The optimization scheme also exhibits favorable convergence behavior and effectively improves system performance. Given its advantages in spectral efficiency and computational efficiency, the proposed MB-NOMA-CDRT method is well suited to cellular IoT scenarios. Future work will focus on exploring the mathematical conditions necessary to fully characterize and exploit the mutualistic transmission mechanism.
Advancements in Quantum Circuit Design for ARIA: Implementation and Security Evaluation
LI Lingchen, LI Pei, MO Shenyong, WEI Yongzhuang, YE Tao
Available online  , doi: 10.11999/JEIT250440
Abstract:
  Objective  ARIA is established as the Korean national standard block cipher (KS X 1213) in 2003 to meet the demand for robust cryptographic solutions across government, industrial, and commercial sectors in South Korea. Designed by a consortium of Korean cryptographers, the algorithm adopts a hardware-efficient architecture that supports 128-, 192-, and 256-bit key lengths, providing a balance between computational performance and cryptographic security. This design allows ARIA to serve as a competitive alternative to the Advanced Encryption Standard (AES), with comparable encryption and decryption speeds suitable for deployment in resource-constrained environments, including embedded systems and high-performance applications. The security of ARIA is ensured by its Substitution–Permutation Network (SPN) structure, which incorporates two distinct substitution layers and a diffusion layer to resist classical cryptanalytic methods such as differential, linear, and related-key attacks. This robustness has promoted its adoption in secure communication protocols and financial systems within South Korea and internationally. With the emergence of quantum computing, challenges to classical ciphers arise. Quantum algorithms such as Grover’s algorithm reduce the effective key strength of symmetric ciphers, necessitating reassessment of their post-quantum security. In this study, ARIA’s quantum circuit implementation is optimized through tower-field decomposition and in-place circuit optimization, enabling a comprehensive evaluation of its resilience against quantum adversaries.  Methods  The quantum resistance of ARIA is evaluated by estimating the resources required for exhaustive key search attacks under Grover’s algorithm. Grover’s quantum search algorithm achieves quadratic speedup, effectively reducing the security strength of a 128-bit key to the classical equivalent of 64 bits. To ensure accurate assessment, the quantum circuits for ARIA’s encryption and decryption processes are optimized within Grover’s framework, thereby reducing the required quantum resources. The core technique employed is tower-field decomposition, which transforms high-order finite field operations into equivalent lower-order operations, yielding compact computational representations. Specifically, the S-box and linear layer circuits are optimized using automated search tools to identify efficient combinations of low-order field operations. The resulting quantum circuits are then applied to estimate Grover-attack resource requirements, and the results are compared against the National Institute of Standards and Technology (NIST) post-quantum security standards.  Results and Discussions  Optimized quantum circuits for all four ARIA S-boxes are constructed using tower-field decomposition and automated circuit search tools (Fig. 7, Table 2). By integrating these with the linear layer, a complete quantum encryption circuit is implemented, and Grover-attack resource requirements are re-evaluated (Tables 5 and 6). Detailed implementation data are provided in the Clifford+T gate set. The experimental results show that ARIA-192 does not meet the NIST Level 3 post-quantum security standard, indicating vulnerabilities to quantum adversaries. In contrast, ARIA-128 and ARIA-256 comply with Level 1 and Level 3 requirements, respectively. Further optimization is theoretically feasible through methods such as pseudo-key techniques. Future research may focus on developing automated circuit search tools to extend this framework, enabling systematic post-quantum security evaluations of ARIA and comparable symmetric ciphers (e.g., AES, SM4) within a generalized assessment model.  Conclusions  This study investigates the quantum resistance of classical cryptographic algorithms in the context of quantum computing, with a particular focus on the Korean block cipher ARIA. By leveraging the distinct algebraic structures of ARIA’s four S-boxes, tower-field decomposition is applied to design optimized quantum circuits for all S-boxes. Additionally, the circuit depth of the ARIA linear layer is optimized through an in-place quantum circuit implementation, resulting in a more efficient realization of the ARIA algorithm in the quantum setting. A complete quantum encryption circuit is constructed by integrating these optimization components, and the security of the ARIA family of algorithms is evaluated against quantum adversaries using Grover’s key search attack model. The results demonstrate improved implementation efficiency under the newly designed quantum scheme. However, ARIA-192 exhibits resistance below the NIST Level 3 quantum security threshold, indicating a potential vulnerability to quantum attacks.
Research and Design of a Ballistocardiogram-Based Heart Rate Variability (HRV) Monitoring Device Integrated into Pilot Helmets
ZHAO Yanpeng, LI Falin, LI Xuan, YU Haibo, CAO Zhengtao, ZHANG Yi
Available online  , doi: 10.11999/JEIT250342
Abstract:
  Objective  Conventional Heart Rate Variability (HRV) monitoring in aviation is limited by bulky wearable devices that require direct skin contact, are prone to electromagnetic interference during flight, and suffer from electrode displacement during high-G maneuvers. These constraints hinder continuous physiological monitoring, which is critical for flight safety. This study presents a non-contact monitoring approach integrated into pilot helmets, utilizing BallistoCardioGram (BCG) technology to detect cardiac mechanical activity via helmet-mounted inertial sensors. The objective is to establish a novel physiological monitoring paradigm that eliminates the need for skin–electrode interfaces while achieving measurement accuracy suitable for aviation operational standards.  Methods  Hardware ConfigurationA patented BCG sensing module is embedded within the occipital stabilization system of flight protective helmets. Miniaturized, high-sensitivity inertial sensors interface with proprietary signal conditioning circuits that execute a three-stage physiological signal refinement process. First, primary analog amplification scales microvolt-level inputs to measurable voltage ranges. Second, a fourth-order Butterworth bandpass filter (0.5–20 Hz) isolates cardiac mechanical signatures. Third, analog-to-digital conversion quantizes the signals at a 250 Hz sampling rate. Physical integration complies with military equipment standards for helmet structural integrity and ergonomic performance, ensuring full compatibility with existing flight gear without compromising protection or pilot comfort during extended missions.Computational Framework A multi-layer signal processing architecture is implemented to extract physiological features. Raw BCG signals undergo five-level discrete wavelet transformation using Daubechies-4 basis functions, effectively separating cardiac components from respiratory modulation and motion-induced artifacts. J-wave identification is achieved through dual-threshold detection: morphological amplitudes exceeding three times the local baseline standard deviation and temporal positioning within 200 ms sliding analysis windows. Extracted J–J intervals are treated as functional analogs of ElectroCardioGram (ECG)-derived R–R intervals. Time-domain HRV metrics are computed as follows: (1) Standard Deviation of NN intervals (SDNN), representing overall autonomic modulation; (2) Root Mean Square of Successive Differences (RMSSD), indicating parasympathetic activity; (3) Percentage of adjacent intervals differing by more than 50 ms (pNN50). Frequency-domain analysis applies Fourier transformation to quantify Low-Frequency (LF: 0.04–0.15 Hz) and High-Frequency (HF: 0.15–0.4 Hz) spectral powers. The LF/HF ratio is used to assess sympathetic–parasympathetic balance. The entire processing pipeline is optimized for real-time execution under in-flight operational conditions.  Results and Discussions  System validation is conducted under simulated flight conditions to evaluate physiological monitoring performance. Signal acquisition is found to be reliable across static, turbulent, and high-G scenarios, with consistent capture of BCG waveforms. Quantitative comparisons with synchronized ECG recordings show strong agreement between measurement modalities: (1) SDNN: 95.80%; (2) RMSSD: 94.08%; (3) LF/HF ratio: 92.86%. These results demonstrate that the system achieves physiological measurement equivalence to established clinical standards. Artifact suppression is effectively performed by the wavelet-based signal processing framework, which maintains waveform integrity under conditions of aircraft vibration and rapid gravitational transition—conditions where conventional ECG monitoring often fails. Among tested sensor placements, the occipital position exhibits the highest signal-to-noise ratio. Operational stability is maintained during continuous 6-hour monitoring sessions, with no observed signal degradation. This long-duration robustness indicates suitability for extended flight operations.Validation results indicate that the BCG-based approach addresses three primary limitations associated with ECG systems in aviation. The removal of electrode–skin contact mitigates the risk of contact dermatitis during prolonged wear. Non-contact sensing eliminates susceptibility to electromagnetic interference generated by radar and communication systems. Furthermore, mechanical coupling ensures signal continuity during abrupt gravitational changes, which typically displace ECG electrodes and cause signal dropout. The wavelet decomposition method is particularly effective in attenuating rotorcraft harmonic vibrations and turbulence-induced high-frequency noise. Autonomic nervous system modulation is reliably captured through pulse transit time variability, which aligns with neurocardiac regulation indices derived from ECG. Two operational considerations are identified. First, respiratory coupling under hyperventilation may introduce artifacts that require additional filtering. Second, extreme cervical flexion exceeding 45 degrees may degrade signal quality, indicating the potential benefit of redundant sensor configurations under such conditions.  Conclusions  This study establishes a functional, helmet-integrated BCG monitoring system capable of delivering medical-grade HRV metrics without compromising flight safety protocols. The technology represents a shift from contact-based to non-contact physiological monitoring in aviation settings. Future system development will incorporate: (1) Infrared eye-tracking modules to assess blink interval variability for objective fatigue evaluation; (2) Dry-contact electroencephalography sensors to quantify prefrontal cortex activity and assess cognitive workload; (3) Multimodal data fusion algorithms to generate unified indices of physiological strain. The integrated framework aims to enable real-time pilot state awareness during critical operations such as aerial combat maneuvers, hypoxia exposure, and emergency responses. Further technology maturation will prioritize operational validation across diverse aircraft platforms and environmental conditions. System implementation remains fully compliant with military equipment specifications and is positioned for future translation to commercial aviation and human factors research. Broader applications include astronaut physiological monitoring during spaceflight missions and enhanced safety systems in high-performance motorsports.
Adversarial Transferability Attack on Deep Neural Networks Through Spectral Coefficient Decay
QIAN Yaguan, KONG Yaxin, CHEN Kecheng, SHEN Yunkai, BAO Qiqi, JI Shouling
Available online  , doi: 10.11999/JEIT250157
Abstract:
  Objective  The rapid advancement of machine learning has accelerated the development of artificial intelligence applications based on big data. Deep Neural Networks (DNNs) are now widely used in real-world systems, including autonomous driving and facial recognition. Despite their advantages, DNNs remain vulnerable to adversarial attacks, particularly in image recognition, which exposes significant security risks. Attackers can generate carefully crafted adversarial examples that appear visually indistinguishable from benign inputs but cause incorrect model predictions. In critical applications, such attacks may lead to system failures. Studying adversarial attack strategies is crucial for designing effective defenses and improving model robustness in practical deployments.  Methods  This paper proposes a black-box adversarial attack method termed Spectral Coefficient Attenuation (SCA), based on Spectral Coefficient Decay (SCD). SCA perturbs spectral coefficients during the iterative generation of adversarial examples by attenuating the amplitude of each frequency component. This attenuation introduces diverse perturbation patterns and reduces the dependency of adversarial examples on the source model, thereby improving their transferability across models.  Results and Discussions  This study generated adversarial examples using three naturally trained deep models based on Convolutional Neural Network (CNN) architectures and evaluated their attack success rates against several widely used convolutional image recognition models. Compared with MI-FGSM, the proposed method achieved average increases in attack success rates of 4.7%, 4.1%, and 4.9%, respectively. To further assess the effectiveness of the SCA algorithm, additional experiments were conducted on vision models based on Transformer architectures. In these settings, the SCA method yielded average success rate improvements of 4.6%, 4.5%, and 3.7% compared with MI-FGSM.  Conclusions  The proposed method attenuates the energy of individual spectral components during the iterative generation of adversarial examples in the frequency domain. This reduces reliance on specific features of the source model, mitigates the risk of overfitting, and enhances black-box transferability across models with different architectures. Extensive experiments on the ImageNet validation set confirm the effectiveness of the SCA method. Whether targeting CNN-based or Transformer-based models, the SCA algorithm consistently achieves high transferability and robust attack performance.
Depression Screening Method Driven by Global-Local Feature Fusion
ZHANG Siyong, QIU Jiefan, ZHAO Xiangyun, XIAO Kejiang, CHEN Xiaofu, MAO Keji
Available online  , doi: 10.11999/JEIT250035
Abstract:
  Objective  Depression is a globally prevalent mental disorder that poses a serious threat to the physical and mental health of millions of individuals. Early screening and diagnosis are essential to reducing severe consequences such as self-harm and suicide. However, conventional questionnaire-based screening methods are limited by their dependence on the reliability of respondents’ answers, their difficulty in balancing efficiency with accuracy, and the uneven distribution of medical resources. New auxiliary screening approaches are therefore needed. Existing Artificial Intelligence (AI) methods for depression detection based on facial features primarily emphasize global expressions and often overlook subtle local cues such as eye features. Their performance also declines in scenarios where partial facial information is obscured, for instance by masks, and they raise privacy concerns. This study proposes a Global-Local Fusion Axial Network (GLFAN) for depression screening. By jointly extracting global facial and local eye features, this approach enhances screening accuracy and robustness under complex conditions. A corresponding dataset is constructed, and experimental evaluations are conducted to validate the method’s effectiveness. The model is deployed on edge devices to improve privacy protection while maintaining screening efficiency, offering a more objective, accurate, efficient, and secure depression screening solution that contributes to mitigating global mental health challenges.  Methods  To address the challenges of accuracy and efficiency in depression screening, this study proposes GLFAN. For long-duration consultation videos with partial occlusions such as masks, data preprocessing is performed using OpenFace 2.0 and facial keypoint algorithms, combined with peak detection, clustering, and centroid search strategies to segment the videos into short sequences capturing dynamic facial changes, thereby enhancing data validity. At the model level, GLFAN adopts a dual-branch parallel architecture to extract global facial and local eye features simultaneously. The global branch uses MTCNN for facial keypoint detection and enhances feature extraction under occlusion using an inverted bottleneck structure. The local branch detects eye regions via YOLO v7 and extracts eye movement features using a ResNet-18 network integrated with a convolutional attention module. Following dual-branch feature fusion, an integrated convolutional module optimizes the representation, and classification is performed using an axial attention network.  Results and Discussions  The performance of GLFAN is evaluated through comprehensive, multi-dimensional experiments. On the self-constructed depression dataset, high accuracy is achieved in binary classification tasks, and non-depression and severe depression categories are accurately distinguished in four-class classification. Under mask-occluded conditions, a precision of 0.72 and a precision of 0.690 are obtained for depression detection. Although these values are lower than the precision of 0.87 and precision of 0.840 observed under non-occluded conditions, reliable screening performance is maintained. Compared with other advanced methods, GLFAN achieves higher recall and F1 scores. On the public AVEC2013 and AVEC2014 datasets, the model achieves lower Mean Absolute Error (MAE) values and shows advantages in both short- and long-sequence video processing. Heatmap visualizations indicate that GLFAN dynamically adjusts its attention according to the degree of facial occlusion, demonstrating stronger adaptability than ResNet-50. Edge device tests further confirm that the average processing delay remains below 17.56 milliseconds per frame, and stable performance is maintained under low-bandwidth conditionsThe performance of GLFAN is evaluated through comprehensive, multi-dimensional experiments. On the self-constructed depression dataset, high accuracy is achieved in binary classification tasks, and non-depression and severe depression categories are accurately distinguished in four-class classification. Under mask-occluded conditions, a precision of 0.72 and a recall of 0.690 are obtained for depression detection. Although these values are lower than the precision of 0.87 and recall of 0.840 observed under non-occluded conditions, reliable screening performance is maintained. Compared with other advanced methods, GLFAN achieves higher recall and F1 scores. On the public AVEC2013 and AVEC2014 datasets, the model achieves lower Mean Absolute Error (MAE) values and shows advantages in both short- and long-sequence video processing. Heatmap visualizations indicate that GLFAN dynamically adjusts its attention according to the degree of facial occlusion, demonstrating stronger adaptability than ResNet-50. Edge device tests further confirm that the average processing delay remains below 17.56 frame/s, and stable performance is maintained under low-bandwidth conditions.  Conclusions  This study proposes a depression screening approach based on edge vision technology. A lightweight, end-to-end GLFAN is developed to address the limitations of existing screening methods. The model integrates global facial features extracted via MTCNN with local eye-region features captured by YOLO v7, followed by effective feature fusion and classification using an Axial Transformer module. By emphasizing local eye-region information, GLFAN enhances performance in occluded scenarios such as mask-wearing. Experimental validation using both self-constructed and public datasets demonstrates that GLFAN reduces missed detections and improves adaptability to short-duration video inputs compared with existing models. Grad-CAM visualizations further reveal that GLFAN prioritizes eye-region features under occluded conditions and shifts focus to global facial features when full facial information is available, confirming its context-specific adaptability. The model has been successfully deployed on edge devices, offering a lightweight, efficient, and privacy-conscious solution for real-time depression screening.
A Novel Transient Execution Attack Exploiting Loop Prediction Mechanisms
GUO Jiayi, QIU Pengfei, YUAN Jie, LAN Zeru, WANG Chunlu, ZHANG Jiliang, WANG Dongsheng
Available online  , doi: 10.11999/JEIT250361
Abstract:
  Objective  Modern processors rely heavily on branch prediction to improve pipeline efficiency; however, the transient execution windows created by speculative execution expose critical security vulnerabilities. While prior research has primarily examined conditional branch instructions, this study identifies a previously overlooked attack surface: loop instructions (LOOP, LOOPZ, LOOPNZ) and JRCXZ in x86 architectures, which use the RCX register to determine branch outcomes. These instructions produce significantly longer transient windows than JCC instructions, posing heightened threats to hardware-level isolation. This work demonstrates the exploitability of these instructions, quantifies their transient execution behavior, and validates practical attack scenarios.  Methods  This study employs a systematic methodology to investigate the speculative behavior of loop instructions and assess their exploitability. First, the microarchitectural behavior of LOOP, LOOPZ, LOOPNZ, and JRCXZ instructions is reverse-engineered using Performance Monitoring Counters (PMCs), with a focus on their dependency on RCX register values and interaction with the branch prediction unit. Speculative durations of loop and JCC instructions are compared using cycle-accurate profiling via the RDPMC instruction, which accesses fixed-function PMCs to record clock cycles. Based on these observations, exploit primitives are constructed by manipulating RCX values to induce speculative execution paths. The feasibility of these primitives is evaluated through four real-world attack scenarios on Intel CPUs: (1) Cross-user/kernel data leakage through speculative memory access following mispredicted loop exits. (2) Covert channel creation between Simultaneous MultiThreading (SMT) threads by measuring timing differences between correctly and incorrectly predicted branches during speculative execution. (3) SGX enclave compromise via speculative access to secrets gated by RCX-controlled branching. (4) Kernel Address Space Layout Randomization (KASLR) bypass using page fault timing during transient execution of loop-based probes. Each scenario is tested on real hardware under controlled conditions to assess reliability, reproducibility, and attack robustness.  Results and Discussions  The proposed transient execution attack targeting loop instructions (LOOP, LOOPZ, LOOPNZ) and JRCXZ offers notable advantages over traditional Spectre exploits. These RCX-dependent instructions exhibit transient execution windows that are, on average, 40% longer than those of conventional JCC branches (Table 1). The extended speculative duration significantly improves attack reliability: in cross-user/kernel boundary experiments, the proposed method achieves an average data leakage accuracy of 90%, compared to only 10% for JCC-based techniques under identical conditions. The attack also demonstrates high efficacy in bypassing hardware isolation mechanisms. In Intel SMT environments, a covert channel is established with 97.5% accuracy and a throughput of 256.9 kbit/s (Table 4), exploiting timing discrepancies between correctly and incorrectly predicted branches during speculative execution. In trusted execution environments, the attack achieves 98% accuracy in extracting secret values from Intel SGX enclaves, highlighting the susceptibility of RCX-controlled speculation to enclave compromise. Additionally, KASLR is completely defeated by exploiting speculative page fault timing during loop instruction execution. Kernel base addresses are recovered deterministically in all test cases (Fig. 4), demonstrating the critical security implications of this attack vector.  Conclusions  This study identifies a critical vulnerability in modern speculative execution mechanisms by demonstrating that loop instructions (LOOP, LOOPZ, LOOPNZ) and JRCXZ—which rely on the RCX register for branch decisions, serve as novel vectors for transient execution attacks. The key contributions are threefold: (1) These instructions generate speculative execution windows that are, on average, 40% longer than those of JCC instructions. (2) Practical exploits are demonstrated across key hardware isolation boundaries—including user/kernel space, SMT, and Intel SGX enclaves, with success rates exceeding 90% in targeted scenarios. (3) The findings expose critical limitations in current Spectre defenses, indicating that existing mitigations are insufficient to address RCX-dependent speculative paths, thereby motivating the need for specialized countermeasures.
A Survey of Processor Hardware Vulnerability
LAN Zeru, QIU Pengfei, WANG Chunlu, ZHAO Yaxuan, JIN Yu, ZHANG Zhihao, WANG Dongsheng
Available online  , doi: 10.11999/JEIT250357
Abstract:
  Significance  Processor security is a cornerstone of computer system security, providing a trusted execution environment for upper-layer systems and applications. However, the increasing complexity of processor microarchitectures and the widespread integration of performance-driven optimization mechanisms have introduced significant security risks. These mechanisms, primarily designed to enhance performance and energy efficiency, often lack comprehensive security evaluation, thereby expanding the potential attack surface. Therefore, numerous microarchitectural security vulnerabilities have emerged, presenting critical challenges in architectural security research.  Progress  Although recent years have witnessed notable progress in the study of hardware vulnerabilities, several key issues remain unresolved. First, the landscape of hardware vulnerabilities is both diverse and complex, yet existing literature lacks a consistent and systematic classification framework. This gap complicates researchers’ efforts to understand, compare, and generalize vulnerability characteristics. Second, current studies predominantly focus on individual vulnerability discovery or specific attack implementations, with limited attention to modeling the full vulnerability lifecycle. A comprehensive research framework including vulnerability identification, attack instantiation, and exploitation is still lacking. One pressing challenge is how to efficiently and systematically convert potential vulnerabilities into practical, high-risk attack paths. In addition, unlike software vulnerabilities, hardware vulnerabilities are inherently more difficult to mitigate and impose higher defense costs. These characteristics highlight the need for a more structured and integrated approach to hardware vulnerability research.  Contributions  This paper systematically reviews and analyzes processor hardware vulnerabilities reported in major architecture security conferences and academic journals since 2010. It first outlines four primary methods for discovering hardware vulnerabilities and, based on prior studies, proposes a three-step attack model and a novel attack scenario framework. The paper then categorizes and describes existing hardware vulnerabilities according to their behavioral characteristics and consolidates eight evaluation metrics for side-channel vulnerabilities derived from related research. To assess the feasibility and scope of various attack types, representative vulnerabilities are selected for experimental validation across multiple processor platforms, with in-depth analysis of the results. In addition, the study provides a systematic evaluation of current defense and mitigation mechanisms for hardware vulnerabilities. Finally, it discusses future research directions from both offensive and defensive perspectives.  Prospects   Future research in processor hardware security is expected to focus on new attack surfaces introduced by increasingly diversified microarchitectural optimizations. Key areas will include the development of system-level collaborative defense mechanisms, automated verification tools, and integrated strategies to enhance awareness and precision in mitigating hardware-level information leakage risks.
MOS-gated Prebond Through-Silicon Via Testing
DOU Xianrui, LIANG Huaguo, HUANG Zhengfeng, LU Yingchun, CHEN Tian, LIU Jun
Available online  , doi: 10.11999/JEIT250285
Abstract:
  Objective  As the miniaturization of semiconductor chips approaches physical limitations, integrated chip technologies have become essential to meet the demand for high-performance, low-cost devices in the post-Moore era. Through-Silicon Via (TSV) is a key process in advanced packaging that requires precise testing to ensure reliable interconnections. Quantitative test methods can estimate defect sizes based on test responses; however, variations in Process, Voltage, and Temperature (PVT) hinder accurate defect characterization, making the associated overhead of data capture and analysis difficult to justify. Current techniques often require long test time, with some necessitating two test cycles. While leakage defect detection has reached high accuracy, the detection of resistive open defects—sometimes only tens of milliohms in fault-free states—remains inadequate. This study presents a method that improves detection accuracy for resistive open defects and reduces both test area and time overhead, offering a more efficient and practical TSV testing solution.  Methods  Previous studies indicate that rising-edge testing provides higher resolution than falling-edge testing and enables simultaneous differentiation of multiple defect types. Based on this principle, a symmetric testing scheme through a single rising-edge test is proposed. To reduce the area overhead associated with shared test structures, MOS gates are employed as selection switches. NMOS transistors, due to their strong 0 and weak 1 characteristics, are placed at the driving end to enable rapid discharge and reset of the reference capacitor voltage. PMOS transistors, exhibiting strong 1 and weak 0 characteristics, are positioned at the receiving end to block interference from low-voltage signals. A two-stage comparator is then employed to amplify the voltage difference between the reference capacitor and the test TSV during the charging phase, producing two intermediate voltage levels. These are subsequently converted into standard high or low logic levels by a Schmitt trigger inverter. Based on the output logic level, both the presence and type of defect can be determined from a single test.  Results and Discussions  The effectiveness of the proposed method is verified through HSPICE simulations using the Nangate 45 nm open cell library. The detection accuracy for different defect types is modulated by adjusting the Width-to-Length (W/L) ratio of the MOS transistors, as shown in (Table 2). For instance, reducing the W/L ratio of NMOS transistors enhances the detection sensitivity to leakage defects. Specific W/L ratios can therefore be selected to meet targeted testing requirements. (Table 3) presents the results under PVT variations. Although the accuracy shows minor fluctuations, these remain within acceptable limits. A temperature variation of approximately 27 °C results in only a 1 Ω deviation in resistive open defect detection, and a 1 MΩ range in leakage defect accuracy. Even under the worst-case PVT condition, the minimum detection threshold for resistive open defects reaches 94 Ω, which exceeds the capability of existing methods.  Conclusions  A prebond TSV testing scheme based on MOS gating is proposed to address the high area and time overheads and limited accuracy of conventional approaches. The scheme adopts a symmetric structure between the reference capacitor and the test TSV to mitigate capacitance variation caused by fabrication inconsistencies. A two-stage comparator amplifies the voltage difference between the defective TSV and the reference capacitor during charging, thereby enhancing detection resolution. Simulation results indicate that the method detects resistive open defects equal of above 50 Ω and leakage defects equal of below 9 MΩ. Compared with existing methods, the proposed approach significantly reduces both testing area and time. When multiple TSVs share the testing circuitry, only one NMOS and one PMOS transistor are added, further minimizing the average area overhead.
Research on Key Technologies of Side-channel Security Protection for Polynomial Multiplication in ML-KEM/Kyber Algorithm
ZHAO Yiqiang, KONG Jindi, FU Yucheng, ZHANG Qizhi, YE Mao, XIA Xianzhao, SONG Xintong, HE Jiaji
Available online  , doi: 10.11999/JEIT250292
Abstract:
  Objective  As ML-KEM/Kyber is adopted as a post-quantum key encapsulation mechanism, securing its hardware implementations against Side-Channel Attacks (SCAs) has become critical. Although Kyber offers mathematically proven security, its physical implementations remain susceptible to timing-based side-channel leakage, particularly during Polynomial Point-Wise Multiplication (PWM), a core operation in decryption. Existing countermeasures, such as masking and static hiding, struggle to balance security, resource efficiency, and hardware feasibility. This study proposes a dynamic randomization strategy to disrupt execution timing patterns in PWM, thereby improving side-channel resistance in Kyber hardware designs.  Methods  A randomized pseudo-round hiding technique is developed to obfuscate the timing profile of PWM computations. The approach incorporates two key mechanisms: (1) dynamic insertion of redundant modular operations (e.g., dummy additions and multiplications), and (2) two-level pseudo-random scheduling based on Linear Feedback Shift Registers (LFSRs). These mechanisms randomize the execution order of PWM operations while reusing existing butterfly units to reduce hardware overhead. The design is implemented on a Xilinx Spartan-6 FPGA and evaluated using Correlation Power Analysis (CPA) and Test Vector Leakage Assessment (TVLA).  Results and Discussions  Experimental results demonstrate a substantial improvement in side-channel resistance. In unprotected implementations, attackers could recover Kyber’s long-term secret key using as few as 897 to 1,650 power traces. With the proposed countermeasure applied, no successful key recovery occurred even after 10,000 traces, representing more than a 100-fold increase in the number of traces required for key extraction. TVLA results (Fig. 6) confirm the suppression of leakage, with t-test values maintained near the threshold (|t| < 4.5). The resource overhead remains within acceptable bounds: the area-time product increases by 17.99%, requiring only 157 additional Look-Up Tables (LUTs) and 99 Flip-Flops (FFs) compared with the unprotected design. The proposed architecture outperforms existing masking and hiding schemes (Table 3), delivering stronger security with lower resource consumption.  Conclusions  This work presents an efficient and lightweight countermeasure against timing-based SCAs for Kyber hardware implementations. By dynamically randomizing PWM operations, the design significantly enhances side-channel security while maintaining practical resource usage. Future research will focus on optimizing pseudo-round scheduling to reduce latency, extending protection to Kyber’s Fujisaki–Okamoto (FO) transformation modules, and generalizing the method to other Number-Theoretic Transform (NTT)-based lattice cryptographic algorithms such as Dilithium. These developments support the secure and scalable deployment of post-quantum cryptographic systems.
Gate-level Side-Channel Protection Method Based on Hybrid-order Masking
ZHAO Yiqiang, LI Zhengyang, ZHANG Qizhi, YE Mao, XIA Xianzhao, LI Yao, HE Jiaji
Available online  , doi: 10.11999/JEIT250198
Abstract:
  Objective  Side-Channel Analysis (SCA) presents a significant threat to the hardware implementation of cryptographic algorithms. Among various sources of side-channel leakage, power consumption is particularly vulnerable due to its ease of extraction and interpretation, making power analysis one of the most prevalent SCA techniques. To address this threat, masking has been widely adopted as a countermeasure in hardware security. Masking introduces randomness to disrupt the correlation between sensitive intermediate data and observable side-channel information, thereby enhancing resistance to SCA. However, existing masking approaches face notable limitations. Algorithm-level masking requires comprehensive knowledge of algorithmic structure and does not reliably strengthen hardware-level security. Masking applied at the Register Transfer Level (RTL) is prone to structural alterations during hardware synthesis and is constrained by the need for logic optimization, limiting scalability. Gate-level masking offers certain advantages, yet such approaches depend on precise localization of leakage and often incur unpredictable overhead after deployment. Furthermore, many masking schemes remain susceptible to higher-order SCA techniques. To overcome these limitations, there is an urgent need for gate-level masking strategies that provide robust security, maintain acceptable overhead, and support scalable deployment in practical hardware systems.  Methods  To address advances in SCA techniques and the limitations of existing masking schemes, this paper proposes a hybrid-order masking method. The approach is specifically designed for gate-level netlist circuits to provide fine-grained and precise protection. By considering the structural characteristics of encryption algorithm circuits, the method integrates masking structures of different orders according to circuit requirements, introduces randomness to sensitive variables, and substantially improves resistance to side-channel attacks. In parallel, the approach accounts for potential hardware overhead to maintain practical feasibility. Theoretical security is verified through statistical evaluation combined with established SCA techniques. An automated deployment framework is developed to facilitate rapid and efficient application of the masking scheme. The framework incorporates functional modules for circuit topology analysis, leakage identification, and masking deployment, supporting a complete workflow from circuit analysis to masking implementation. The security performance of the masked design is further assessed through correlation-based evaluation methods and simulation.  Results and Discussions  The automated masking deployment tool is applied to implement gate-level masking for Advanced Encryption Standard (AES) circuits. The security of the masked design is evaluated through first-order and higher-order power analysis in simulation. The correlation coefficient and Minimum Traces to Disclosure (MTD) parameter serve as the primary evaluation metrics, both widely used in side-channel security assessment. The MTD reflects the number of power traces required to extract the encryption key from the circuit. In first-order power analysis, the unmasked design exhibits a maximum correlation value of 0.49 for the correct key (Fig. 6(a)), and the correlation curve for the correct key is clearly separated from those of incorrect keys. By contrast, the masked design reduces the correlation to approximately 0.02 (Fig. 6(b)), with no evidence of successful key extraction. Based on the MTD parameter, the unmasked design requires 116 traces for key disclosure, whereas the masked design requires more than 200,000 traces, reflecting an improvement exceeding 1724 times (Fig. 7). Higher-order power analysis yields consistent results. The unmasked design demonstrates an MTD of 120 traces, indicating clear vulnerability, whereas the masked design maintains a maximum correlation near 0.02 (Fig. 8) and an MTD greater than 200,000 traces (Fig. 9), corresponding to a 1667-fold improvement. In terms of hardware overhead, the masked design shows a 1.2% increase in area and a 41.1% reduction in maximum operating frequency relative to the unmasked circuit.  Conclusions  This study addresses the limitations of existing masking schemes by proposing a hybrid-order masking method that disrupts the conventional definition of protection order. The approach safeguards sensitive data during cryptographic algorithm operations and enhances resistance to SCA in gate-level hardware designs. An automated deployment tool is developed to efficiently integrate vulnerability identification and masking protection, supporting practical application by hardware designers. The proposed methodology is validated through correlation analysis across different orders. The results demonstrate that the method improves resistance to power analysis attacks by more than 1600 times and achieves significant security enhancement with minimal hardware overhead compared to existing masking techniques. This work advances the current knowledge of masking strategies and provides an effective approach for improving hardware-level security. Future research will focus on extending the method to broader application scenarios and enhancing performance through algorithmic improvements.
Analyzing and Mitigating Asymmetric Residual Stress in 3D NAND Scaling Based on Process-dependent Modeling
CUI Hanwen, GAO Yanze, ZHANG Kun, WANG Shizhao, TIAN Zhiqiang, GUO Yuzheng, XIA Zhiliang, ZHANG Zhaofu, HUO Zongliang, LIU Sheng
Available online  , doi: 10.11999/JEIT250410
Abstract:
  Objective  To improve the performance of 3D NAND architecture, a series of horizontal and vertical miniaturization strategies have been proposed. While these designs increase storage density, they also introduce integration challenges. In particular, thermo-mechanical stress during fabrication has become a critical limitation on device yield and performance. This study establishes a high-precision process mechanics model of 3D NAND based on a local Representative Volume Element (RVE) finite element modeling framework, accounting for the multilayer stacked structure and various block architecture designs. By systematically investigating stress evolution during fabrication, the analysis identifies the root causes of stress non-uniformity and characterizes the dynamic distribution of mechanical stress under different miniaturization schemes. These findings have practical relevance for yield improvement and device reliability, addressing key challenges in advancing 3D NAND storage density.  Methods  This study constructs a high-precision, device-level finite element model of 3D NAND based on the theory of RVE. The simulation of thermal stress evolution throughout the manufacturing process uses the element birth/death technique in Abaqus. The baseline model features a representative 3D NAND structure comprising 8 Nitride/Oxide (N/O) bilayers, each 25 nm thick. Within a 40-nm-wide slit, 15 storage pillars, each with a diameter of 24 nm and spaced at 36 nm intervals, are arranged in a staggered configuration. To explore the effect of stacking layer number on stress evolution, modified models with 6 and 10 N/O layers are also developed. In addition, to examine the effect of different block architecture transitions, models incorporating 5 and 10 pillars per block are analyzed. The material properties used are consistent with those reported in previous studies, where both the calibration of material parameters and the modeling methodology are validated.  Results and Discussions  Process-dependent simulations were conducted to examine the evolution of stress distribution during key 3D NAND fabrication steps and to assess the effects of vertical stacking layers (Fig. 7) and block architecture designs (Fig. 8). The results show that metal volume fraction, the number of pillars in the array region, and the presence of oxide stairs are primary factors influencing stress distribution. A higher metal volume fraction markedly increases internal stress due to thermal expansion mismatch. Asymmetric metal layouts in the Word Line (WL) and Bit Line (BL) directions intensify stress anisotropy between these axes. Pillars in the array region help alleviate stress concentration by generating tensile zones during nitride/metal thermal deformation, thereby reducing the overall compressive stress. In contrast, oxide stairs constrain deformation along the WL direction, inhibiting stress relaxation and resulting in localized compressive regions. These combined mechanisms indicate that increasing the number of WL layers tends to enhance stress asymmetry, whereas block architectures with a larger number of pillars reduce the degree of stress non-uniformity.  Conclusions  Using a process mechanics model based on the RVE approach, this study explored stress evolution in 3D NAND fabrication. The effects of two major scaling strategies—vertical layer stacking and horizontal block architecture conversion—were systematically analyzed with respect to stress magnitude and directional asymmetry. The results show that asymmetric stress distribution originates during the step etching stage and peaks following WL and slot filling. As the number of vertical stacking layers increases, structural compressive stress intensifies, particularly in the WL and BL directions. Increasing the number of layers from 6 to 10 results in an 8.54 MPa rise in WL compressive stress and a 5.66 MPa rise in BL stress, with the WL–BL stress difference increasing from 20.76 MPa to 24.64 MPa. Larger-area block architectures effectively mitigate stress asymmetry. Compared with the 5-pillar configuration, the 15-pillar architecture reduces WL–BL stress asymmetry by 22.4%. The composite structure of oxide and tungsten, combined with the constraint effects of pillars and stepped oxide on sacrificial layer deformation, plays a central role in modulating stress levels and directional distribution in 3D NAND structures.
Data Enhancement for Few-shot Radar Countermeasure Reconnaissance via Temporal-Conditional Generative Adversarial Networks
HUANG Linxuan, HE Minghao, YU Chunlai, FENG Mingyue, ZHANG Fuqun, ZHANG Yinan
Available online  , doi: 10.11999/JEIT250280
Abstract:
  Objective  Radar electronic warfare systems rely on Pulse Descriptor Words (PDWs) to represent radar signals, capturing key parameters such as Time of Arrival (TOA), Carrier Frequency (CF), Pulse Width (PW), and Pulse Amplitude (PA). However, in complex electromagnetic environments, the scarcity of PDW data limits the effectiveness of data-driven models in radar pattern recognition. Conventional augmentation methods (e.g., geometric transformations, SMOTE) fall short in addressing three core issues: (1) failure to capture temporal physical laws (e.g., gradual frequency shifts and pulse modulation patterns); (2) distributional inconsistencies (e.g., frequency band overflows and PW discontinuities); and (3) weak coupling between PDWs and modulation types, leading to reduced classification accuracy. This study proposes Time-CondGAN, a temporal-conditional generative adversarial network designed to generate physically consistent and modulation-specific PDW sequences, enhancing few-shot radar reconnaissance performance.  Methods   Time-CondGAN integrates three core innovations: (1) Multimodal Conditional Generation Framework: A label encoder maps discrete radar modulation types (e.g., VS, HRWS, TWS in Table 2) into 128-dimensional feature vectors. These vectors are temporally expanded and concatenated with latent noise to enable category-controllable generation. The generator architecture (Fig. 4) employs bidirectional Gated Recurrent Units (GRUs) to capture long-range temporal dependencies. (2) Multi-task Discriminator Design: The discriminator (Fig. 5) is designed for joint adversarial discrimination and signal classification. A shared bidirectional GRU with an attention mechanism extracts temporal features, while classification loss constrains the generator to maintain category-specific distributions. (3) Temporal-Statistical Joint Optimization: The training process incorporates multiple loss components: Supervisor Loss (Eq. 7): A bidirectional GRU-based supervisor enforces temporal consistency. Feature Matching Loss (Eq. 9): Encourages alignment between high-level features of real and synthetic signals. Adversarial and Classification Losses (Eqs. (8)–(9)): Promote distribution realism and accurate category separation. A two-stage training strategy is adopted to improve stability, consisting of pre-training (Fig. 2) followed by adversarial training (Fig. 3).  Results and Discussions  Time-CondGAN demonstrates strong performance across three core dimensions. (1) Physical plausibility is achieved through accurate modeling of radar signal dynamics. Compared with TimeGAN, Time-CondGAN reduces the Kullback-Leibler (KL) divergence by 28.25% on average. Specifically, the PW distribution error decreases by 50.20% (KL = 4.68), the TOA interval error decreases by 20.5% (KL = 0.636), and the CF deviation is reduced by 13.29% (KL = 7.93), confirming the suppression of non-physical signal discontinuities. (2) Downstream task enhancement highlights the model’s few-shot generation capability. With only 10 real samples, classification accuracy improves markedly: VGG16 accuracy increases by 37.2% (from 43.0% to 59.0%) and LSTM accuracy by 28.6% (from 49.0% to 63.0%), both substantially outperforming conventional data augmentation methods. (3) Ablation studies validate the contribution of key modules. Removing the conditional encoder increases the PW KL divergence by 107.4%. Excluding the supervisor loss degrades CF continuity by 76.2%, and omitting feature matching results in a 44.6% misalignment in amplitude distribution.  Conclusions  This study establishes Time-CondGAN as an effective solution for radar PDW generation under few-shot conditions, addressing three key limitations: temporal fragmentation is resolved through bidirectional GRU supervision, mode collapse is alleviated via multimodal conditioning, and modulation specificity is maintained by classification constraints. The proposed framework offers operational value, enabling over 35% improvement in recognition accuracy under low-intercept conditions. Future work will incorporate radar propagation physics to refine amplitude modeling (current KL = 8.51), adopt meta-learning approaches for real-time adaptation in dynamic battlefield environments, and expand the model to multi-radar cooperative scenarios to support heterogeneous electromagnetic contexts.
A Lightweight Semantic Visual Simultaneous Localization and Mapping Framework for Inspection Robots in Dynamic Environments
YU Haoyang, LI Yansheng, XIAO Lingli, ZHOU Jiyuan
Available online  , doi: 10.11999/JEIT250301
Abstract:
  Objective  In complex dynamic environments such as industrial parks and urban roads, inspection robots depend heavily on visual Simultaneous Localization And Mapping (SLAM) systems. However, the presence of moving objects often causes feature drift and map degradation, reducing SLAM performance. Furthermore, conventional semantic segmentation models typically require extensive computational resources, rendering them unsuitable for embedded platforms with limited processing capabilities, thereby constraining SLAM deployment in autonomous inspection tasks. To address these challenges, this study proposes a lightweight semantic visual SLAM framework designed for inspection robots operating in dynamic environments. The framework incorporates a semantic segmentation-based dynamic feature rejection method to achieve real-time identification of dynamic regions at low computational cost, thereby improving SLAM robustness and mapping accuracy.  Methods  Building upon the 11th generation lightweight YOLO segmentation model (YOLOv11n-seg), a systematic lightweight redesign is implemented. to enhance performance under constrained computational resources. First, the original neck is replaced with DyCANet, a lightweight multi-scale feature fusion module that integrates dynamic point sampling and channel attention to improve semantic representation and boundary segmentation. DyCANet combines DySample, a dynamic upsampling operator that performs content-aware spatial sampling with minimal overhead, and ChannelAttention_HSFPN, a hierarchical attention structure that strengthens multi-scale integration and highlights critical semantic cues, particularly for small or occluded objects in complex scenes. Second, a Dynamic Convolution module (DynamicConv) is embedded into all C3k2 modules to enhance the adaptability and efficiency of feature extraction. Inspired by the Mixture-of-Experts framework, DynamicConv applies a conditional computation mechanism that dynamically adjusts kernel weights based on the input feature characteristics. This design allows the network to extract features more effectively across varying object scales and motion patterns, improving robustness against dynamic disturbances with low computational cost. Third, the original segmentation head is replaced by the Reused and Shared Convolutional Segmentation Head (RSCS Head), which enables decoder structure sharing across multi-scale branches. RSCS reduces redundant computation by reusing convolutional layers and optimizing feature decoding paths, further improving overall model efficiency while maintaining segmentation accuracy. These architectural modifications result in DHSR-YOLOSeg, a lightweight semantic segmentation model that significantly reduces parameter count and computational cost while preserving performance. DHSR-YOLOSeg is integrated into the tracking thread of ORB-SLAM3. to provide real-time semantic information. This enables dynamic object detection and the removal of unstable feature points during localization, thereby enhancing the robustness and trajectory consistency of SLAM in complex dynamic environments.  Results and Discussions  Ablation experiments on the COCO dataset demonstrate that, compared with the baseline YOLOv11n-seg, the proposed DHSR-YOLOSeg achieves a 13.8% reduction in parameter count, a 23.1% decrease in Giga Floating Point Operations (GFLOPs), and an approximate 2% increase in mean Average Precision at IoU 0.5 (mAP50) (Table 1). On the KITTI dataset, DHSR-YOLOSeg reaches an inference speed of 60.19 frame/s, which is 2.14% faster than YOLOv11n-seg and 275% faster than the widely used Mask R-CNN (Table 2). For trajectory accuracy evaluation on KITTI sequences 00~10, DHSR-YOLOSeg outperforms ORB-SLAM3 in 8 out of 10 sequences, achieving a maximum Root Mean Square Error (RMSE) reduction of 16.76% and an average reduction of 8.78% (Table 3). Compared with DynaSLAM and DS-SLAM, the proposed framework exhibits more consistent error suppression across sequences, improving both trajectory accuracy and stability. In terms of runtime efficiency, DHSR-YOLOSeg achieves an average per-frame processing time of 48.86 ms on the KITTI dataset, 18.44% and 41.38% lower than DS-SLAM and DynaSLAM, respectively (Table 4). The per-sequence processing time ranges from 41 to 55 ms, which is comparable to the 35.64 ms of ORB-SLAM3, indicating that the integration of semantic segmentation introduces only a modest computational overhead.  Conclusions  This study addresses the challenge of achieving robust localization for inspection robots operating in dynamic environments, particularly in urban road settings characterized by frequent interference from pedestrians, vehicles, and other moving objects. To this end, a semantic-enhanced visual SLAM framework is proposed, in which a lightweight semantic segmentation model, DHSR-YOLOSeg, is integrated into the stereo-based ORB-SLAM3 pipeline. This integration enables real-time identification of dynamic objects and removal of their associated feature points, thereby improving localization robustness and trajectory consistency. The DHSR-YOLOSeg model incorporates three key architectural components—DyCANet for feature fusion, DynamicConv for adaptive convolution, and the RSCS Head for efficient multi-scale decoding. Together, these components reduce the model’s size and computational cost while preserving segmentation performance, providing an efficient and deployable perception solution for resource-constrained platforms. Experimental findings show that: (1) ablation tests on the COCO dataset confirm substantial reductions in complexity with preserved accuracy, supporting embedded deployment; (2) frame rate comparisons on the KITTI dataset demonstrate superior performance over both lightweight and standard semantic segmentation methods, meeting real-time SLAM requirements; (3) trajectory evaluations indicate that the dynamic feature rejection strategy effectively mitigates localization errors in dynamic scenes; and (4) the overall system maintains high runtime efficiency, ensuring a practical balance between semantic segmentation and real-time localization performance. However, current experiments are conducted under ideal conditions using standardized datasets, without fully reflecting real-world challenges such as multi-sensor interference or unstructured environments. Moreover, the trade-offs between model complexity and accuracy for each lightweight module have not been systematically assessed. Future work will focus on multimodal sensor fusion and adaptive dynamic perception strategies to enhance the robustness and applicability of the proposed system in real-world autonomous inspection scenarios.
Edge Network Data Scheduling Optimization Method Integrating Improved Jaya and Cluster Center Selection Algorithm
YANG Wensheng, PAN Chengsheng
Available online  , doi: 10.11999/JEIT250317
Abstract:
  Objective  The rapid advancement of technologies such as artificial intelligence and the Internet of Things has placed increasing strain on traditional centralized cloud computing architectures, which struggle to meet the communication and computational demands of large-scale data processing. Due to the physical separation between cloud servers and end-users, data transmission typically incurs considerable latency and energy consumption. Therefore, edge computing—by deploying computing and storage resources closer to users, has emerged as a viable paradigm for supporting data-intensive and latency-sensitive applications. However, effectively addressing the challenges of data-intensive services in edge computing environments, such as efficient edge node clustering and resource scheduling, remains a key issue. This study proposes a data scheduling optimization method for edge networks that integrates an improved Jaya algorithm with a cluster center selection strategy. Specifically, for data-intensive services, the method partitions edge nodes into clusters and identifies optimal cluster centers. Data are first aggregated at these centers before being transmitted to the cloud. By leveraging cluster-based aggregation, the method facilitates more efficient data scheduling and improved resource management in edge environments.  Methods  The proposed edge network data scheduling optimization method comprises two core components: a shortest-path selection algorithm based on an improved Jaya algorithm and an optimal cluster center selection algorithm. The scheduling framework accounts for both the shortest communication paths among edge nodes and the availability of network resources. The improved Jaya algorithm incorporates a cosine-based nonlinear decay function and a multi-stage search strategy to dynamically optimize inter-node paths. The nonlinear decay function modulates the variation of random factors across iterations, allowing adaptive adjustment of the algorithm’s exploration capacity. This mechanism helps prevent premature convergence and reduces the likelihood of becoming trapped in local optima during the later optimization stages. To further enhance performance, a multi-stage search strategy divides the optimization process into two phases: an exploration phase during early iterations, which prioritizes global search across the solution space, and an exploitation phase during later iterations, which refines solutions locally. This staged approach improves the trade-off between convergence speed and solution accuracy, increasing the algorithm’s robustness in complex edge network environments. Based on the optimized paths and available bandwidth, a criterion is established for selecting the initial cluster center. Subsequently, a selection scheme for additional cluster centers is formulated by evaluating inter-cluster center distances. Finally, a partitioning method assigns edge nodes to their respective clusters based on the optimized topology.  Results and Discussions  The simulation experiments comprise two parts: performance evaluation of the improved Jaya algorithm (Jaya*) and analysis of the cluster partitioning scheme. To assess convergence speed and optimization accuracy, three benchmark test functions are used to compare Jaya* with four existing algorithms: Simulated Annealing (SA), Genetic Algorithm (GA), Ant Colony Optimization (ACO), and the standard Jaya algorithm. Building on these results, two additional experiments—cluster center selection and cluster partitioning—are conducted to evaluate the feasibility and effectiveness of the proposed optimal cluster center selection algorithm for resource scheduling. A parameter sensitivity analysis using the multi-modal Rastrigin function is performed to investigate the effects of different population sizes and maximum iteration counts on optimization accuracy and stability (Table 2 and Table 3). The optimal configuration is determined to be \begin{document}$ {\text{po}}{{\text{p}}_{{\text{size}}}} = 50 $\end{document} and \begin{document}$ {t_{\max }} = 500 $\end{document}, which achieves a favorable balance between accuracy and computational efficiency. Subsequently, a multi-algorithm comparison experiment is carried out under consistent conditions. The improved Jaya algorithm outperforms the four alternatives in convergence speed and optimization accuracy across three standard functions: Sphere (Fig. 4), Rastrigin (Fig. 5), and Griewank (Fig. 6). The algorithm also demonstrates superior stability. Its convergence trajectory is characterized by a rapid initial decline followed by gradual stabilization in later stages. Based on these findings, the cluster center selection algorithm is applied to tactical edge networks of varying scales—25, 38, and 50 nodes (Fig. 7). The parameter mi is calculated (Fig. 8), and various numbers of cluster centers are set to complete center selection and cluster member assignment (Table 5). Evaluation using the Average Sum of Squared Errors (AvgSSE) under different cluster center counts reveals that the minimum AvgSSE for all three network sizes occurs when the number of cluster centers is 4 (Table 6), indicating that this configuration yields the optimal clustering outcome. Therefore, the proposed method effectively selects cluster centers and derives the optimal clustering configuration (Fig. 9), while maintaining low clustering error and enhancing the efficiency and accuracy of resource scheduling. Finally, in a 38-node edge network scenario with four cluster centers, a multi-algorithm cluster partitioning comparison is conducted (Table 7). The improved Jaya algorithm achieves the best AvgSSE result of 16.22, significantly outperforming the four baseline algorithms. These results demonstrate its superiority in convergence precision and global search capability.  Conclusions  To address data resource scheduling challenges in edge computing environments, this study proposes an edge network data scheduling optimization method that integrates an improved Jaya algorithm with a cluster center selection strategy. The combined approach achieves high clustering accuracy, robustness, and generalization performance. It effectively enhances path planning precision and central node selection, leading to improved data transmission performance and resource utilization in edge networks.
HNN4RP: Heterogeneous Graph Neural Network for Rug Pull Detection on Ethereum
LI Chenchen, JIN Hai, WU Minrui, XIAO Jiang
Available online  , doi: 10.11999/JEIT250160
Abstract:
  Objective  The rapid expansion of Decentralized Finance (DeFi) has been accompanied by a rise in fraudulent activities, with Rug Pull scams posing a significant threat to the security, stability, and credibility of blockchain ecosystems. Due to the permissionless and decentralized nature of DeFi platforms, malicious actors can exploit smart contracts to deceive investors, artificially inflate token prices, and withdraw liquidity abruptly, leaving investors with worthless assets. The anonymous and pseudonymous characteristics of blockchain transactions further hinder fraud detection and regulatory oversight, making Rug Pull scams one of the most challenging forms of financial fraud in the DeFi sector. Although traditional fraud detection approaches, such as rule-based heuristics and statistical analysis, provide basic risk assessment, they often fail to capture the complex transaction patterns, dynamic interactions, and heterogeneous relationships within blockchain networks. To address these limitations, this study proposes Heterogeneous graph Neural Network for Rug Pull detection (HNN4RP), a fraud detection framework based on Heterogeneous Graph Neural Networks (GNNs), specifically designed for the early identification of Rug Pull scams. By leveraging the graph structure of blockchain transactions, HNN4RP models Ethereum transaction data as a heterogeneous graph to capture the relationships among smart contracts, liquidity pools, and trader addresses. The framework integrates advanced representation learning techniques to reveal hidden dependencies, detect anomalous patterns, and differentiate between legitimate and fraudulent behaviors in a scalable, data-driven manner. Furthermore, HNN4RP incorporates temporal dynamics and contextual information to adapt to evolving fraud tactics.  Methods  HNN4RP models Ethereum transactions as a heterogeneous graph, distinguishing between regular nodes and smart contract nodes while preserving both structured and unstructured transactional data. By leveraging a graph-based representation, the model captures complex interactions within the Ethereum network, supporting comprehensive analysis of transactional behavior. This approach explicitly represents key relationships between entities, including token transfers, contract interactions, and liquidity movements, thereby facilitating more accurate fraud detection.The core components of HNN4RP include a Random Walk with Restart (RWR) sampling mechanism, which mitigates biases caused by skewed transaction distributions and enables balanced exploration of node neighborhoods. This technique captures long-range dependencies in the transaction network while preserving the local structural context, which is essential for detecting anomalous behaviors associated with fraud. By dynamically adjusting the probability of revisiting specific nodes, RWR improves the model’s focus on relevant transactional pathways, reducing the effects of data sparsity and imbalance.In addition to RWR-based sampling, HNN4RP incorporates multi-modal feature encoding to capture both structured transactional relationships and unstructured node attributes, such as smart contract code and token metadata. This integration enables the model to leverage a diverse set of informative features, improving its ability to differentiate between legitimate and fraudulent activities. By encoding both numerical and textual representations of transaction data, the framework enhances the robustness of fraud detection against evasive tactics employed by malicious actors.Furthermore, HNN4RP employs a BiLSTM-based temporal fusion mechanism to retain the sequential dependencies within transaction histories, effectively modeling temporal behavioral patterns. This component is essential for capturing the evolution of fraudulent activities over time, as Rug Pull scams often exhibit time-dependent characteristics. By learning the progression of transactional patterns, the model enhances fraud classification accuracy and enables early detection of fraudulent schemes before they escalate. Through this comprehensive approach, HNN4RP provides a scalable, adaptive, and high-precision solution for fraud detection in decentralized financial ecosystems.  Results and Discussions  The performance of the proposed HNN4RP model is evaluated against three categories of baseline methods: attribute-based approaches, random walk-based algorithms, and deep learning models. Across all dataset scales, HNN4RP consistently outperforms these baselines across three predefined evaluation metrics, demonstrating superior fraud detection capability. Notably, on the D2 and D3 datasets, HNN4RP achieves an accuracy of 96% and an F1 score of 0.95 (Table 2). Ablation experiments indicate that both attribute-based features and textual representations contribute significantly to Rug Pull detection performance (Fig. 3). Parameter sensitivity analysis shows that the optimal neighborhood size is 16, and the best embedding dimension is 27 (Figs. 4 and 5). Further experiments incorporating different types of noise interference confirm that HNN4RP exhibits strong robustness to data perturbations, with the F1 score declining by no more than 7% under moderate noise levels (Table 3). Early fraud detection tests demonstrate that HNN4RP effectively identifies fraudulent activities at an early stage, providing reliable warning signals before scams fully develop (Table 4).  Conclusions  This study addresses the emerging challenge of Rug Pull scams in the DeFi and Web3 domains and proposes a deep learning-based detection model, HNN4RP, with significant implications for the stability and security of decentralized financial ecosystems. By formalizing the Rug Pull detection task using heterogeneous graph modeling, this work tackles three major challenges in Ethereum-based fraud detection: biased transaction exploration, incomplete feature representation, and insufficient temporal pattern recognition. These challenges are addressed through the integration of RWR-based random walks, heterogeneous node feature extraction, and BiLSTM-based temporal feature fusion. Experimental results on real Ethereum transaction datasets demonstrate that HNN4RP achieves 96% accuracy and provides reliable early warnings within four hours before fraudulent activities occur, confirming its practical applicability for early fraud detection in decentralized environments.
Efficient Blind Detection Clustering Algorithm for Reconfigurable Intelligent Surface-aided Spatial Modulation Systems
ZHANG Lijuan, SHA Sha, ZHONG Huaqian
Available online  , doi: 10.11999/JEIT250162
Abstract:
  Objective  Reconfigurable Intelligent Surface (RIS)-aided Spatial Modulation (SM) systems (RIS-SM) represent an advanced integration of two promising technologies, offering considerable potential to enhance spectral and energy efficiency in future Sixth-Generation (6G) networks. However, the passive nature of RIS poses a significant challenge to acquiring accurate Channel State Information (CSI), which is essential for conventional signal detection methods. Most traditional algorithms assume perfect CSI, an unrealistic assumption in practical deployments that constrains both performance and scalability. To overcome this limitation, this study proposes an enhanced K-means clustering detection algorithm that reframes the blind detection problem as a clustering task. By exploiting the statistical channel distribution properties inherent to RIS-SM systems, the proposed algorithm achieves efficient blind signal detection without prior CSI and with low computational complexity. These findings contribute to the optimization of RIS-SM system performance and support the development of robust blind detection strategies in complex wireless communication scenarios.  Methods  This study considers a typical RIS-SM architecture in which the RIS comprises multiple passive reflecting elements and the receiver is equipped with multiple antennas. The wireless channel is modeled as an equivalent channel derived from the unique propagation characteristics of the RIS-SM system. By exploiting the statistical distribution of the diagonal elements of this equivalent channel, an improved K-means clustering detection algorithm is proposed, eliminating the requirement for prior CSI. The algorithm’s performance is assessed through extensive simulations. The communication channel is modeled as a Rayleigh fading environment with Additive White Gaussian Noise (AWGN). Phase Shift Keying (PSK) and Quadrature Amplitude Modulation (QAM) schemes are adopted. Bit Error Rate (BER) is used as the primary performance metric to compare the proposed approach against conventional detection methods. Simulations are carried out on the MATLAB platform. Results demonstrate that the proposed algorithm significantly lowers computational complexity while achieving near-optimal detection accuracy, highlighting its suitability for practical deployment in RIS-SM systems.  Results and Discussions  Simulation results indicate that the proposed improved K-means clustering detection algorithm markedly enhances the performance of RIS-SM systems under unknown CSI conditions. As illustrated in Fig. 4 and Fig. 5, the proposed method achieves BER performance that closely approaches that of ideal KMC (KMC(ideal)) and Maximum Likelihood (ML) detectors. It outperforms several benchmark algorithms, including traditional KMC, multi-initialization KMC (KMC(\begin{document}$ P $\end{document})), improved KMC (IKMC), K-means++ (KMC++), and Constrained Clustering-Based Detection (CCBD). While repeated initialization in KMC (\begin{document}$ P $\end{document}) can alleviate poor local minima, it substantially increases computational complexity. IKMC improves centroid initialization but yields limited enhancement in detection accuracy. KMC++ exhibits performance degradation at high Signal-to-Noise Ratio (SNR) due to persistent error floors. CCBD utilizes structural priors to improve clustering but shows performance deterioration when the number of received signals is small. Furthermore, Fig. 6 demonstrates that the proposed algorithm converges more rapidly than other clustering-based methods, particularly under low SNR conditions, highlighting its robustness and efficiency in noisy environments. Table 1 provides additional evidence that the proposed method achieves a favorable balance between detection accuracy and computational cost, supporting its practical value for blind detection in RIS-SM systems.  Conclusions  This study proposes an improved K-means clustering detection algorithm for RIS-SM systems, reformulating the blind detection task as an unsupervised clustering problem. By exploiting the distribution properties of the diagonal elements of the equivalent channel, the algorithm enhances the initialization of clustering centroids, enabling efficient blind signal detection without prior CSI. Simulation results confirm that the proposed method achieves near-optimal detection performance across a wide range of SNR levels and system configurations. These findings offer meaningful contributions to both the theoretical development and practical deployment of RIS-SM systems. The proposed approach is particularly relevant for the design of low-complexity and high-efficiency detection schemes in next-generation wireless networks, including Beyond-5G (B5G) and 6G systems. This work also provides a foundation for future research on blind detection algorithms and their integration with emerging technologies such as RIS, contributing to the advancement of intelligent and adaptive wireless communication systems.
Optimized Design of Non-Transparent Bridge for Heterogeneous Interconnects in Hyper-converged Infrastructure
ZHENG Rui, SHEN Jianliang, LV Ping, DONG Chunlei, SHAO Yu, ZHU Zhengbin
Available online  , doi: 10.11999/JEIT250272
Abstract:
  Objective  The integration of heterogeneous computing resource clusters into modern Hyper-Converged Infrastructure (HCI) systems imposes stricter performance requirements in latency, bandwidth, throughput, and cross-domain transmission stability. Traditional HCI systems primarily rely on the Ethernet TCP/IP protocol, which exhibits inherent limitations, including low bandwidth efficiency, high latency, and limited throughput. Existing PCIe Switch products typically employ Non-Transparent Bridges (NTBs) for conventional dual-system connections or intra-server communication; however, they do not meet the performance demands of heterogeneous cross-domain transmission within HCI environments. To address this limitation, a novel Dual-Mode Non-Transparent Bridge Architecture (D-MNTBA) is proposed to support dual transmission modes. D-MNTBA combines a fast transmission mode via a bypass mechanism with a stable transmission mode derived from the Traditional Data Path Architecture (TDPA), thereby aligning with the data characteristics and cross-domain streaming demands of HCI systems. Hardware-level enhancements in address and ID translation schemes enable D-MNTBA to support more complex mappings while minimizing translation latency. These improvements increase system stability and effectively support the cross-domain transmission of heterogeneous data in HCI systems.  Methods  To overcome the limitations of traditional single-pass architectures and the bypass optimizations of the TDPA, the proposed D-MNTBA incorporates both a fast transmission path and a stable transmission path. This dual-mode design enables the NTB to leverage the data characteristics of HCI systems for telegram-based streaming, thereby reducing dependence on intermediate protocols and data format conversions. The stable transmission mode ensures reliable message delivery, while the fast transmission mode—enhanced through hardware-level optimizations in address and ID translation—supports high-real-time cross-domain communication. This combination improves overall transmission performance by reducing both latency and system overhead. To meet the low-latency demands of the bypass transmission path, the architecture implements hardware-level enhancements to the address and ID conversion modules. The address translation module is expanded with a larger lookup table, allowing for more complex and flexible mapping schemes. This enhancement enables efficient utilization of non-contiguous and fragmented address spaces without compromising performance. Simultaneously, the ID conversion module is optimized through multiple conversion strategies and streamlined logic, significantly reducing the time required for ID translation.  Results and Discussions  Address translation in the proposed D-MNTBA is validated through emulation within a constructed HCI environment. The simulation log for indirect address translation shows no errors or deadlocks, and successful hits are observed on BAR2/3. During dual-host disk access, packet header addresses and payload content remain consistent, with no packet loss detected (Fig. 14), indicating that indirect address translation is accurately executed under D-MNTBA. ID conversion performance is evaluated by comparing the proposed architecture with the TDPA implemented in the PEX8748 chip. The switch based on D-MNTBA exhibits significantly shorter ID conversion times. A maximum reduction of approximately 34.9% is recorded, with an ID conversion time of 71 ns for a 512-byte payload (Fig. 15). These findings suggest that the ID function mapping method adopted in D-MNTBA effectively reduces conversion latency and enhances system performance. Throughput stability is assessed under sustained heavy traffic with payloads ranging from 256 to 2048 bytes. The maximum throughputs of D-MNTBA, the Ethernet card, and PEX8748 are measured at 1.36 GB/s, 0.97 GB/s, and 0.9 GB/s, respectively (Fig. 16). Compared to PEX8748 and the Ethernet architecture, D-MNTBA improves throughput by approximately 51.1% and 40.2%, respectively, and shows the slowest degradation trend, reflecting superior stability in heterogeneous cross-domain transmission. Bandwidth comparison reveals that D-MNTBA outperforms TDPA and the Ethernet card, with bandwidth improvements of approximately 27.1% and 19.0%, respectively (Fig. 17). These results highlight the significant enhancement in cross-domain transmission performance achieved by the proposed architecture in heterogeneous environments.  Conclusions  This study proposes a Dual-Mode D-MNTBA to address the challenges of heterogeneous interconnection in HCI systems. By integrating a fast transmission path enabled by a bypass architecture with the stable transmission path of the TDPA, D-MNTBA accommodates the specific data characteristics of cross-domain transmission in heterogeneous environments and enables efficient message routing. D-MNTBA enhances transmission stability while improving system-wide performance, offering robust support for high-real-time cross-domain transmission in HCI. It also reduces latency and overhead, thereby improving overall transmission efficiency. Compared with existing transmission schemes, D-MNTBA achieves notable gains in performance, making it a suitable solution for the demands of heterogeneous domain interconnects in HCI systems. However, the architectural enhancements, particularly the bypass design and associated optimizations, increase logic resource utilization and power consumption. Future work should focus on refining hardware design, layout, and wiring strategies to reduce logic complexity and resource consumption without compromising performance.
Human-Machine Fusion Intelligent Decision-Making: Concepts, Framework, and Applications
LI Zhe, WANG Ke, WANG Biao, ZHAO Ziqi, LI Yafei, GUO Yibo, HU Yazhou, WANG Hua, LV Pei, XU Mingliang
Available online  , doi: 10.11999/JEIT250260
Abstract:
  Objective  The exponential growth of data volume, advances in computational power, and progress in algorithmic theory have accelerated the development of Artificial Intelligence (AI). Although AI offers unprecedented opportunities across industries, it continues to face limitations such as dependence on large datasets, poor interpretability of learning and decision-making mechanisms, limited robustness, and susceptibility to hallucinations. To overcome these challenges, integrating human cognitive decision-making capabilities and human-like cognitive models into AI systems is essential. This integration gives rise to a new form of intelligence—Human-Machine Fusion Intelligence—which combines physiological and physical characteristics. The core concept is to harness the complementary strengths of humans and machines in information processing and decision-making: humans provide intuitive judgment and contextual understanding, whereas machines are capable of high-speed computation and large-scale data analysis. By establishing a synergistic, collaborative “partnership,” Human-Machine Fusion Intelligent Decision-Making seeks to optimize decision quality through coordinated organic and probabilistic integration. This paradigm holds significant potential to improve decision reliability in mission-critical contexts, such as military operations, medical procedures, and autonomous driving, thus offering both theoretical research value and practical application relevance.  Methods  This review adopts a systematic research approach to examine Human-Machine Fusion Intelligence in decision-making across four core dimensions. First, it presents a theoretical analysis of the fundamental concepts underpinning Human-Machine Fusion Intelligence and highlights its unique advantages in complex decision-making contexts. Second, it proposes a general framework for Human-Machine Fusion Intelligent Decision-Making systems, emphasizing two key components: situational awareness and collaborative decision-making. Based on this framework, decision-making approaches are categorized into three types according to task characteristics and the nature of human-machine interaction: human-led, machine-led, and human-machine collaborative decision-making. Third, the review synthesizes recent practical advancements in representative application domains. Finally, it examines emerging trends in the development of Human-Machine Fusion Intelligent Decision-Making.  Results and Discussions  Unlike prior reviews that focus primarily on specific application domains, this article presents a comprehensive overview of Human-Machine Fusion Intelligence across four key dimensions: conceptual foundations, system framework, practical applications, and current challenges and future prospects. The core contributions of this review are summarized in the following four areas: First, it elucidates the advantages of Human-Machine Fusion Intelligent Decision-Making systems: (1) Improved decision-making accuracy—By combining machines’ strengths in data processing and logical reasoning with human capabilities in handling unstructured problems and ethically complex decisions, the system enables dynamic adjustment through a human-in-the-loop mechanism. (2) Enhanced interpretability of decision outcomes—The decision-making process bridges the cognitive gap between humans and machines, providing a transparent, traceable decision path and clarifying accountability boundaries. (3) Greater system robustness—By integrating machines’ risk monitoring and adaptive capabilities with human experiential judgment in complex or uncertain environments, the system establishes a closed-loop collaboration that balances technological rationality with human cognition. Second, the article highlights that Human-Machine Fusion systems cannot operate independently in safety-critical contexts due to imperfect trust mechanisms and ethical constraints. In response, it proposes a hierarchical architecture comprising two key layers: (1) Situational awareness layer, including three core processes: multimodal data perception, cross-modal information fusion, and situational analysis. (2) Collaborative decision-making layer, which distinguishes three decision-making paradigms based on task characteristics and human-machine interaction mode: (a) Human-led decision-making, suited for tasks with high uncertainty and open-ended conditions, where an enhanced intelligence model with a human-in-the-loop is adopted. (b) Machine-led decision-making, appropriate for tasks with lower uncertainty, emphasizing hybrid intelligence through cognitive model integration in automated workflows. (c) Human-machine collaborative decision-making, applicable when human and machine strengths are complementary, allowing for equal, synergistic cooperation to optimize decision efficiency. Third, the article synthesizes recent technological progress, summarizing representative applications of Human-Machine Fusion Intelligent Decision-Making in mission-critical domains such as the military, healthcare, and autonomous driving. Finally, it identifies six key directions for future development: optimization of multimodal perception, fusion of semantic and feature spaces, construction of deep collaborative feedback loops, dynamic task allocation mechanisms, enhancement of system reliability, and development of ethical guidelines. These directions aim to advance efficient collaboration and sustainable evolution of human-machine intelligence.  Conclusions  Human-Machine Fusion Intelligent Decision-Making offers substantial research value and strong application potential for advancing emerging industries and enabling new intelligent paradigms. Although several exploratory efforts have been made, the field remains in its infancy, lacking a unified and mature theoretical or technological foundation. Key scientific and engineering challenges persist, including the optimization of multimodal perception and data fusion, bridging the semantic gap between human cognition and machine-represented feature spaces, and achieving deep integration of human and machine intelligence. Continued interdisciplinary collaboration will be essential to drive theoretical progress and technological innovation, further unlocking the potential of Human-Machine Fusion Intelligent Decision-Making.
A P-band Dual-polarized Ultra-thin Absorptive-transmissive Electromagnetic Surface Using Frequency Selective Surface
SUN Daifei, YANG Huanhuan, LI Tong, LIAO Jiawei, WU Tianhao, ZOU Jing, YANG Qi, CAO Xiangyu
Available online  , doi: 10.11999/JEIT250309
Abstract:
  Objective  Frequency Selective Surfaces (FSS), as artificial ElectroMagnetic (EM) periodic structures, regulate the transmission and reflection of EM waves. Radomes integrating FSS can protect antennas, preserve the aerodynamic profile of radio-frequency systems, shape spatial scattering field distributions, and suppress backward Radar Cross Section (RCS). However, when illuminated by multiple radars, such radomes often fail to maintain low detectability due to their inability to achieve bistatic low RCS. Recent efforts have focused on developing absorptive structures based on FSS, where active FSS-based absorbers offer adaptive tunability across frequency and time domains. Nonetheless, achieving absorption in the P-band remains challenging due to the inherent limitations of existing dielectric materials. While FSS bandpass properties are frequently employed in radomes and the tunability of active FSS supports the design of reconfigurable absorbers, the two functionalities have largely been pursued independently, resulting in limited multifunctional surface designs. This study proposes a P-band ultra-thin absorber using FSS composed of cascaded unit cells with gradually curved meander lines. By exploiting the distinct equivalent circuit characteristics of absorbing and transmitting FSS structures, an integrated system is developed that enables both EM wave transmission and tunable wideband absorption in the P-band.  Methods  This paper proposes a novel design method for a dual-polarized, ultra-thin absorptive-transmissive EM surface element operating in the P-band, based on the FSS technique. The method uses cascaded elements with a gradient-bending structure to increase the effective current path length and incorporates lumped components to achieve wideband tunable absorption at low frequencies. By analyzing the equivalent circuit characteristics of both absorptive and transmissive FSS-based elements, an integrated absorptive-transmissive structure is developed. The difference in their equivalent circuits effectively suppresses mutual coupling, enabling the relatively independent design of absorptive and transmissive functions. To demonstrate this approach, a dual-polarized ultra-thin EM surface element is designed that simultaneously exhibits high transmittance and tunable wideband absorptivity. The step-by-step design process is presented, and the operating mechanism of the proposed element is thoroughly analyzed.  Results and Discussions  Both simulation and experimental results confirm that the proposed integrated element achieves dual-polarized absorptive performance in the P-band and dual-polarized transmissive performance in the C-band. The element features an ultra-thin profile, requires few lumped components, and exhibits a broad operational bandwidth. Notably, the proposed method combines equivalent circuit modeling with field-based analysis to facilitate the design of multifunctional EM surfaces, thereby streamlining the integration of absorptive and transmissive functionalities within a single structure.  Conclusions  Structural absorbers face a fundamental trade-off between achieving efficient low-frequency absorption and maintaining a low profile or lightweight design, making the development of P-band absorbing surfaces particularly challenging. FSS, commonly used in radomes, also offer potential for tunable absorber design. This study integrates both functionalities to develop a multifunctional EM surface capable of simultaneous wave absorption and transmission, based on an FSS architecture. When irradiated from one direction, the surface achieves absorptivity above 0.9 across 0.34~1.1 GHz. When irradiated from the opposite direction, it exhibits transmittivity exceeding 0.8 over 4.26~4.48 GHz. These two functions operate with relative independence. The proposed structure features a wide operational bandwidth, ultra-thin profile, and minimal reliance on electronic components. The method enables not only effective P-band absorption but also the integrated design of multifunctional EM surfaces. It offers strong scalability and holds significant potential for future applications.
Global Navigation Satellite System Interference Source Localization Algorithm via Single-antenna Synthetic Aperture on the Unmanned Aerial Vehicle Platform
WANG Lu, WANG Xuan, WU Renbiao
Available online  , doi: 10.11999/JEIT250169
Abstract:
  Objective  As Radio Frequency (RF) spectrum congestion intensifies and electromagnetic interference incidents occur more frequently, the reliable operation of the Global Navigation Satellite System (GNSS) is increasingly threatened. To ensure GNSS functionality, interference sources must be accurately monitored and localized. Conventional localization methods are constrained by the limited aperture of physical antenna arrays and often incur high deployment costs. With the gradual opening of low-altitude airspace, small Unmanned Aerial Vehicles (UAVs) are increasingly used as platforms for remote sensing, electronic reconnaissance, and communication tasks, offering a practical alternative for GNSS interference localization. Passive Synthetic Aperture (PSA) techniques have emerged as a solution to overcome the aperture limitations of physical arrays and are well suited for integration with UAV platforms. In this study, PSA technology is combined with a UAV-based system to locate GNSS interference sources. A unified signal model is established to describe various interference types. Based on this model, a slant-range-accurate Range-Doppler (RD) algorithm is derived and further integrated with the Back Projection (BP) algorithm to perform joint localization. This combined RD–BP approach enhances positioning accuracy while reducing computational complexity, demonstrating the feasibility and efficiency of the proposed UAV-based PSA localization framework.  Methods  This approach integrates a RD positioning algorithm based on a precise slant range expression with a BP algorithm. First, an accurate slant range expression is derived using the interference source direction and Doppler frequency. Applying this expression and the principle of stationary phase, the signal is transformed into the RD domain. Multiple matched filters are then constructed in the Doppler domain using the range gate. The outputs of these filters are converted into the range-azimuth domain for peak detection, yielding a coarse estimate of the interference source location. However, this initial estimate is susceptible to errors due to factors such as platform sampling rate and azimuth sampling interval. To refine the estimate, the BP algorithm is employed for fine positioning. A grid is defined around the coarse estimate, and each grid point is used to compute the projected signal. These projected signals are coherently accumulated with the received signals and integrated along the azimuth direction. Peak searching is then performed on the accumulated results to obtain the final, high-precision localization result.  Results and Discussions  Simulation results validate the effectiveness and excellent performance of the proposed RD-BP algorithm. As shown in (Fig. 7), (Fig. 10), and (Fig. 13), the algorithm successfully localizes interference sources after preprocessing for narrowband interference, Linear Frequency Modulated (LFM) interference, and spoofing interference. These results indicate the algorithm’s broad applicability across various interference types. (Fig. 14) illustrates the Root Mean Square Error (RMSE) under two scenarios in which the interference source is either close or distant. In both cases, RMSE decreases as the Signal-to-Noise Ratio (SNR) increases. When the source is nearby, RMSE changes slowly, indicating stable performance. When the source is distant, the algorithm maintains strong performance at higher SNRs. (Fig. 15) shows that under SNR conditions of –10 dB and 15 dB, RMSE increases with target distance. However, the RD-BP algorithm consistently outperforms other methods in accuracy, demonstrating its robustness and adaptability across varying distances. (Fig. 16) further confirms that RMSE stabilizes as the synthetic aperture length increases under both –10 dB and 15 dB SNR conditions. When the aperture length exceeds 100 m, the RD-BP algorithm achieves superior performance relative to other methods and approaches the Cramér–Rao Lower Bound (CRLB), indicating near-optimal estimation accuracy. (Table 1) compares the computational and localization performance of the proposed algorithm with that of the RD algorithm alone. While the RD algorithm is sensitive to platform sampling frequency, flight velocity, and other operational parameters, the RD-BP algorithm mitigates these limitations. It improves localization accuracy while maintaining manageable computational complexity, offering an effective trade-off between precision and efficiency.  Conclusions  This study addresses the problem of locating GNSS interference sources using low-altitude, low-speed UAV platforms. A localization method based on single-antenna synthetic aperture technology is proposed. The study begins by categorizing different types of GNSS interference and their respective preprocessing strategies. An RD algorithm is then derived based on the precise slant range representation, leveraging the relative direction of the interference source with respect to the UAV to achieve coarse localization. To address the accuracy limitations inherent in the RD algorithm due to initial condition constraints, an improved approach is proposed by incorporating the BP algorithm. This enhancement enables refined localization within a narrowed search region. Simulation results confirm that the proposed algorithm can reliably localize common GNSS interference sources under various interference conditions. Compared to conventional methods such as array antenna direction finding, the proposed technique requires only a single antenna mounted on the UAV platform, thereby increasing system flexibility and reducing both cost and design complexity. This demonstrates the feasibility of the method for rapid and efficient localization of GNSS interference sources.
Global–local Co-embedding and Semantic Mask-driven Aging Approach
LIU Yaohui, LIU Jiaxin, SUN Peng, SHEN Zhe, LANG Yubo
Available online  , doi: 10.11999/JEIT250430
Abstract:
  Objective  Facial age progression has become increasingly important in applications such as criminal investigation and digital identity authentication, making it a key research area in computer vision. However, existing mainstream facial age progression networks face two primary limitations. First, they tend to overemphasize the embedding of age-related features, often at the expense of preserving identity-consistent multi-scale attributes. Second, they fail to effectively eliminate interference from non-age-related elements such as hair and glasses, leading to suboptimal performance in complex scenarios. To address these challenges, this study proposes a global–local co-embedding and semantic mask-driven aging method. The global–local co-embedding strategy improves the accuracy of input portrait reconstruction while reducing computational cost during the embedding phase. In parallel, a semantic mask editing mechanism is introduced to remove non-age-related features—such as hair and eyewear—thereby enabling more accurate embedding of age-related characteristics. This dual strategy markedly enhances the model’s capacity to learn and represent age-specific attributes in facial imagery.  Methods  A Global–Local Collaborative Embedding (GLCE) strategy is proposed to achieve high-quality latent space mapping of facial images. Distinct learning objectives are assigned to separate latent subspaces, which enhances the representation of fine-grained facial features while preserving identity-specific information. Therefore, identity consistency is improved, and both training time and computational cost are reduced, increasing the efficiency of feature extraction. To address interference from non-age-related elements, a semantic mask-driven editing mechanism is employed. Semantic segmentation and image inpainting techniques are integrated to accurately remove regions such as hair and glasses that hinder precise age modeling. A differentiable generator, DsGAN, is introduced to align the transferred latent codes with the embedded identity-preserving codes. Through this alignment, the expression of age-related features is enhanced, and identity information is better retained during the age progression process.  Results and Discussions  Experimental results on benchmark datasets, including CCAD and CelebA, demonstrate that GLS-Age outperforms existing methods such as IPCGAN, CUSP, SAM, and LATS in identity confidence assessment. The age distributions of the generated portraits are also more closely aligned with those of the target age groups. Qualitative analysis further shows that, in cases with hair occlusion, GLS-Age produces more realistic wrinkle textures and enables more accurate embedding of age-related features compared with other methods. Simultaneously, it significantly improves the identity consistency of the synthesized facial images.  Conclusions  This study addresses core challenges in facial age progression, including identity preservation, inadequate detail modeling, and interference from non-age-related factors. A novel Global–Local collaborative embedding and Semantic mask-driven Aging method (GLS-Age) is proposed to resolve these limitations. By employing a differentiated latent space learning strategy, the model achieves hierarchical decoupling of structural and textural features. When integrated with semantic-guided portrait editing and a differentiable generator for latent space alignment, GLS-Age markedly enhances both the fidelity of age feature expression and the consistency of identity retention. The method demonstrates superior generalization and synthesis quality across multiple benchmark datasets, effectively reproducing natural wrinkle patterns and age-related facial changes. These results confirm the feasibility and advancement of GLS-Age in facial age synthesis tasks. Furthermore, this study establishes a compact, high-quality dataset focused on Asian facial portraits, supporting further research in image editing and face generation within this demographic. The proposed method not only contributes technical support to practical applications such as cold case resolution and missing person identification in public security but also offers a robust data and modeling framework for advancing human age-based simulation technologies. Future work will focus on enhancing controllable editing within latent spaces, improving anatomical plausibility in skull structure transformations, and strengthening model performance across extreme age groups, including infants and the elderly. These efforts aim to expand the application of facial age progression in areas such as forensic analysis, humanitarian family search, and social security systems.
Low-Complexity Spectrum -Efficiency Optimization Algorithm for Cell-Free Massive MIMO-NOMA Systems
ZHOU Wei, YANG Yu, XIANG Bo, ZHANG Yi, HUANG Hua
Available online  , doi: 10.11999/JEIT250189
Abstract:
  Objective  With the evolution of wireless communication toward ultra-dense networks, optimizing spectrum efficiency in cell-free massive Multiple-Input Multiple-Output (MIMO) Non-Orthogonal Multiple Access (NOMA) systems faces the challenge of balancing algorithmic complexity and performance. Traditional user clustering methods, such as random clustering, exhibit high randomness and easily converge to suboptimal solutions, whereas exhaustive search is computationally prohibitive. Similarly, power allocation schemes with rigid fixed thresholds often fail to accommodate dynamic user demands, resulting in imbalanced resource utilization and reduced fairness. To address these limitations, this study proposes a low-complexity joint optimization algorithm. By collaboratively designing user clustering and power allocation, the algorithm maximizes the system’s sum spectrum efficiency while guaranteeing the quality of service for low-rate users, thereby offering an efficient resource allocation strategy for cell-free massive MIMO-NOMA systems.  Methods  A downlink sum spectrum efficiency maximization model is first constructed and decomposed into two sub-problems: user clustering and power allocation. A clustering algorithm based on cluster head selection and channel difference maximization is then proposed, which reduces the complexity of pairing searches by optimizing the selection of cluster heads. Based on the clustering results, a minimum-rate enhancement constraint mechanism is incorporated. The resulting non-convex power allocation problem is subsequently transformed into a convex optimization form using the Successive Convex Approximation (SCA) method.  Results and Discussions  The relationship between sum spectrum efficiency and the number of users under different clustering algorithms is shown in (Fig. 2). Under perfect Successive Interference Cancellation (SIC) conditions, the proposed algorithm achieves performance comparable to that of the statistical clustering algorithm, whereas the combined clustering algorithm yields the lowest efficiency. Under imperfect SIC, the proposed algorithm maintains the highest spectrum efficiency, while the performance of the statistical algorithm decreases. The sum spectrum efficiency is markedly reduced by imperfect SIC, particularly when the number of users is small. This reduction arises from insufficient elimination of intra-cluster interference, leading to increased residual interference. Notably, the NOMA system supports twice the user capacity of Orthogonal Multiple Access (OMA), confirming its superior spectrum resource utilization. The relationship between sum spectrum efficiency and the number of Access Points (APs) is shown in (Fig. 3). Spectrum efficiency improves substantially as the number of APs increases, owing to enhanced channel hardening and interference suppression. Under perfect SIC, both the proposed algorithm and the statistical clustering algorithm achieve similar performance, exceeding the combined algorithm. The advantage of the proposed algorithm is further demonstrated under imperfect SIC. Increasing the number of deployed APs strengthens anti-interference capability and expands user capacity, verifying the efficiency and robustness of the proposed clustering algorithm under both perfect and imperfect SIC conditions. The effect of intra-cluster user numbers on spectrum efficiency is evaluated in (Fig. 4). Spectrum efficiency increases with the number of APs, largely due to improved coverage and channel estimation accuracy. However, efficiency decreases significantly as intra-cluster user numbers increase, which is attributed to aggravated intra-cluster interference and constraints on power resource allocation. The convergence of the proposed algorithm is demonstrated in (Fig. 5), where the optimal solution is reached within approximately seven iterations. After power allocation optimization, the efficiency gap between imperfect and perfect SIC remains below 0.2 bit/(s·Hz). Compared with Full Power Control (FPC), the Proposed Power Control (PPC) scheme effectively mitigates residual interference and achieves performance close to perfect SIC under practical conditions. The relationship between sum spectrum efficiency and AP numbers with different antenna configurations is presented in (Fig. 6). Efficiency continuously improves with increasing AP numbers and per-AP antenna numbers, owing to stronger channel hardening effects. Across all AP configurations, the PPC scheme demonstrates clear advantages over FPC, with the benefits becoming more pronounced in densely deployed networks. Power allocation optimization further enhances efficiency as antenna numbers increase. Finally, the cumulative distribution of per-user spectrum efficiency under different power control schemes is illustrated in (Fig. 7). The PPC scheme substantially reduces the proportion of inefficient users (0-0.2 bit/(s·Hz)), thereby improving system fairness. By contrast, the FPC scheme performs slightly better in the high-efficiency region, but this advantage comes at the expense of user fairness. In user-intensive scenarios, the PPC scheme effectively balances system stability and fairness by ensuring minimum rates for weak users.  Conclusions  A low-complexity joint optimization algorithm is presented to address the challenge of spectrum efficiency optimization in cell-free massive MIMO-NOMA systems. Through theoretical analysis and simulation, the spectrum efficiency and computational complexity of the system are compared under different user clustering algorithms and power allocation schemes. The results show that the proposed clustering algorithm significantly enhances system performance across various AP deployments and antenna configurations, while reducing computational complexity by 47.5% compared with the statistical clustering algorithm. Furthermore, the joint power allocation scheme demonstrates clear advantages over FPC in terms of spectrum efficiency and user fairness, verifying the effectiveness and practicality of the proposed algorithm.
Low-Light Object Detection via Joint Image Enhancement and Feature Adaptation
QIAO Chengping, JIN Jiakun, ZHANG Junchao, ZHU Zhengliang, CAO Xiangxu
Available online  , doi: 10.11999/JEIT250302
Abstract:
  Objective  Object detection has advanced significantly under normal lighting conditions, supported by numerous high-accuracy, high-speed deep learning algorithms. However, in low-light environments, images exhibit reduced brightness, weak contrast, and severe noise interference, leading to blurred object edges and loss of color information, which substantially degrades detection accuracy. To address this challenge, this study proposes an end-to-end low-light object detection algorithm that balances detection accuracy with real-time performance. Specifically, an end-to-end network is designed to enhance feature quality and improve detection accuracy in real time under low-light conditions.  Methods  To improve object detection performance under low-light conditions while maintaining detection accuracy and real-time processing, this study proposes an end-to-end low-light image object detection method. Detection accuracy is enhanced through joint learning of image enhancement and feature adaptation, with the overall network structure shown in Fig. 1. First, a data augmentation module synthesizes low-light images from normal-light images. The paired normal-light and low-light images are mixed using the MixUp function provided by YOLOv5 to generate the final low-light images. These synthesized images are input into the low-light image enhancement module. In parallel, the matched normal-light images are provided as supervision to train the image enhancement network. Subsequently, both the enhanced low-light images and the corresponding normal-light images are fed into the object detection module. After processing through the YOLOv5 backbone, a matching loss is computed to guide feature adaptation.  Result and Discussions   The experiments are conducted primarily on the Polar3000 and LLVIP datasets. Fig. 3 presents the detection results obtained using YOLOv5 with different image enhancement methods applied to the Polar3000 dataset. Most existing methods tend to misclassify overexposed regions as bright Unmanned Aerial Vehicles (UAVs). In contrast, the proposed method demonstrates accurate object detection in low-light conditions without misidentifying overexposed areas as UAVs (Fig. 3). Furthermore, the detection performance of the proposed method, termed MAET, is compared with that of a standalone YOLOv5 model. Quantitative experiments show that the proposed method outperforms both image-enhancement-first detection pipelines and recent low-light object detection methods across both experimental groups A and B, irrespective of low-light fine-tuning. On the LLVIP dataset, the proposed method achieves a detection accuracy of 91.7% (Table 1), while on the Polar3000 dataset, it achieves 92.3% (Table 2). The model also demonstrates superior generalization performance on the ExDark and DarkFace datasets (Tables 4 and 3). Additionally, compared to the baseline YOLOv5 model, the proposed method increases parameter size by only 2.5% while maintaining real-time detection speed (Table 5).  Conclusions  This study proposes a low-light object detection method based on joint learning of image enhancement and feature adaptation. The approach simultaneously optimizes image enhancement loss, feature matching loss, and object detection loss within an end-to-end framework. It improves image illumination, preserves fine details, and aligns the features of enhanced images with those acquired under normal lighting conditions, enabling high-precision object detection in low-light environments. Comparative experiments on the LLVIP and Polar3000 datasets demonstrate that the proposed method achieves improved detection accuracy while maintaining real-time performance. Furthermore, the method achieves the best generalization results on the ExDark and DarkFace datasets. Future work will explore low-light object detection based on multimodal data fusion of visible and infrared images to further enhance detection performance in extremely dark conditions.
Automated Discovery of Exploitable Instruction Patterns for KASLR Circumvention
LI Zhouyang, QIU Pengfei, QING Yu, WANG Chunlu, WANG Dongsheng
Available online  , doi: 10.11999/JEIT250366
Abstract:
  Objective  Kernel Address Space Layout Randomization (KASLR) remains a core defense against kernel-level exploits; however, its robustness is increasingly undermined by microarchitectural side-channel attacks that exploit specific processor instructions. Existing research has largely concentrated on isolated attack vectors, lacking a systematic evaluation of the entire x86 instruction set. This study addresses this limitation by developing an automated framework to identify and characterize KASLR-bypass instructions comprehensively, assess their attack efficacy across multiple Intel processor generations, and derive defensible instruction patterns to inform the reinforcement of current security mechanisms.  Methods  This study systematically addresses three core challenges in analyzing instruction-level mechanisms for bypassing KASLR. The first challenge is achieving comprehensive coverage of the x86 Instruction Set Architecture (ISA), which includes thousands of historical and modern instructions characterized by variable-length encoding and complex microarchitectural dependencies. To address this, the proposed framework combines static and dynamic analysis. Instruction semantics are extracted statically from Intel Software Developer Manuals and uops.info XML datasets. Dynamic profiling on Intel Core processors is used to verify instruction support across processor generations. Byte-level pattern matching is applied to accurately handle variable-length encodings. The second challenge concerns the generation of attack-compliant machine code that satisfies strict encoding requirements and bypasses compiler-level checks. This is achieved using a template-driven approach, which modifies a CLFLUSH-based attack prototype by replacing inline assembly instructions through pattern substitution. Memory operands are redirected to target addresses preloaded into the EDX register, with boundary values used to ensure operand validity. For nonstandard or undocumented instructions, self-modifying code techniques dynamically inject opcodes at runtime, thereby bypassing compiler restrictions and enabling broader instruction coverage. The third challenge focuses on evaluating attack effectiveness through accurate localization of kernel symbols. To this end, the framework applies a dual-verification strategy. RDTSC instructions are used to timestamp memory probes across 512 predefined address slots. Differential timing analysis identifies latency outliers (i.e., maximum and minimum values), indicating potential KASLR bypasses. Signal handlers suppress exceptions caused by access to privileged or unmapped memory regions, while debug symbol cross-referencing is used to confirm actual kernel address leakage. All generated code undergoes Monte Carlo simulation to reduce false positives and ensure statistical robustness.  Results and Discussions  Experiments are performed on Intel Core i7-11700K, i7-12700K, and i7-13700 processors (Table 1). In the Assembly-Level Instruction Analysis (Fig. 4), 699 assembly instructions are identified as effective KASLR bypass vectors on the i7-11700K. Variations in support for AVX512 instruction set extensions account for differences in the attack surface, with the number of effective instructions decreasing slightly to 542 on the i7-12700K and 547 on the i7-13700, reflecting minor microarchitectural differences. In the Byte-Level Instruction Analysis (Table 2), 39 one-byte, 121 two-byte, and 24 three-byte opcodes are found to bypass KASLR without relying on predefined assembly semantics. These opcodes demonstrate consistent attack efficacy across all evaluated processors, indicating similar behavioral patterns across Intel architectures. Overall, the results—supported by (Fig. 4, Table 2, Table 3)—demonstrate two principal findings: comprehensive coverage of the x86 ISA and cross-generation consistency of effective KASLR bypass instructions. Although the current study focuses on Intel processors, the findings raise open questions regarding the vulnerability of AMD processors that share the same ISA, as well as ARM-based platforms used in Android devices and Apple M series chips. Future work is intended to extend the framework to analyze KASLR bypass vectors on non-Intel architectures. Furthermore, an automated analysis framework is proposed to assess KASLR attack efficacy through differential analysis. To enhance detection across heterogeneous architectures and instruction sets, future efforts will incorporate data preprocessing techniques to improve the scalability and precision.  Conclusions  KASLR remains a critical defense against kernel memory exploitation; however, its resilience is increasingly challenged by instruction-dependent microarchitectural side-channel attacks. This study presents an automated framework that systematically identifies potential KASLR-bypass instructions, quantifies their attack effectiveness across multiple Intel processor generations, and derives actionable defense signatures to address emerging threats. The findings reveal a significantly underestimated attack surface: hundreds of x86 instructions, at both the assembly and byte levels, are capable of leaking sensitive address information. The broader implications of this work are threefold: (1) Defensive Improvement: The experimental results may be directly applied to enhance signature-based detection systems. (2) Hardware–Software Co-Design: The consistent vulnerability observed across Intel microarchitectures highlights the need to redesign timing isolation mechanisms at the hardware level. (3) Methodological Contribution: The proposed dual-analysis framework offers a generalizable approach for evaluating instruction-level attack surfaces, with applicability to other contexts such as cache-based side-channel attacks. Future research will extend this methodology to alternative architectures, including ARM and RISC-V, and explore the integration of machine learning techniques.
A Scalable CPU–FPGA Heterogeneous Cluster for Real-time Power System Simulation
YANG Hangyu, TANG Yongming, LIU Jiyuan, CAO Yang, ZOU Dehu, XU Mingwang, YUAN Xiaodong, HAN Huachun, GU Wei, LI He
Available online  , doi: 10.11999/JEIT250355
Abstract:
  Objective  This study aims to design and implement a scalable CPU–FPGA heterogeneous cluster for real-time simulation of high-frequency power electronic systems. With the increasing adoption of wide-bandgap semiconductor devices such as SiC and GaN, modern power systems exhibit complex switching dynamics that require sub-microsecond timestep resolution. This work focuses on the real-time modeling and simulation of 80 Voltage Source Converters (VSCs), equivalent to 480 switches, representing a typical scenario in renewable-integrated power grids with high switching frequency. Three major technical challenges are addressed: (1) enabling efficient task scheduling across multiple FPGAs to support large-scale parallel computation while maintaining load balance; (2) reducing hardware resource usage through precision-aware hybrid quantization that preserves accuracy with reduced bitwidth; and (3) minimizing CPU–FPGA communication latency via a high-throughput, low-latency data exchange framework to ensure stable synchronization between slow and fast subsystems. This work contributes to the development of a practical and extensible platform for simulating future power systems with complex electronic components.  Methods  To enable real-time simulation with sub-microsecond resolution, the system partitions the power system model into a slow subsystem (AC/DC network) and a fast subsystem (multiple VSCs), following a decoupled computation strategy. A Computation Load-Aware Scheduling (CLAS) strategy is employed to allocate tasks across four Xilinx XCKU060 FPGAs (Fig. 1 and Fig. 2), supporting parallel simulation of up to 80 VSCs. The slow subsystem is executed on the CPU using high-precision floating-point arithmetic with a 50 μs timestep. The fast subsystem is implemented on the FPGAs using fixed-point arithmetic at a 1 μs timestep (Fig. 3 and Fig. 4). A hybrid-precision quantization scheme is adopted: voltage-processing modules use Q(48,30) format to retain numerical precision, whereas current-dominant modules use Q(48,20) to avoid overflow. The FPGA-based Matrix–Vector Multiplication (MVM) is partitioned into two sub-modules (Sub MVM1 and Sub MVM2), leveraging row-level parallelism and pipelined streaming to achieve 400 ns latency per cycle. For communication, a Data Plane Development Kit (DPDK)-based zero-copy framework with lock-free queues is implemented between the CPU and FPGA, reducing latency to 29 μs and enabling reliable synchronization between fast and slow subsystems.  Results and Discussions  The proposed system successfully achieves real-time simulation of a wind farm model comprising 80 VSCs using four Xilinx XCKU060 FPGA boards. Each FPGA supports 20 VSCs operating at a 1 μs timestep, with a computation latency of 400 ns, demonstrating the system’s ability to satisfy stringent real-time constraints. The hybrid-precision quantization strategy yields substantial resource savings relative to a 64-bit fixed-point baseline: LookUp Table (LUT) usage is reduced by 32.0%, Flip-Flops (FFs) by 24.2%, and Digital Signal Processors (DSPs) by 43.8%, while preserving simulation accuracy (Table 1). These optimizations support scalable deployment without loss of fidelity. Communication between the CPU and FPGA is handled by a DPDK-based zero-copy framework with lock-free queues, achieving an end-to-end latency of 29 μs. This ensures robust synchronization between the slow and fast subsystems. Compared with existing FPGA-based designs, the proposed architecture provides a more resource-efficient solution (Table 1), delivering sub-microsecond simulation performance with reduced hardware cost and enabling multi-VSC deployment per FPGA. These findings highlight the platform’s applicability for large-scale industrial power system simulation (Fig. 6).  Conclusions  This study presents a CPU–FPGA heterogeneous cluster designed for real-time simulation of large-scale power systems. The system employs a decoupled, CLAS strategy that enables efficient resource distribution across multiple FPGAs. Real-time requirements are fully met, and the use of hybrid-precision quantization substantially reduces FPGA resource consumption without sacrificing accuracy. The system demonstrates scalability and efficiency by supporting up to 80 VSCs across four FPGA boards. Compared with existing solutions, the proposed architecture achieves the lowest resource utilization while maintaining sub-microsecond resolution, making it a practical platform for industrial-grade power system simulation.
Multi-Dimensional Resource Management Optimization Strategy for Multi-Group Target Tracking
LIN Mushen, CAO Bingxia, YAN Fenggang, MENG Xiangtian, SHA Minghui, LI Zhanguo, JIN Ming
Available online  , doi: 10.11999/JEIT250152
Abstract:
  Objective  Although existing resource management strategies can improve the performance of Multiple Target Tracking (MTT), they generally assume that targets are well separated. With the development of swarm intelligence, multiple maneuvering targets can adopt group merging and splitting tactics to enhance task efficiency. This often results in the presence of several closely spaced targets within a single tracking beam, making them difficult to distinguish. In such cases, networked radars must coordinate the allocation of signal bandwidth and pulse width to increase both the resolution of adjacent targets and overall MTT performance. However, previous studies have not accounted for the effect of signal resolution capability on MTT performance. When adjacent targets emerge in the surveillance region, resource allocation schemes derived from existing methods become insufficient. To address this limitation, this study proposes a joint Beam Selection and Multi-dimensional Resource Management (BSMRM) strategy. By jointly optimizing the transmit beam, power, signal bandwidth, and pulse width, the proposed approach markedly improves MTT performance for multiple groups of clustered maneuvering targets and enhances the resolution of adjacent targets within the surveillance area.  Methods  This study proposes a BSMRM strategy for netted radar systems to improve target resolution capability and MTT accuracy. First, the Bayesian Cramér-Rao Lower Bound (BCRLB) is derived under the Probabilistic Data Association (PDA) fusion rule and the Interacting Multiple Model-Extended Kalman Filter (IMM-EKF) framework. Resolution performance is quantified using the normalized magnitude of the ambiguity function for both observed and unobserved targets located within the beam. A utility function is then formulated using the logarithmic barrier method to quantify global performance. Multi-dimensional resource management is modeled as a constrained optimization problem, with the objective of maximizing global performance under system resource constraints. A three-stage solver is developed to address the optimization problem. Simulation results show that the proposed BSMRM strategy achieves comparable target resolution and MTT accuracy to state-of-the-art algorithms, while demonstrating superior efficiency in spectrum usage and power allocation.  Results and Discussions  As shown in (Table 1) and (Fig. 1), this study evaluates the effectiveness and robustness of the proposed BSMRM strategy in a complex scenario where four groups of swarm targets maneuver within the surveillance region of three radar nodes. The performance of the BSMRM strategy is assessed through four simulation experiments. Experiment 1 presents the allocation of transmission resources, including beam selection and power distribution. The results are shown in (Fig. 2), where color-coded regions represent normalized transmit power levels determined by the BSMRM strategy. Experiment 2 reports the allocation of waveform resources—specifically, signal bandwidth and pulse width. In this experiment, resolution thresholds corresponding to various target states are first established. The BSMRM strategy is then used to adjust the signal bandwidth and pulse width of each transmit beam to enable effective discrimination of swarm targets. Experiment 3 analyzes tracking performance through numerical simulations, comparing the proposed strategy with benchmark algorithms. A sequence of 60 tracking frames with a time interval of T = 2.5 s is simulated, with 300 Monte Carlo runs. Root Mean Square Error (RMSE) is used as the evaluation metric. (Fig. 4) displays the BCRLB and RMSE curves for each target group under the BSMRM strategy. The RMSE converges toward the corresponding BCRLBs, indicating that the proposed method meets tracking accuracy requirements across all target groups. When sufficient transmit power is available, the BSMRM strategy drives the accuracy to the desired thresholds, thereby maximizing utility. (Fig. 5) compares the normalized average power consumption of three strategies. (Figs. 6 and 7) show resolution performance curves for the BSMRM and JBPAWLS strategies, respectively. (Fig. 8) compares spectrum occupancy across two operational phases to assess spectral efficiency under resolution constraints. Experiment 4 evaluates the computational complexity of the BSMRM strategy. Results show that, with parallel computing, the strategy satisfies the real-time requirements of resource scheduling.  Conclusions  This study proposes a BSMRM strategy for MTT in a C-MIMO radar network. The strategy jointly optimizes beam selection, transmit power, signal bandwidth, and pulse length to improve both target resolvability and tracking accuracy. An efficient three-stage solver is developed to address the resulting optimization problem. Simulation results show that the proposed strategy achieves high-resolution MTT while reducing spectral occupancy. Under sufficient power conditions, tracking accuracy is comparable to that of state-of-the-art algorithms. In the same simulation environment, the average CPU runtime per frame is 2.8617 s, which is reduced to 0.3573 s with GPU-based parallel processing—meeting the real-time requirements for resource allocation. Moreover, the BSMRM strategy demonstrates flexibility and robustness under limited power constraints.
Network Protocol Fuzzing: Method Classification and Research Progress
LAI Guoqing, ZHU Yuefei, CHEN Di, LU Bin, LIU Long, ZHANG Zihao
Available online  , doi: 10.11999/JEIT250188
Abstract:
  Significance   Network security vulnerabilities arising from flaws in network protocol implementations cause substantial losses and adverse societal effects, exemplified by the Heartbleed vulnerability in OpenSSL (CVE-2014-0160). This flaw allows attackers to extract encryption keys from public servers, enabling decryption of traffic or unauthorized authentication on less secure systems. Approximately 500,000 internet servers are affected when the vulnerability is publicly disclosed. Against this backdrop of escalating network security risks, ensuring the security of network protocol software becomes a critical research area. Fuzzing, a software testing technique, emerges as one of the most widely used approaches for identifying vulnerabilities in network protocol implementations due to its ease of deployment and high efficiency. The core concept of fuzzing is to improve software security by generating and sending crafted test cases to the target program. Despite significant progress in this field, Network Protocol Fuzzing (NPF) still faces technical challenges. Currently, no systematic and up-to-date review of NPF research exists, limiting researchers’ ability to grasp recent advances. This paper conducts a comprehensive review of NPF techniques, aiming to provide researchers in the field of network protocol security with valuable references for tool selection and optimization.  Progress   Since the proposal of AFLNET in 2020, considerable progress occurs in the field of NPF, addressing key challenges such as protocol packet construction, protocol state awareness, and network communication efficiency optimization. Specifically, for protocol packet construction, researchers propose machine learning-based packet generation methods that integrate both generation and mutation strategies. Mutation operators and guidance techniques are designed to target specific protocol characteristics. In terms of protocol state awareness, state tracking capabilities are enhanced through state variable capture and state model construction. Furthermore, state fuzzing has been widely employed to detect state machine bugs. For protocol stack efficiency optimization, researchers improve testing efficiency by refining communication mechanisms and applying snapshot techniques.  Conclusions  To comprehensively summarize the research progress in NPF, this paper first clarifies the unique characteristics of network protocol software compared with other fuzzing targets. These characteristics include strict protocol message formats, asynchronous network interactions, and complex protocol state maintenance. A problem-oriented classification framework for NPF is proposed, structured around three core challenges: protocol packet construction, protocol state awareness, and protocol stack efficiency optimization. Based on this framework, research advancements in NPF over the past eight years are systematically reviewed, with a technical analysis of the capabilities, and limitations of existing approaches. This review highlights several key challenges in the field. For input construction, major limitations include weak validity of generated inputs, input space explosion, and restrictions imposed by encryption and authentication mechanisms. In terms of state awareness, the field faces insufficient protocol state space exploration and low levels of test intelligence and automation. Regarding performance optimization, technical challenges include slow network communication speed, limited scalability across different protocol implementations, and inadequate adaptability to complex network environments. This study provides both theoretical foundations and practical references to guide future research and technological development in the NPF domain.  Prospects   Future research in NPF can integrate emerging technologies such as Artificial Intelligence (AI) to enhance the intelligence and automation of testing processes. For example, combining AI methods with program analysis techniques may enable a deeper understanding of protocol behavior and more efficient generation of test packets. Developing state representations tailored to different protocol characteristics, implementing real-time protocol state mapping, and applying state prediction and reasoning based on LLM can further improve the efficiency and applicability of NPF tools. In addition, introducing technologies such as parallelization, distributed computing, modular test architectures, and integrated network simulation with virtualization can significantly enhance testing scalability and adaptability. Through the integration of emerging technologies and interdisciplinary research, NPF is expected to play an increasingly critical role in network protocol security, providing robust technical support for building secure and reliable network environments.
Security and Reliability-Optimal Offloading for Mobile Edge Computing in Low-latency Industrial IoT
HU Yulin, YU Xinlan, GAO Wei, ZHU Yao
Available online  , doi: 10.11999/JEIT250262
Abstract:
  Objective  With the advancement of Industry 4.0, the increasing demand for mission-critical and latency-sensitive applications is driving the evolution of Industrial Internet of Things (IIoT) networks. Mobile Edge Computing (MEC) has emerged as a key solution to alleviate the computational constraints of User Equipment (UE), thus enhancing IIoT system performance. However, ensuring high reliability and security remains a critical challenge in multi-access MEC networks. To support Ultra-Reliable and Low-Latency Communication (URLLC), short packet transmission is employed, introducing Finite BlockLength (FBL) effects that impacts communication performance. Additionally, queuing delays at the servers present a significant bottleneck in meeting latency requirement. While existing studies predominantly focus on secure transmission under the Infinite BlockLength (IBL) assumption, comprehensive investigations into Physical Layer Security (PLS) and reliability in the FBL regime remain scarce. To address this gap, this paper aims to jointly optimize the security and reliability of MEC networks. Specifically, an analytical expression for end-to-end Leakage Failure Probability (LFP) is derived to assess the security and reliability performance. Furthermore, an efficient two-phase time length allocation algorithm is proposed to enhance overall system performance.  Methods  This study examines the uplink URLLC scenario in a multi-access MEC network. First, leveraging FBL theory, the communication phase within frame structure is modeled, and the decoding error probabilities of both the server and the eavesdropper are characterized. Subsequently, Extreme Value Theory (EVT) is employed to model the delay violation probability in the computation phase. Based on these formulations, the end-to-end LFP is derived to comprehensively assess the system’s PLS and reliability. Given the stringent end-to-end latency constraints in URLLC, a fundamental trade-off between security and reliability arises. To investigate this trade-off, an optimization problem is formulated to minimize LFP while satisfying the latency constraint, with the time length of the communication and computation phases as the optimization variables. However, due to the presence of multiplicative non-convex Q-functions in the objective function, solving this optimization problem poses significant challenges. To overcome this, a local point expansion technique is employed to approximate the non-convex terms as summations, facilitating problem reformulation. Additionally, auxiliary variables are introduced to enable convex relaxation of the Q-functions, transforming the problem into a convex optimization framework. The reformulated problem is then efficiently solved using the Successive Convex Approximation (SCA) algorithm. In response to the issue of performance degradation in algorithm convergence when the task scale increases, an Accelerated-SCA (A-SCA) algorithm is introduced. By leveraging the problem structure, a closed-form expression for the time allocation strategy corresponding to the system's optimal reliability is derived, which improves the initial point selection mechanism. Finally, simulation results validate the accuracy, convergence, and adaptability of the proposed algorithm, while also revealing the trade-off between security and reliability.  Results and Discussions  The proposed FBL coding-based time allocation scheme is evaluated through numerical simulations to validate its effectiveness. A comparative analysis with the IBL coding-based scheme is conducted, demonstrating the performance gains achieved through the joint optimization of communication and computation phases, while highlighting the limitations of IBL coding. Furthermore, the quality of the sub-optimal solution obtained via the SCA algorithm is assessed by benchmarking against the exhaustive search method. Additionally, both constrained and unconstrained resource scenarios are considered to explore the trade-offs between security and reliability, as well as between time resource and LFP. First, (Fig. 2) analyzes the convergence behavior of the proposed SCA-based time allocation scheme, demonstrating that the sub-optimal solution obtained closely approximates the global optimum. Furthermore, by optimizing the initial point selection with the A-SCA algorithm, the convergence speed and scalability are significantly enhanced. (Fig. 3) evaluates the variation of LFP with the eavesdropper’s channel gain under different task data sizes, revealing the fundamental trade-off between security and reliability, along with the associated performance compensation mechanism. (Fig. 4) illustrates the LFP trend as a function of the server’s channel gain, comparing the proposed scheme with IBL coding-based scheme. The results indicate that even in scenarios with low secure channel capacity, the proposed scheme achieves a 21.4% performance improvement, thereby validating its superiority in joint optimization across both phases. (Fig. 5) investigates the impact of task workload on LFP, while (Fig. 6) explores an unconstrained resource scenario, analyzing the influence of transmit power on LFP and the corresponding service time requirements.  Conclusions  This study explores the joint optimization of security and reliability in multi-access MEC networks supporting URLLC. In the FBL regime, the decoding error probability is modeled in the communication phase, while the delay violation probability is characterized using EVT in the computation phase. Based on these models, the LFP is derived as a key metric to quantify the system’s end-to-end PLS and reliability. Given the stringent latency constraints, a two-phase time length allocation optimization problem is formulated to minimize LFP, balancing security and reliability. To address the problem’s non-convexity, this paper first decouples the coupled variables through local linearization and reconstructs the problem. Next, convex upper and lower bounds of the complex Q-function are derived, enabling convex relaxation of the objective function. Subsequently, the SCA algorithm is applied iteratively to obtain a sub-optimal solution for the original problem. To tackle the issue of deteriorating convergence performance and increased complexity as the task scale grows, an A-SCA algorithm is proposed, which optimizes the initial point selection to address this challenge. Finally, simulation results validate the convergence, accuracy, and scalability of the proposed algorithm, demonstrating its performance advantages through comparison with existing approaches. Additionally, the impact of key system parameters on performance is systematically analyzed. The proposed LFP metric exhibits significant extensibility, being applicable to collaborative optimization of physical layer security and network layer security in URLLC systems, reliable task offloading in MEC, and throughput optimization in MIMO systems. Furthermore, the solution framework for LFP optimization problem, including problem decoupling and reconstruction, introduction of auxiliary variables for Q-function convex relaxation, iterative solving via SCA, and optimized initial point selection, provides valuable insights for solving multi-variable non-convex optimization problems in the FBL regime.
Joint Beamforming and Antenna Position Optimization in Movable Antenna Empowered ISAC Systems
LI Zhendong, BA Jianle, SU Zhou, ZHAO Weichun, CHEN Wen, ZHU Zhengyu
Available online  , doi: 10.11999/JEIT250146
Abstract:
  Objective  With the rapid advancement of information technology, mobile communication is transitioning from the fifth generation (5G) to the sixth generation (6G), with a central goal of evolving from the Internet of Things to the Internet of Intelligence. Integrated Sensing and Communication (ISAC) has emerged as a key technology for next-generation wireless systems. By jointly designing sensing and communication functionalities, ISAC substantially improves spectral efficiency and overall system performance, and is regarded as a core technology for future intelligent networks. However, conventional fixed antenna arrays are limited by insufficient spatial degrees of freedom, making them inadequate for meeting the stringent sensing and communication requirements in dynamic and complex environments. To address this challenge, Movable Antenna (MA) is proposed as a novel architecture. MA systems enable antenna elements to move within designated transmit and receive regions, allowing real-time adjustment of their positions according to instantaneous channel states and system demands. This study proposes an MA-empowered ISAC framework that dynamically reconfigures antenna positions to exploit the spatial characteristics of wireless channels, with the objective of minimizing transmit power. The proposed design contributes new insights into energy-efficient ISAC system development and offers theoretical and practical relevance for future wireless communication networks.  Methods  This study formulates a joint optimization problem that considers the discrete positions of MA elements, beamforming vectors, and sensing signal covariance matrices. To address the inherent coupling among optimization variables and the presence of binary discrete variables, a Discrete Binary Particle Swarm Optimization (BPSO) framework is adopted. The algorithm iteratively determines the discrete antenna positions through a fitness-based search. Based on the obtained positions, Semi-Definite Relaxation (SDR) and Successive Convex Approximation (SCA) techniques are applied to manage non-convex constraints and solve for the beamforming vectors and sensing covariance matrices. This approach yields suboptimal yet effective solutions in complex non-convex optimization scenarios, thereby enhancing system performance. In terms of system modeling, a Dual-Function Radar and Communication Base Station (DFRC BS) equipped with MA is considered. The DFRC BS transmits downlink ISAC signals, communicates with multiple single-antenna downlink users, and senses a radar target. Antenna elements are restricted to a discrete set of candidate positions, facilitating practical system deployment. The radar sensing channel is modeled as a Line-of-Sight (LoS) channel, whereas the communication channels follow a field-response-based model. The DFRC BS transmits narrowband ISAC signals through beamforming, which are used concurrently for radar target detection and downlink communication.  Results and Discussions  Simulation results demonstrate that the proposed MA-empowered ISAC system achieves superior performance in transmission power optimization compared with conventional fixed antenna array systems. Under the constraint of sensing Signal-to-Interference-plus-Noise Ratio (SINR), transmit power increases with more stringent sensing requirements. However, the MA-empowered ISAC system substantially lowers the required transmit power, with a maximum observed reduction of 101.1 W. Under the communication SINR constraint and fixed sensing requirements, transmit power also rises as communication demands grow. In this setting, the MA-empowered ISAC system again shows a clear advantage, reducing transmit power by up to 134.6 W compared with traditional systems. Furthermore, as the number of downlink users increases—while maintaining consistent sensing and communication requirements—transmit power increases accordingly. Even under these conditions, the MA-empowered system continues to outperform its fixed-array counterpart. Beamforming pattern simulations further confirm that the BPSO-based optimization framework achieves multi-beam alignment and provides a degree of interference suppression, highlighting its effectiveness in managing dual-function transmission within MA-enabled ISAC systems. Overall, the introduction of MA significantly enhances ISAC system performance in terms of energy efficiency and spatial adaptability.  Conclusions  This paper proposes a joint optimization framework based on Discrete BPSO for MA-empowered ISAC systems, which effectively reduces transmit power while maintaining sensing and communication performance. The optimization process begins by iteratively solving a fitness function to determine the discrete positions of the MA elements. SDR and SCA techniques are employed during this process to address non-convex constraints. Once the MA element positions are established, the associated beamforming vectors and sensing signal covariance matrices are computed or further optimized. By dynamically adjusting antenna positions, the system gains additional spatial degrees of freedom, enabling more efficient utilization of the spatial characteristics of wireless channels. This work provides a new approach for energy-efficient design in ISAC systems and offers meaningful guidance for the development of next-generation wireless networks. Future research will continue to explore advanced optimization algorithms for MA positioning and system performance enhancement, aiming to meet sensing and communication requirements in highly dynamic environments. Additionally, the applicability of MA in other wireless scenarios—such as vehicular networks and aerial communication platforms—will be studied to facilitate broader adoption of MA-based technologies.
Available online  , doi: 10.11999/JEIT250112
Abstract:
A Test Vector CODEC Scheme Based on BRAM-Segmented Synchronous Table Lookup
YI Maoxiang, ZHANG Jiatong, LU Yingchun, LIANG Huaguo, MA Lixiang
Available online  , doi: 10.11999/JEIT250053
Abstract:
  Objective  Logic testing using Automatic Test Equipment (ATE) is a critical step in integrated circuit (IC) manufacturing test to ensure chip quality. Enhancing logic test efficiency is essential to reducing digital IC testing costs. During testing, IC test data are typically stored in the main memory of the ATE user board and sequentially read to generate channel test waveforms. The time required to read test data directly affects test efficiency. Traditional Test Data Compression (TDC) approaches, which often require preprocessing such as X-bit filling, are suited only for scan testing and thus do not meet broader test engineering needs. Meanwhile, advances in Field-Programmable Gate Array (FPGA) technology have enabled the customization of high-speed Block RAM (BRAM) resources. This study proposes a test vector coding scheme based on component statistics, in which the Device Under Test (DUT) test vectors are encoded and corresponding component coding tables are generated and stored in the FPGA BRAM. A table lookup circuit is implemented to achieve synchronous, parallel output of all test vector components, effectively reducing the external data read time and improving logic test efficiency.  Methods  Each bit symbol in an IC test vector comprises four components: drive (DC), measurement (MC), high impedance (ZC), and residual value (RV). The proposed scheme performs statistical encoding of each component across all bit symbols in the DUT’s test vectors and generates shared DC, MC, and ZC coding tables. The encoding process includes: (1) scanning and extracting each vector from the DUT test project files; (2) determining the bit component values and residual values for all channels; (3) for each component, compiling and deduplicating all generated codes, reassigning deleted code references to reserved codes to form the final coding tables; and (4) determining the combined component addresses and residual values. Using a Xilinx Kintex-7 FPGA development board and the Vivado tool, three BRAM modules are configured, and a BRAM table lookup control circuit is designed (Fig. 4). Prior to testing, the component coding tables are downloaded to the FPGA BRAM, and the combined address and residual values of the three component codes for each test vector are stored in off-chip SDRAM. During operation, the lookup circuit uses the combined address to synchronously and in parallel output the three components, which—together with the residual value—drive the waveform generator to produce the channel test waveform.  Results and Discussions  The functionality of the BRAM-segmented synchronous table lookup circuit is verified through simulation. Three BRAM modules with 64-bit width and customized segment address depth are configured. The COE files of the component encoding tables are downloaded to the target BRAMs via a UART interface, using address generation control logic. The corresponding addresses are then applied to the lookup circuit. A complete simulation is conducted by integrating the segmented lookup module, data strobe module, address allocation module, and data transmission module, enabling validation of the BRAM data download, segmented table lookup, and I/O processes within the FPGA (Fig. 6Fig. 8). Results confirm that the synchronized parallel output from the lookup circuit matches the three component codes of the predefined test vectors (Fig. 9Fig. 13). The SDRAM read time is also analyzed. Under the same configuration parameters, the proposed encoding scheme reduces the read time of each test vector by 66.7% compared with a direct encoding storage scheme (Table 3), indicating a significant improvement in logic test efficiency. A qualitative comparison with traditional TDC schemes—including dictionary coding, Frequency-Directed Run-length (FDR) coding and run-length coding—is presented in Table 4. The results indicate that the proposed scheme, which utilizes high-speed BRAM embedded in modern FPGAs, supports non-scan parallel logic testing with high decoding speed and low overhead, while fully satisfying the original test project requirements.  Conclusions  A test vector encoding and decoding scheme based on component statistics and BRAM-segmented synchronous table lookup is proposed and implemented. The segmented lookup circuit is designed, and its functional correctness is verified through simulation. Compared with direct encoding, the proposed scheme achieves a 66.7% reduction in logic test time. In contrast to traditional TDC approaches, it offers lower hardware overhead by leveraging embedded high-speed BRAM. The scheme supports ATE-based parallel non-scan logic testing and meets the original engineering design goals, providing a practical foundation for optimizing the logic test function module of the ATE user board.
Review of Research Progress on TSV Technology in 3D IC Packaging
ZHANG Qianfan, HE Xi, TIAN Yu, FENG Guangyin
Available online  , doi: 10.11999/JEIT250377
Abstract:
  Significance   Three-Dimensional Integrated Circuits (3D ICs) have emerged as a key research direction in the post-Moore era due to their advantages in low latency and high integration density. As electronic devices demand higher performance and smaller form factors, 3D ICs offer a compelling solution by vertically stacking multiple chip layers to achieve enhanced integration. A core enabler of 3D IC technology is Through-Silicon Via (TSV) technology, which facilitates high-density vertical interconnects across layers. TSVs have contributed significantly to performance improvements in 3D ICs but also pose challenges in thermal management, power integrity, and signal integrity, all of which can affect device reliability and operational stability. Addressing these challenges is essential for the continued advancement of 3D IC systems. This review outlines recent research on TSV technology, with an emphasis on thermal, electrical, and signal integrity issues, as well as current strategies for mitigating these limitations.  Progress   This review systematically summarizes the progress in TSV technology, focusing on the following areas: Thermal Management: Thermal dissipation is a critical concern in 3D ICs due to elevated power densities resulting from multilayer stacking. While TSVs improve interconnect performance, they can also introduce vertical heat flow paths that lead to localized overheating and reduced reliability. To manage this, various thermal modeling approaches—such as Finite Element Analysis (FEA) and thermal stacking simulations—have been developed to predict temperature distributions and optimize thermal performance. These models inform the layout of TSVs and guide the incorporation of Thermal TSVs (TTSVs) to enhance heat dissipation. Researchers have also explored the use of high-thermal-conductivity materials, such as carbon nanotubes and graphene, to improve thermal pathways. Optimizing TSV density and employing multi-layer thermal redistribution techniques have further advanced thermal management, contributing to better device performance and longer operational lifetimes. Power Integrity: Power integrity is a major design constraint in 3D ICs, given the complex power delivery networks required in stacked architectures. TSVs, acting as vertical power conduits, can introduce issues such as voltage drops, electromigration, and power noise. Several approaches have been proposed to address these issues. Layout optimization—particularly through uniform TSV distribution and the integration of Backside Power Delivery Networks (BPDNs)—helps reduce power delivery path lengths and mitigate voltage loss. Dynamic Voltage And Frequency Scaling (DVFS) is also employed to adapt power usage under varying workloads, particularly in high-performance computing environments. Additional methods include the use of decoupling capacitors (DECAPs) and Fully Integrated Voltage Regulators (FIVRs), which help suppress power noise and maintain stability across multiple voltage domains. Signal Integrity: TSV-based interconnects must maintain signal integrity at increasingly high frequencies, but parasitic inductance and capacitance inherent to TSVs can degrade signal quality through reflection, crosstalk, and delay mismatch. These effects become especially pronounced in high-density, high-speed interconnect architectures. To address this, electromagnetic shielding—using grounded TSVs and metallic isolation structures—has been shown to reduce crosstalk and enhance signal fidelity. The use of low-dielectric constant (low-k) materials further minimizes parasitic capacitance and improves signal propagation speed. Differential TSV designs and advanced interconnect architectures have also been proposed to reduce interference and enhance signal integrity. These improvements are essential for achieving reliable high-speed data transmission in storage and processing applications.  Conclusions  While TSV technology has advanced substantially in addressing the thermal, power, and signal integrity challenges of 3D ICs, several limitations persist. These include scalability constraints, power delivery reliability under high-density integration, and diminished signal transmission quality at high frequencies. These challenges highlight the need for continued innovation in TSV design and integration to meet the demands of next-generation 3D IC systems. Several promising research directions are emerging. First, there is a growing need for higher-precision multiphysics coupling models. As 3D ICs progress toward large-scale heterogeneous integration, high-speed data communication, and extreme energy efficiency, more accurate modeling of the thermal, electrical, and signal interactions associated with TSVs is required. This calls for enhanced integration of multiphysics simulations into the Electronic Design Automation (EDA) workflow to enable co-simulation across electrical, thermal, and signal domains. Second, co-optimization of BPDNs and nano-TSVs (nTSVs) is becoming increasingly important. As chip dimensions decrease and stacking complexity grows, traditional front-side power delivery approaches no longer meet the required power densities. Improved BPDN strategies, in conjunction with nTSV integration, will support higher stacking capability and improved energy efficiency. Third, the exploration of new materials and TSV array structures offers additional opportunities. Carbon-based nanomaterials, used as TSV fillers or liners, can alleviate thermal expansion mismatch and improve resistance to electromigration. Incorporating air gaps or low-k dielectrics as insulating liners can reduce parasitic capacitance and enhance high-speed signal performance. Meanwhile, novel TSV array architectures can increase interconnect density and improve redundancy and fault tolerance. Finally, the adoption of AI-driven TSV optimization holds considerable promise. TSV layout design currently depends heavily on manual heuristics. The application of artificial intelligence to automate TSV placement and power network distribution can significantly reduce design time and accelerate the transition toward more intelligent 3D integration design paradigms.
Study on Time Slot Allocation and Monitoring Performance of UAT2 Data Link Status Bitmap
TANG Xinmin, TANG Shengjia, WEN Jingyu, GU Junwei
Available online  , doi: 10.11999/JEIT250251
Abstract:
  Objective  With the advancement of Urban Air Mobility (UAM), the rapid growth in aircraft numbers under such scenarios requires improved time slot allocation algorithms for the Universal Access Transceiver 2 (UAT2) data link to enhance surveillance capacity. This study analyzes the original time slot allocation algorithm for UAT2, identifying limitations related to downlink message time slot management. An improved allocation algorithm based on a state bitmap with a random drift mechanism is proposed considering message characteristics. Additionally, three transmission schemes to expand the number of time slots are proposed, and a surveillance capacity calculation model is established. The performance of each transmission scheme under different slot allocation strategies is simulated and evaluated. The research addresses the challenge of insufficient surveillance capacity in UAT2 data link under high-density UAM scenarios and provides an optimized approach for time slot allocation and surveillance performance improvement.  Methods  The study begins with an analysis of the original UAT2 time slot allocation algorithm, which is limited in high-density aircraft environments due to its pseudo-random mechanism based on geographic coordinates. The proposed algorithm introduces a state bitmap, implemented as a bit vector table where each bit indicates whether the corresponding time slot is occupied. When an aircraft selects a time slot, it first generates a pseudo-random number using the algorithm specified in DO-282C MOPS. The state bitmap is then checked: if the corresponding bit is 0, the time slot is selected; if 1, a piecewise random drift mechanism adjusts the slot selection, as shown in Formula (17). The drift mechanism segments the time slots to distribute selections more evenly and reduce collision probability. The algorithm also applies a slot retention period T, allowing an aircraft to occupy the same time slot for T consecutive UAT frames before reselecting, as defined by Formulas (19)–(21). To further expand time slot availability, three transmission schemes are proposed: Multi-ary Continuous Phase Frequency Shift Keying (MCPFSK) with modulation orders M = 4 and 8, constant modulation index with increased symbol rate, and constant carrier frequency difference with reduced modulation index. The available number of time slots for each scheme is calculated using Formula (26). A surveillance capacity model incorporating bit error rate and collision probability is established, as expressed by Formulas (27)–(30).  Results and Discussions  Simulation results demonstrate that the improved algorithm substantially outperforms both the original algorithm and the fixed time-window dynamic slot allocation algorithm. Under the original transmission scheme, the improved algorithm reduces slot collision probability by 16.78% and increases slot utilization by 16.12% compared to the original algorithm (Fig. 6, Fig. 7). Relative to the algorithm described in [Ref. 19], the collision probability decreases by 10.80%, and slot utilization increases by 10.48%. For the expanded time slot schemes, when maximum surveillance capacity is reached, the improved algorithm reduces collision probability by 16.14% and increases slot utilization by 16.13% relative to the original algorithm (Table 3). Among these schemes, the 8CPFSK expansion achieves the highest surveillance capacity of 3913, with a slot utilization rate of 79.37% (Fig. 13). Real-time performance testing indicates that even in high-density scenarios, the improved algorithm maintains scheduling times within 120 ms, meeting the real-time operational requirements of UAT2 (Fig. 11). Bit Error Rate (BER) simulations reveal that the MCPFSK scheme provides superior anti-interference performance, whereas the constant carrier frequency difference scheme exhibits the highest BER (Fig. 12).  Conclusions  A slot allocation algorithm for the UAT2 data link based on state bitmaps and a random drift mechanism is proposed in this study. Compared to the original algorithm, the proposed method reduces slot collision probability by 16.78% and improves slot utilization by 16.12% under the original transmission scheme. When compared to the fixed time-window dynamic slot allocation algorithm described in [Ref. 19], collision probability decreases by 10.80%, and slot utilization increases by 10.48%. Three transmission schemes designed to expand slot availability are also proposed. Simulation results show that as the number of available slots increases, the performance advantage of the improved algorithm becomes more pronounced. Across all transmission schemes, when maximum surveillance capacity is reached, the improved algorithm reduces slot collision probability by approximately 16% and increases slot utilization by approximately 17%. In addition, a surveillance capacity calculation model for the UAT2 data link is established. Quantitative simulation results based on slot collision probability and BER performance confirm that the effective surveillance capacity of the improved algorithm is significantly higher than that of both the original algorithm and the fixed time-window dynamic slot allocation algorithm. It is also demonstrated that the proposed algorithm achieves optimal surveillance performance when all aircraft are equipped with both transmission and reception capabilities. Future research will focus on optimizing slot allocation algorithms for scenarios where aircraft possess only transmission capability.
A Distributed KBB Index Tree Multi-Keyword Fuzzy Sorting Search Scheme
SUN Jin, SONG Nana, WANG Lu, KANG Mengna, YE Kexin
Available online  , doi: 10.11999/JEIT250151
Abstract:
  Objective  The rapid advancement of information technology has driven significant transformation in the medical domain. As a cornerstone of medical informatization, Electronic Health Records (EHRs) play a critical role in improving healthcare service efficiency and supporting medical research. Effective use of patient EHRs can enhance diagnostic and treatment processes and offer reliable data support for disease prevention and novel drug discovery. However, conventional EHR systems face several challenges, including data silos, secure sharing complexity, and heightened privacy risks. Data silos hinder the seamless exchange and integration of EHRs across institutions, limiting the potential for collaborative medical practices. A central challenge in secure data sharing lies in ensuring data interoperability without compromising patient privacy—a problem of urgent concern in modern healthcare systems. Moreover, privacy breaches not only jeopardize patient welfare but also damage the credibility and trustworthiness of healthcare institutions. To address these issues, this paper proposes a distributed multi-keyword fuzzy sorting search scheme based on a Keywords Balanced Binary (KBB) index tree. The scheme leverages the decentralized, tamper-proof, and traceable features of blockchain technology to create a secure and efficient framework for EHR sharing. By eliminating data silos, the blockchain facilitates cross-institutional interoperability, while smart contracts maintain strict privacy protections throughout the data sharing process.  Methods   The scheme first applies a K-means clustering algorithm to categorize EHRs, followed by hierarchical encrypted storage via the Interplanetary File System (IPFS). This dual-layer approach ensures distributed data management while protecting user privacy and enhancing storage robustness and fault tolerance. To improve search performance, the Porter stemming algorithm standardizes query keywords, reducing ambiguity from semantic variations and enabling consistent semantic matching. A KBB index tree is then constructed over the clustered EHRs to support fuzzy keyword matching and high retrieval precision. This structure incorporates adaptive sorting to reduce search latency. Access control is enforced using smart contracts, which implement fine-grained, role-based authentication to ensure that only authorized users can retrieve EHRs, thereby minimizing the risk of data leaks. During query processing, ranked searches are conducted within the KBB index tree. Once matching encrypted EHRs are located, data retrieval is performed via IPFS. Blockchain-based hashing ensures the integrity and immutability of the retrieved data, protecting it from tampering or corruption. Finally, users decrypt the data with private keys to access the original EHR content, completing the secure retrieval process.  Results and Discussions  Simulation experiments demonstrate that the proposed scheme offers superior implementation efficiency compared to related methods in terms of both time and storage overhead. The index tree construction time is shorter than that reported in[Ref. 7] and[Ref. 8], with an efficiency improvement of 80%–85% (Fig. 3). Although the trap generation time is longer than in[Ref. 7] and[Ref. 15] (Fig. 4), this increase stems from the scheme’s support for multiuser search, whereas[Ref. 7] only supports single-user search. The combined search time for indexing and trap generation is lower than in[Ref. 7],[Ref. 8], and[Ref. 15], with search efficiency improved by 50.5% and 75.8%, respectively (Fig. 5). The ciphertext generation and decryption times are also markedly better than those in the comparative literature, with improvements of 33.3% and 60%, respectively (Fig. 6, Fig. 7). Furthermore, the storage costs for keys and ciphertext in this scheme are lower than those of the other methods (Fig. 8, Fig. 9).  Conclusions  The proposed scheme substantially improves multi-keyword search efficiency, enabling users to rapidly locate target information within large-scale EHR datasets while enhancing the accuracy of search results. The fuzzy search functionality allows users to retrieve data even when the exact form of the keywords is unknown, increasing the flexibility and applicability of the search process. The encryption algorithm has been rigorously tested in experimental analyses, demonstrating that it maintains semantic security under the known ciphertext model, thereby ensuring EHR confidentiality. In addition, the algorithm performs well in terms of implementation efficiency, meeting the practical needs of EHR applications in medical contexts while maintaining strong security guarantees. By combining high-precision multi-keyword indexing with robust privacy protection, this scheme offers a scalable and secure framework for EHR sharing that meets the interoperability and confidentiality requirements of modern healthcare systems.
A Particle-Swarm-Confinement-based Zonotopic Space Filtering Algorithm and Its Application to State of Charge Estimation for Lithium-Ion Batteries
HUO Leiting, WANG Ziyun, WANG Yan
Available online  , doi: 10.11999/JEIT250437
Abstract:
  Objective  The State of Charge (SOC) is a critical indicator for evaluating the remaining capacity and health status of lithium-ion batteries, which are widely deployed in electric vehicles, portable electronics, and energy storage systems. Accurate SOC estimation is essential for maintaining safe operation, extending battery life, and optimizing energy utilization. However, practical SOC estimation is complicated by measurement uncertainties and disturbances, particularly Unknown But Bounded (UBB) noise arising from sensor errors, environmental fluctuations, and battery aging. Conventional filtering algorithms, such as Kalman filters, often depend on probabilistic noise assumptions and tend to perform poorly when actual noise characteristics deviate from Gaussian distributions. This study addresses these limitations by proposing a Particle-Swarm-Confinement-based Zonotopic Space Filtering (PSC-ZSF) algorithm to enhance estimation robustness and reduce conservatism, with specific emphasis on high-dimensional dynamic systems such as lithium-ion battery SOC estimation.  Methods  The PSC-ZSF algorithm combines the robustness of set-membership filtering with the global optimization capabilities of Particle Swarm Optimization (PSO), integrating geometric uncertainty representation with heuristic search strategies. A zonotopic feasible state set is first constructed by propagating system model predictions and refining them with measurement updates, thereby representing the bounded uncertainty in system states. A swarm of particles is then randomly initialized within this zonotopic space to explore potential state estimates. Particle movement follows PSO-based velocity and position updates, leveraging both individual experience and swarm intelligence to identify optimal state estimates. Fitness functions quantify the consistency between candidate states and observed measurements, guiding particle convergence toward more plausible regions. To maintain algorithm stability, a boundary detection mechanism identifies particles that exceed the zonotopic feasible region. Out-of-bound particles are projected back into the feasible set by solving a quadratic programming problem that minimizes positional distortion while preserving spatial characteristics. Additionally, a dynamic contraction strategy adaptively tightens the zonotopic boundaries by scaling the normal vectors of the defining hyperplanes, effectively shrinking the search space as the particle swarm converges. This contraction improves estimation precision and reduces conservatism without incurring excessive computational overhead. The approach utilizes Minkowski sum properties intrinsic to zonotopes and utilizes efficient geometric computations to balance accuracy and efficiency. For experimental validation, the PSC-ZSF algorithm is applied to SOC estimation of lithium-ion batteries modeled by a discrete-time equivalent circuit that incorporates polarization resistance and capacitance effects. Real-world data are collected from a 18650 lithium-ion battery undergoing constant current discharge at room temperature. The system model considers UBB process and measurement noise, with parameters calibrated through empirical measurements. The performance of the proposed method is benchmarked against Ellipsoidal Set-Membership Filtering (ESMF) and Zonotopic Set-Membership Filtering (ZSMF) methods by comparing feasible state set volumes and the tightness of estimated boundaries.  Results and Discussions  The proposed PSC-ZSF algorithm demonstrates reliable confinement of particle swarms within the zonotopic feasible region throughout iterative optimization, effectively preventing particle divergence and improving estimation stability and reliability (Fig. 1). Comparative analysis indicates that PSC-ZSF consistently achieves significantly smaller feasible state set volumes at each time step compared to ESMF and ZSMF methods, reflecting reduced estimation redundancy and improved compactness (Fig. 3). The ESMF method guarantees that the true state remains enclosed; however, it produces overly conservative ellipsoidal bounds, especially under conditions of rapid system dynamics, which compromises estimation informativeness and responsiveness. The ZSMF method improves upon this by employing zonotopic bounds but still yields relatively broad estimation regions due to fixed zonotope geometries and cautious boundary updates. In contrast, PSC-ZSF adaptively refines the zonotopic boundaries based on real-time particle swarm distributions, leading to consistently tighter, more accurate boundaries that closely track the true SOC and polarization voltage trajectories (Figs. 4 and 5). This adaptive boundary contraction strategy enhances estimation precision while preserving robustness. Moreover, computational complexity analysis shows that although particle projection and boundary scaling introduce additional per-iteration operations, the accelerated convergence of PSC-ZSF reduces overall iteration requirements. This trade-off ensures computational feasibility for real-time SOC estimation in battery management systems.  Conclusions  This study proposes a Particle-Swarm-Confinement-Based Zonotopic Space Filtering (PSC-ZSF) algorithm that integrates set-membership filtering with PSO to address state estimation under unknown but bounded noise. The PSC-ZSF algorithm ensures that particle swarms remain confined within a zonotopic feasible region through optimal projection and dynamically contracts the zonotope boundaries via hyperplane scaling. This approach improves estimation accuracy and reduces conservatism. Application to lithium-ion battery SOC estimation confirms the approach’s superiority over conventional approaches, providing more precise and stable state boundaries while maintaining computational efficiency suitable for real-time applications. Future work will focus on extending the PSC-ZSF algorithm to complex dynamic systems such as autonomous vehicle navigation and smart grid state estimation to further assess scalability and practical applicability.
A Survey on System and Architecture Optimization Techniques for Mixture-of-Experts Large Language Models
WANG Zehao, ZHU Zhenhua, XIE Tongxin, WANG Yu
Available online  , doi: 10.11999/JEIT250407
Abstract:
The Mixture-of-Experts (MoE) framework has become a pivotal approach for enhancing the knowledge capacity and inference efficiency of Large Language Models (LLMs). Conventional methods for scaling dense LLMs have reached significant limitations in training and inference due to computational and memory constraints. MoE addresses these challenges by distributing knowledge representation across specialized expert sub-networks, enabling parameter expansion while maintaining efficiency through sparse expert activation during inference. However, the dynamic nature of expert activation introduces substantial challenges in resource management and scheduling, necessitating targeted optimization at both the system and architectural levels. This survey focuses on the deployment of MoE-based LLMs. It first reviews the definitions and developmental trajectory of MoE, followed by an in-depth analysis of current system-level optimization strategies and architectural innovations tailored to MoE. The paper concludes by summarizing key findings and proposing prospective optimization techniques for MoE-based LLMs.  Significance   The MoE mechanism offers a promising solution to the computational and memory limitations of dense LLMs. By distributing knowledge representation across specialized expert sub-networks, MoE facilitates model scaling without incurring prohibitive computational costs. This architecture alleviates the bottlenecks associated with training and inference in traditional dense models, marking a notable advance in LLM research. Nonetheless, the dynamic expert activation patterns inherent to MoE introduce new challenges in resource scheduling and management. Overcoming these challenges requires targeted system- and architecture-level optimizations to fully harness the potential of MoE-based LLMs.  Progress   Recent advancements in MoE-based LLMs have led to the development of various optimization strategies. At the system level, approaches such as automatic parallelism, communication–computation pipelining, and communication operator fusion have been adopted to reduce communication overhead. Memory management has been improved through expert prefetching, caching mechanisms, and queue scheduling policies. To address computational load imbalance, both offline scheduling methods and runtime expert allocation strategies have been proposed, including designs that leverage heterogeneous CPU–GPU architectures. In terms of hardware architecture, innovations include dynamic adaptation to expert activation patterns, techniques to overcome bandwidth limitations, and near-memory computing schemes that improve deployment efficiency. In parallel, the open-source community has developed supporting tools and frameworks that facilitate the practical deployment and optimization of MoE-based models.  Conclusions  This survey presents a comprehensive review of system and architectural optimization techniques for MoE-based LLMs. It highlights the importance of reconciling parameter scalability with computational efficiency through the MoE framework. The dynamic nature of expert activation poses significant challenges in scheduling and resource management, which this survey systematically addresses. By evaluating current optimization techniques across both system and hardware layers, the paper offers key insights into the state of the field. It also proposes directions for future work, providing a reference for researchers and practitioners seeking to improve the performance and scalability of MoE-based models. The findings emphasize the need for continued innovation across algorithm development, system engineering, and architectural design to fully realize the potential of MoE in real-world applications.  Prospects   Future research on MoE-based LLMs is expected to advance the integration of algorithm design, system optimization, and hardware co-design. Key research directions include resolving load imbalance and maximizing resource utilization through adaptive expert scheduling algorithms, refining system frameworks to support dynamic sparse computation more effectively, and exploring hardware paradigms such as near-memory computing and hierarchical memory architectures. These developments aim to deliver more efficient and scalable MoE model deployments by fostering deeper synergy between software and hardware components.
Hybrid Far-Near Field Channel Estimation for XL-RIS Assisted Communication Systems
SHAO Kai, HUA Fanyu, WANG Guangyu
Available online  , doi: 10.11999/JEIT250306
Abstract:
  Objective  With the rapid development of sixth-generation mobile communication, Extra-Large Reconfigurable Intelligent Surfaces (XL-RIS) have attracted significant attention due to their potential to enhance spectral efficiency, expand coverage, and reduce energy consumption. However, conventional channel estimation methods, primarily based on Far-Field (FF) or near-field (NF) models, face limitations in addressing the hybrid far-NF environment that arises from the coexistence of NF spherical waves and FF planar waves in XL-RIS deployments. These limitations restrict the intelligent control capability of RIS technology due to inaccurate channel modeling and reduced estimation accuracy. To address these challenges, this paper constructs a hybrid-field channel model for XL-RIS and proposes a robust channel estimation method to resolve parameter estimation challenges under coupled FF and NF characteristics, thereby improving channel estimation accuracy in complex propagation scenarios.  Methods  For channel estimation in XL-RIS-aided communication systems, several key challenges must be addressed, including the modeling of hybrid far-NF cascaded channels, separation of FF and NF channel components, and individual parameter estimation. To capture the hybrid-field effects of XL-RIS, a hybrid-field cascaded channel model is constructed. The RIS-to-User Equipment (UE) channel is modeled as a hybrid far-NF channel, whereas the Base Station (BS)-to-RIS channel is characterized under the FF assumption. A unified representation of FF and NF models is established by introducing equivalent cascaded angles for the angle of departure and angle of arrival on the RIS side. The XL-RIS hybrid-field cascaded channel is parameterized through BS arrival angles, RIS-UE cascaded angles, and distances. To reduce the computational complexity of joint parameter estimation, a Two-Stage Hybrid-Field (TS-HF) channel estimation scheme is proposed. In the first stage, the BS arrival angle is estimated using the MUltiple SIgnal Classification (MUSIC) algorithm. In the second stage, a Hybrid-Field forward spatial smoothing Rank-reduced MUSIC (HF-RM) algorithm is proposed to estimate the parameters of the RIS-UE hybrid-field channel. The received signals are pre-processed using a forward spatial smoothing technique to mitigate multipath coherence effects. Subsequently, the Rank-reduced MUSIC (RM) algorithm is applied to separately estimate the FF and NF angle parameters, as well as the NF distance parameter. During this stage, a power spectrum comparison scheme is designed to distinguish FF and NF angles based on power spectral characteristics, thereby providing high-precision angular information to support NF distance estimation. Finally, channel attenuation is estimated using the least squares method. To validate the effectiveness of the proposed hybrid-field channel estimation scheme, comparative analyses are conducted against FF, NF, and the proposed TS-HF-RM schemes. The FF estimation approximates the hybrid-field channel using a FF channel model and estimates FF angle parameters with the MUSIC algorithm, referred to as the TS-FF-M scheme. The NF estimation applies a NF channel model to characterize the hybrid channel and estimates angle and distance parameters using the RM algorithm, referred to as the TS-NF-RM scheme. To further evaluate the estimation performance, additional benchmark schemes are considered, including the Two-Stage Near-Field Orthogonal Matching Pursuit (TS-NOMP) scheme, the Two-Stage Hybrid Orthogonal Matching Pursuit with Prior (TS-HOMP-P) scheme that requires prior knowledge of FF and NF quantities, and the Two-Stage Hybrid Orthogonal Matching Pursuit with No Prior (TS-HOMP-NP) scheme that operates without requiring such prior information.  Results and Discussions  Compared with the TS-FF-M and TS-NF-RM schemes, the proposed TS-HF-RM approach achieves effective separation and accurate estimation of both FF and NF components by jointly modeling the hybrid-field channel. The method consistently demonstrates superior estimation accuracy across a wide range of Signal-to-Noise Ratio (SNR) conditions (Fig. 4). These results confirm both the necessity of hybrid-field channel modeling and the effectiveness of the proposed estimation scheme. Experimental findings show that the TS-HF-RM approach significantly improves channel estimation performance in XL-RIS-assisted communication systems. Further comparative analysis reveals that the TS-HF-RM scheme outperforms TS-NOMP and TS-HOMP-P by mitigating power leakage effects and overcoming limitations associated with unknown path numbers through distinct processing of FF and NF components. Without requiring prior knowledge of the propagation environment, the proposed method achieves lower Normalized Mean Square Error (NMSE) while demonstrating improved robustness and estimation precision (Fig. 5). Although TS-HOMP-NP also operates without prior field information, the TS-HF-RM scheme provides superior parameter resolution, attributed to its subspace decomposition principle. Additionally, both the TS-HF-RM and TS-HOMP-P schemes exhibit improved performance as the number of pilot signals increases. However, TS-HF-RM consistently outperforms TS-HOMP-P under low-SNR conditions (0 dB). At high SNR (10 dB) with a limited number of pilot signals (<280), TS-HOMP-P temporarily achieves better performance due to its higher sensitivity to SNR. Nevertheless, the proposed TS-HF-RM approach demonstrates greater stability and adaptability under low-SNR and resource-constrained conditions (Fig. 6).  Conclusions  This study addresses the challenge of hybrid-field channel estimation for XL-RIS by constructing a hybrid-field cascaded channel model and proposing a two-stage estimation scheme. The HF-RM algorithm is specifically designed for accurate hybrid component estimation in the second stage. Theoretical analysis and simulation results demonstrate the following: (1) The hybrid-field model reduces inaccuracies associated with traditional single-field assumptions, providing a theoretical foundation for reliable parameter estimation in complex propagation environments; (2) The proposed TS-HF-RM algorithm enables high-resolution parameter estimation with effective separation of FF and NF components, achieving lower NMSE compared to hybrid-field OMP-based methods.
Available online  , doi: 10.11999/JEIT250305
Abstract:
UAV-Assisted Intelligent Data Collection and Computation Offloading for Railway Wireless Sensor Networks
YAN Li, WANG Junkai, FANG Xuming, LIN Wei, LIANG Yiqun
Available online  , doi: 10.11999/JEIT250340
Abstract:
  Objective  Ensuring the safety and stability of train operations is essential in the advancement of railway intelligence. The growing maturity of Wireless Sensor Network (WSN) technology offers an efficient, reliable, low-cost, and easily deployable approach to monitoring railway operating conditions. However, in complex and dynamic maintenance environments, WSNs encounter several challenges, including weak signal coverage at monitoring sites, limited accessibility for tasks such as sensor node battery replacement, and the generation of large volumes of monitoring data. To address these issues, this study proposes a multi-Unmanned Aerial Vehicle (UAV)-assisted method for data collection and computation offloading in railway WSNs. This approach enhances overall system energy efficiency and data freshness, offering a more effective and robust solution for railway safety monitoring.  Methods  An intelligent data collection and computation offloading system is constructed for multi-UAV-assisted railway WSNs. UAV flight constraints within railway safety protection zones are considered, and wireless sensing services are prioritized to ensure preferential transmission for safety-critical tasks. To balance energy consumption and data freshness, the system optimization objective is defined as the weighted sum of UAV energy consumption, WSN energy consumption, and the Age of Information (AoI). A joint optimization algorithm based on Multi-Agent Soft Actor-Critic (MASAC) is proposed, which balances exploration and exploitation through entropy regularization and adaptive temperature parameters. This approach enables efficient joint optimization of UAV trajectories and computation offloading strategies.  Results and Discussions  (1) Compared with the Multi-Agent Deep Deterministic Policy Gradient (MADDPG), MASAC-Greedy, and MASAC-AOU algorithms, the MASAC-based scheme converges more rapidly and demonstrates greater stability (Fig. 4), ultimately achieving the highest reward. In contrast, MADDPG exhibits slower learning and less stable performance. (2) The comparison of multi-UAV flight trajectories under different algorithms shows that the proposed MASAC algorithm enables effective collaboration among UAVs, with each responsible for monitoring distinct regions while strictly adhering to railway safety protection zone constraints (Fig. 5). (3) The MASAC algorithm yields the best objective function value across all evaluated algorithms (Fig. 6). (4) As the number of sensors and the AoI weight increase, UAV energy consumption rises for all algorithms; however, the MASAC algorithm consistently maintains the lowest energy consumption (Fig. 7). (5) In terms of sensor node energy consumption, MADDPG achieves the lowest value, but at the expense of information freshness (Fig. 8). (6) Regarding average AoI performance, the MASAC algorithm performs best across a range of sensor densities and AoI weight settings, with the greatest improvements observed under higher AoI weight conditions (Fig. 9). (7) The AoI performance comparison by sensor type (Table 2) confirms that the system effectively supports priority-based data collection services.  Conclusions  This study proposes a MASAC-based intelligent data collection and computation offloading scheme for railway WSNs supported by multiple UAVs, addressing critical challenges such as limited WSN battery life and the high real-time computational demands of complex railway environments. The proposed algorithm jointly optimizes UAV flight trajectories and computation offloading strategies by integrating considerations of UAV and WSN energy consumption, data freshness, sensing service priorities, and railway safety protection zone constraints. The optimization objective is to minimize the weighted sum of average UAV energy consumption, average WSN energy consumption, and average WSN AoI. Simulation results demonstrate that the proposed scheme outperforms baseline algorithms across multiple performance metrics. Specifically, it achieves faster convergence, efficient multi-UAV collaboration that avoids resource redundancy and spatial overlap, and superior results in UAV energy consumption, sensor node energy consumption, and average AoI.
An Optimization Design Method for Zero-Correlation Zone Sequences Based on Newton’s Method
HU Enbo, LIU Tao, LI Yubo
Available online  , doi: 10.11999/JEIT250394
Abstract:
  Objective  Sequences with favorable correlation properties are widely applied in radar and communication systems. Sequence sets with zero or low correlation characteristics enhance radar resolution, target detection, imaging quality, and information acquisition, while also improving the omnidirectional transmission capability of massive multiple-input multiple-output (MIMO) systems. Designing aperiodic Zero Correlation Zone (ZCZ) sequence sets with excellent correlation performance is therefore critical for both wireless communication and radar applications. For example, aperiodic Z-Complementary Set (ZCS) sequence sets are often used in omnidirectional precoding for MIMO systems, whereas aperiodic ZCZ sequence sets are employed in integrated MIMO radar-communication systems. These ZCZ sequence sets are thus valuable across a range of system applications. However, most prior studies rely on analytical construction methods, which impose constraints on parameters such as sequence length and the number of sequences, thereby limiting design flexibility and practical applicability. This study proposes a numerical optimization approach for designing ZCS and aperiodic ZCZ sequence sets with improved correlation properties and greater parametric flexibility. The method minimizes the Complementary Peak Sidelobe Level (CPSL) and Weighted Peak Sidelobe Level (WPSL) using Newton’s method to achieve superior sequence performance.  Methods  This study proposes an optimization-based design method using Newton’s method to construct both aperiodic ZCS sequence sets and aperiodic ZCZ sequence sets with low sidelobe levels and flexible parameters. The optimization objective is first formulated using the CPSL and WPSL. The problem is then reformulated as an equivalent system of nonlinear equations, which is solved using Newton’s method. To reduce computation time, partial derivatives are approximated using numerical differentiation techniques. A loop iteration strategy is employed to address multiple constraints during the optimization process. To ensure algorithmic convergence, Armijo’s rule is used for step size selection, promoting stable descent of the objective function along the defined search direction.  Results and Discussions  The aperiodic ZCS sequence set is constructed using Newton’s method. As the number of sequences increases, the CPSL progressively decreases, falling below –300 dB when \begin{document}$M \geqslant 2$\end{document}. The proposed method yields better sidelobe performance than the improved Iterative Twisted Approximation (ITORX) algorithm (Fig. 1). The performance of ZCS sequences generated by both methods is evaluated under different ZCZ conditions. While both approaches achieve low CPSL, Newton’s method yields aidelobe levels closer to the ideal value (Fig. 2). Convergence behavior is assessed using CPSL and the number of iterations. The improved ITROX algorithm typically requires around 20000 iterations to converge, with increasing iterations as ZCZ size grows. In contrast, Newton’s method achieves rapid convergence within approximately 10 iterations (Figs. 3 and 4). The aperiodic ZCZ sequence set constructed using Newton’s method exhibits autocorrelation and cross-correlation peak sidelobe levels below –300 dB within the ZCZ. Moreover, Newton’s method achieves the lowest WPSL, offering the best overall performance among all tested methods (Fig. 5). The smooth convergence curves further confirm the algorithm’s stability when applied to aperiodic ZCZ sequence construction (Fig. 6).  Conclusions  This study proposes an optimization-based algorithm for designing aperiodic ZCS and aperiodic ZCZ sequence sets using Newton’s method, aiming to address the limitations of fixed parameters and high peak sidelobe levels found in existing approaches. Two optimization problems are formulated by minimizing the WPSL and CPSL, respectively. To simplify computation, the optimization tasks are converted into systems of nonlinear equations, which are solved using Newton’s method. The Jacobian matrix is computed via numerical differentiation to reduce computational cost. A loop iteration strategy is introduced to meet multiple constraints in the construction of aperiodic ZCZ sequences. Simulation results confirm that the proposed method yields sequence sets with excellent correlation properties and flexible parameter configurations. By tuning the weighting coefficients, low sidelobe levels can be achieved in specific regions of interest, accommodating different application requirements. The combination of flexible design parameters and favorable correlation performance makes the proposed sequences suitable for a wider range of practical scenarios.
Continuous Federation of Noise-resistant Heterogeneous Medical Dialogue Using the Trustworthiness-based Evaluation
LIU Yupeng, ZHANG Jiang, TANG Shichen, MENG Xin, MENG Qingfeng
Available online  , doi: 10.11999/JEIT250057
Abstract:
  Objective   To address the key challenges of client model heterogeneity, data distribution heterogeneity, and text noise in medical dialogue federated learning, this paper proposes a trustworthiness-based, noise-resistant heterogeneous medical dialogue federated learning method, termed FedRH. FedRH enhances robustness by improving the objective function, aggregation strategy, and local update process, among other components, based on credibility evaluation.  Methods   Model training is divided into a local training stage and a heterogeneous federated learning stage. During local training, text noise is mitigated using a symmetric cross-entropy loss function, which reduces the risk of overfitting to noisy text. In the heterogeneous federated learning stage, an adaptive aggregation mechanism incorporates clean, noisy, and heterogeneous client texts by evaluating their quality. Local parameter updates consider both local and global parameters simultaneously, enabling continuous adaptive updates that improve resistance to both random and structured (syntax/semantic) noise and model heterogeneity. The main contributions are threefold: (1) A local noise-resistant training strategy that uses symmetric cross-entropy loss to prevent overfitting to noisy text during local training; (2) A heterogeneous federated learning approach based on client trustworthiness, which evaluates each client’s text quality and learning effectiveness to compute trust scores. These scores are used to adaptively weight clients during model aggregation, thereby reducing the influence of low-quality data while accounting for text heterogeneity; (3) A local continuous adaptive aggregation mechanism, which allows the local model to integrate fine-grained global model information. This approach reduces the adverse effects of global model bias caused by heterogeneous and noisy text on local updates.  Results and Discussions   The effectiveness of the proposed model is systematically validated through extensive, multi-dimensional experiments. The results indicate that FedRH achieves substantial improvements over existing methods in noisy and heterogeneous federated learning scenarios (Table 2, Table 3). The study also presents training process curves for both heterogeneous models (Figure 3) and isomorphic models (Figure 6), supplemented by parameter sensitivity analysis, ablation experiments, and a case study.  Conclusions   The proposed FedRH framework significantly enhances the robustness of federated learning for medical dialogue tasks in the presence of heterogeneous and noisy text. The main conclusions are as follows: (1) Compared to baseline methods, FedRH achieves superior performance in client-side models under heterogeneous and noisy text conditions. It demonstrates improvements across multiple metrics, including precision, recall, and factual consistency, and converges more rapidly during training. (2) Ablation experiments confirm that both the symmetric cross-entropy-based local training strategy and the credibility-weighted heterogeneous aggregation approach contribute to performance gains.
Dynamic Adaptive Partitioning of Deep Neural Networks Based on Early Exit Mechanism under Edge-End Collaboration
DING Nan, WANG Jiajia, JI Chenghui, HU Chuangye, XU Li
Available online  , doi: 10.11999/JEIT250291
Abstract:
  Objective  The deployment of Deep Neural Networks (DNNs) for inference tasks in industrial intelligence applications is constrained by the complexity of Directed Acyclic Graph (DAG) structures and dynamic resource limitations, making it challenging to simultaneously optimize both latency and accuracy. Existing methods are generally restricted to chain-structured DNNs and lack adaptive mechanisms to accommodate network variability and heterogeneous computational resources. To address these limitations, this paper proposes a Dynamic Adaptive Partitioning framework based on a Deep Early Exit mechanism (DAPDEE), designed to achieve low-latency, high-accuracy inference through edge-end collaborative computing. The significance of this work lies in its potential to provide a generalizable solution suitable for diverse network conditions and computing environments.  Methods  The proposed DAPDEE framework incorporates several technical innovations: First, it abstracts both chain and complex DNN architectures into a unified DAG representation, establishing a general topological foundation for partition optimization. Second, it employs offline optimization of early exit classifiers using a multi-task learning approach, requiring only the loading of pre-trained model parameters during deployment. Combined with real-time indicators, such as network bandwidth and terminal computational load, this enables dynamic selection of optimal exit points and partitioning strategies. Finally, an inverse search mechanism is applied to jointly optimize latency and accuracy, aiming to minimize single-frame end-to-end delay under light workloads and to maximize system throughput under heavy workloads. Through these strategies, the framework enables efficient inference suitable for time-sensitive scenarios, including smart manufacturing and autonomous driving.  Results and Discussions  Experimental results demonstrate that the DAPDEE framework substantially improves performance compared to conventional Device-Only methods under varying network conditions. Specifically, under CAT1, 3G, and 4G networks, DAPDEE achieves latency reductions of up to 7.7% under heavy loads and 7.5% under light loads, with throughput improvements reaching up to 9.9 times. Notably, the accuracy loss remains consistently below 1.2% (Fig. 6, Fig. 7), confirming the framework’s ability to maintain reliable inference performance. These results verify the effectiveness of DAPDEE in adapting to dynamic network environments and heterogeneous computational loads. For instance, when the bandwidth is fixed at 1.1 Mbps (3G), the optimal partition strategy adjusts in response to varying latency constraints, revealing a positive correlation between relaxed latency requirements and deeper exit points (Fig. 6). Conversely, with fixed latency constraints and increasing bandwidth, the partition point progressively shifts toward the terminal device, reflecting enhanced resource utilization on the end side (Fig. 7). Furthermore, practical deployments on a PC and a Raspberry Pi-based intelligent vehicle validate the theoretical performance gains, as demonstrated by the applied partitioning strategies (Algorithm 1, Algorithm 2).  Conclusions  In summary, the proposed DAPDEE framework effectively addresses the challenge of balancing inference efficiency and accuracy in edge-end collaborative scenarios involving complex DAG-structured DNNs. By integrating early exit mechanisms with dynamic partitioning strategies and multidimensional load evaluation, DAPDEE exhibits strong adaptability and robustness under diverse network conditions and resource constraints. These findings advance the current state of DNN partitioning methodologies and offer practical insights for optimizing cloud-edge-terminal architectures and reinforcement learning-based adaptive mechanisms. Nonetheless, areas for further improvement remain. These include incorporating multi-task concurrency, refining the energy consumption model, and enhancing real-time partitioning efficiency for complex DAG topologies. Future research will focus on extending the framework to support multi-task collaborative optimization and reducing the computational complexity of online partitioning algorithms for DAG-structured DNNs.
Dual-Reconfigurable Intelligent Surface Phase Shift Optimization and Unmanned Aerial Vehicle Trajectory Control for Vehicle Communication
CHANG Kuan, ZHANG Lei, WANG Yu, SHANG Yulong, CHEN Weicong, MA Junchao
Available online  , doi: 10.11999/JEIT250274
Abstract:
This study considers a scenario in which an Unmanned Aerial Vehicle (UAV) equipped with a Reconfigurable Intelligent Surface (RIS) cooperates with a fixed RIS to enhance communication with a mobile User Equipment (UE) vehicle. A joint optimization problem is formulated to maximize the UE’s communication rate by controlling the UAV’s flight trajectory and the phase shifts of both RISs. Given the system complexity and environmental dynamics, a solution is proposed that integrates a Deep Deterministic Policy Gradient (DDPG) algorithm with phase-shift alignment to optimize continuous UAV trajectories and RIS configurations. Simulation results confirm that the proposed method achieves stable reward convergence within 1,000 training episodes. Compared with benchmark approaches, the algorithm improves communication rates by at least 3 dB over random trajectory and phase-shift strategies in dual-RIS deployments. The study further presents optimal UAV trajectories under varying base station and RIS placements and evaluates algorithm performance across different vehicle speeds.  Objective   This study investigates a vehicular communication scenario in which a UAV-mounted RIS cooperates with a fixed RIS to assist a mobile UE device. A joint optimization framework is established to maximize UE communication rates during movement by simultaneously optimizing the UAV trajectory and the phase shifts of both RISs. To address system complexity and environmental dynamics, a DDPG algorithm is employed for continuous trajectory control, while a low-complexity phase-shift alignment method configures the RISs. Simulation results show that the proposed algorithm achieves stable reward convergence within 1,000 training episodes and improves communication rates by at least 3 dB compared with randomized trajectory and phase-shift baselines. It also outperforms alternative reinforcement learning approaches, including Twin Delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC). Optimal UAV trajectories are derived for various base station and RIS deployment scenarios, with additional simulations confirming robustness across a range of vehicle speeds.  Methods   This study establishes a Multiple-Input Single-Output (MISO) system in which a UAV-mounted RIS cooperates with a fixed RIS to support mobile vehicular communication, with the objective of maximizing user information rates. To address the complexity of continuous trajectory control under dynamic environmental conditions, a DDPG-based algorithm is developed. The phase shifts of RIS elements are optimized using a low-complexity alignment method. A reward function based on the achievable information rates of vehicular users is designed to guide the agent’s actions and facilitate policy learning. The proposed framework enhances adaptability by dynamically optimizing UAV trajectories and RIS configurations under time-varying channel conditions.  Results and Discussions   (1) The convergence behavior of the DDPG algorithm is verified in Fig. 3, where the reward values progressively converge as the number of training episodes increases. (2) Fig. 4 shows the effect of varying the number of RIS elements on system performance, indicating that additional elements lead to a steady increase in reward values, confirming the channel gain enhancement provided by RIS deployment. (3) As shown in Fig. 5, the DDPG algorithm outperforms baseline methods and demonstrates greater adaptability to target scenarios; concurrently, optimized RIS phase shifts yield significantly higher rewards than random configurations, validating the proposed phase-alignment strategy. (4) Figs. 6–7 highlight notable variations in UAV trajectories and system performance across different base station and RIS deployments, demonstrating the adaptability of the trajectory optimization strategy. Fig. 8 further compares performance across scenarios with optimized UAV trajectories, highlighting the algorithm’s versatility. (5) System performance under different UE mobility speeds is evaluated in Fig. 9, showing a performance decline at higher speeds, indicating strong efficacy in low-speed environments but reduced effectiveness under high-speed conditions. These results collectively illustrate the operational strengths and limitations of the proposed framework in dynamic vehicular communication systems.  Conclusions   This paper investigates a vehicular communication scenario assisted by both fixed and UAV-mounted mobile RISs, aiming to maximize UE information rates under dynamic mobility conditions. A joint optimization framework is developed, combining dual-RIS phase shift alignment based on channel state information with UAV trajectory planning using a DDPG algorithm. The proposed method features a low-complexity design that addresses both network architecture and RIS configuration challenges. Extensive simulations under varying vehicular speeds, RIS element counts, and base station deployments demonstrate the algorithm’s superiority over SAC, TD3, and randomized phase shift strategies. Results further highlight the framework’s adaptability to heterogeneous base station–RIS topologies and reveal performance degradation at higher vehicle speeds, indicating the need for future research into real-time adaptive mechanisms.
Cross Modal Hashing of Medical Image Semantic Mining for Large Language Model
LIU Qinghai, WU Qianlin, LUO Jia, TANG Lun, XU Liming
Available online  , doi: 10.11999/JEIT250529
Abstract:
  Objective  A novel cross-modal hashing framework driven by Large Language Models (LLMs) is proposed to address the semantic misalignment between medical images and their corresponding textual reports. The objective is to enhance cross-modal semantic representation and improve retrieval accuracy by effectively mining and matching semantic associations between modalities.  Methods  The generative capacity of LLMs is first leveraged to produce high-quality textual descriptions of medical images. These descriptions are integrated with diagnostic reports and structured clinical data using a dual-stream semantic enhancement module, designed to reinforce inter-modality alignment and improve semantic comprehension. A structural similarity-guided hashing scheme is then developed to encode both visual and textual features into a unified Hamming space, ensuring semantic consistency and enabling efficient retrieval. To further enhance semantic alignment, a prompt-driven attention template is introduced to fuse image and text features through fine-tuned LLMs. Finally, a contrastive loss function with hard negative mining is employed to improve representation discrimination and retrieval accuracy.  Results and Discussions  Experiments are conducted on a multimodal medical dataset to compare the proposed method with existing cross-modal hashing baselines. The results indicate that the proposed method significantly outperforms baseline models in terms of precision and Mean Average Precision (MAP) (Table 3; Table 4). On average, a 7.21% improvement in retrieval accuracy and a 7.72% increase in MAP are achieved across multiple data scales, confirming the effectiveness of the LLM-driven semantic mining and hashing approach.  Conclusions  Experiments are conducted on a multimodal medical dataset to compare the proposed method with existing cross-modal hashing baselines. The results indicate that the proposed method significantly outperforms baseline models in terms of precision and Mean Average Precision (MAP) (Table 3; Table 4). On average, a 7.21% improvement in retrieval accuracy and a 7.72% increase in MAP are achieved across multiple data scales, confirming the effectiveness of the LLM-driven semantic mining and hashing approach.
A Collaborative Detection Method for Bauxite Quality Parameters Based on the Fusion of G-DPN and Near-Infrared Spectroscopy
ZOU Liang, REN Kelong, WU Hao, XU Zhibin, TAN Zhiyi, LEI Meng
Available online  , doi: 10.11999/JEIT250240
Abstract:
  Objective  Bauxite is a critical non-metallic mineral resource used in aluminum production, ceramic manufacturing, and refractory material processing. As global demand for aluminum and its derivatives continues to rise, improving the efficiency of bauxite resource utilization is essential. Accurate determination of quality parameters supports the reduction of waste from low-grade ores during smelting and improves overall process optimization. However, traditional chemical analyses are time-consuming, costly, complex, and subject to human error. Existing rapid testing methods, often based on machine learning, typically predict individual quality indicators and overlook correlations among multiple parameters. Deep learning, particularly multi-task learning, offers a solution to this limitation. Near-InfraRed (NIR) spectroscopy, a real-time, non-destructive analytical technique, is especially suited for assessing mineral quality. This study proposes a multi-indicator collaborative detection model—Gate-Depthwise Pointwise Network (G-DPN)—based on NIR spectroscopy to enable the simultaneous prediction of multiple bauxite quality parameters. The proposed approach addresses the limitations of conventional methods and supports efficient, accurate, and cost-effective real-time quality monitoring in industrial settings.  Methods  To accurately model the nonlinear relationships between NIR spectral features and bauxite quality parameters while leveraging inter-parameter correlations, this study proposes a dedicated representation model, G-DPN. The model incorporates large-kernel DepthWise Convolution (DWConv) to extract long-range dependencies within individual spectral channels, and PointWise Convolution (PWConv) to enable inter-channel feature fusion. A Spatial Squeeze-and-Excitation (sSE) mechanism is introduced to enhance spatial feature weighting, and residual connections support the integration of deep features. To further improve task differentiation, a Custom Gate Control (CGC) module is added to separate shared and task-specific features. Orthogonal constraints are applied within this module to reduce feature redundancy. Gate-controlled fusion enables each branch to focus on extracting task-relevant information while preserving shared representations. Additionally, quality parameter labels are normalized to address scale heterogeneity, allowing the model to establish a stable nonlinear mapping between spectral inputs and multiple output parameters.  Results and Discussions  This study applies large convolution kernels in DWConv to capture long-range dependencies within individual spectral channels (Fig. 3). Compared with conventional small-sized kernels (e.g., 3×3), which increase the receptive field but exhibit limited focus on critical spectral regions, large kernels enable more concentrated activation in key bands, thereby enhancing model sensitivity (Fig. 4). Empirical results confirm that the use of large kernels improves prediction accuracy (Table 6). Furthermore, compared to Transformer-based models, DWConv with large kernels achieves comparable accuracy with fewer parameters, offering computational efficiency. The CGC module effectively disentangles shared and task-specific features while applying orthogonal constraints to reduce redundancy. Its dynamic fusion mechanism enables adaptive feature sharing across tasks without compromising task-specific learning, thereby mitigating task interference and accounting for sample correlations (Fig. 6). Relative to conventional multi-task learning frameworks, the CGC-based architecture demonstrates superior performance in multi-parameter prediction (Table 6).  Conclusions  This study proposes a deep learning approach that integrates large-kernel DWConv and a CGC module for multi-parameter prediction of bauxite quality using NIR spectroscopy. DWConv captures long-range dependencies within spectral channels, while the CGC module leverages inter-parameter correlations to enhance feature sharing and reduce task interference. This design mitigates the effects of spectral peak overlap and establishes a robust nonlinear mapping between spectral features and quality parameters. Experiments on 424 bauxite samples show that the proposed G-DPN model achieves \begin{document}${R^2}$\end{document} values of 0.9226, 0.9377, and 0.9683 for aluminum, silicon, and iron content, respectively—outperforming conventional machine learning and existing deep learning methods. These results highlight the potential of combining NIR spectroscopy with G-DPN for accurate, efficient, and scalable mineral quality analysis, contributing to the sustainable utilization of bauxite resources.
Hyperspectral Image Denoising Algorithm via Joint Low-Rank Tensor Decomposition and Product Graph Modeling
MA Mou, CAI Mingjiao, SHEN Yu, ZHOU Fang, JIANG Junzheng
Available online  , doi: 10.11999/JEIT250130
Abstract:
Various types of noise in HyperSpectral Images (HSI) can severely degrade subsequent analysis. To address this problem, this study proposes a denoising framework based on Low-Rank Tensor Decomposition and Kronecker product Graph Laplacian Regularization (LRTDKGLR). Spatial and spectral graphs are constructed using Graph Signal Processing (GSP) and fused into a joint spatial-spectral product graph through the Kronecker product. Tucker decomposition is applied to this product graph to extract a unified low-dimensional representation, capturing the global spatial and spectral structures. A Kronecker Graph Laplacian Regularization (KGLR) term is introduced to enforce piecewise smoothness across both spatial and spectral dimensions, enhancing inter-band coherence. The denoising problem is formulated as an optimization task that integrates low-rank decomposition and graph-based regularization, solved efficiently using an Augmented Lagrangian Multiplier (ALM) approach. Experimental results on simulated and real HSI datasets demonstrate that LRTDKGLR achieves superior performance in edge preservation and noise suppression compared with existing HSI denoising methods.  Objective  Traditional denoising approaches often fail to simultaneously preserve spatial details and spectral fidelity due to the high dimensionality and complex spectral characteristics of HSI data. This study aims to develop a method that effectively suppresses mixed noise, such as Gaussian and stripe noise, while preserving critical spatial features and edge structures. The proposed approach leverages LRTDKGLR to enhance inter-dimensional correlations between spatial and spectral domains, ensuring improved noise reduction and detail preservation. Additionally, this research investigates an efficient ALM-based optimization strategy to address the HSI denoising problem, providing a more robust solution for noise reduction in high-dimensional, contaminated environments.  Methods  The proposed LRTDKGLR algorithm performs HSI denoising by effectively modeling spatial-spectral correlations. Separate spatial and spectral graphs are constructed and integrated through the Kronecker product to form a unified product graph representation. This joint graph considers both pixel intensity and the values of neighboring pixels across spatial and spectral dimensions, enabling enhanced denoising through spatial–spectral correlation modeling. Tucker decomposition is applied to extract a low-rank representation that captures global spatial-spectral relationships within the data. This decomposition facilitates effective modeling of the inherent structure of HSI while supporting the denoising process. The KGLR term models smoothness across both spatial and spectral domains, preserving essential spatial details and spectral continuity. The denoising task is formulated as an optimization problem that integrates low-rank tensor decomposition and KGLR. This problem is efficiently solved using the ALM method. The proposed approach achieves a balance between noise suppression and detail preservation, providing an effective solution for HSI denoising in high-dimensional, noisy conditions.  Results and Discussions  The proposed LRTDKGLR algorithm effectively exploits the inherent spatial and spectral correlations in HSI data. By incorporating a low-rank constraint and leveraging the similarity between adjacent spectral bands, the model enhances its capacity to capture inter-band spectral dependencies. Fig. 2 illustrates the denoising performance of various algorithms under Scenario 3, which combines Gaussian and impulse noise, applied to the 106th band of the Washington DC Mall dataset. A magnified view of a local region is also provided. The LRTV, LRTDTV, and LRTDGTV algorithms remove most of the noise but introduce over-smoothing, resulting in the loss of important edge structures. The NGmeet algorithm performs less effectively due to insufficient utilization of spatial neighborhood relationships. Although the LRMR, GLF, and TSLRLN algorithms reduce part of the noise, residual artifacts persist and over-smoothing remains evident. In contrast, the proposed LRTDKGLR algorithm demonstrates substantially improved denoising performance, supported by both visual comparisons and quantitative metrics. Fig. 3 presents the PSNR and SSIM values across multiple spectral bands under Scenario 3. The proposed method achieves higher performance in both metrics across most bands, indicating enhanced noise suppression and improved preservation of structural details. Numerical results summarized in Table 2 further confirm the effectiveness of the proposed approach, which consistently outperforms competing algorithms under identical conditions, providing superior denoising performance. Additionally, Table 3 shows that LRTDKGLR exceeds two representative deep learning-based methods, HSID-CNN and HSI-SDeCNN, across different noise levels. These findings demonstrate the robustness of the proposed method in the presence of Gaussian noise. Compared with deep learning-based approaches, LRTDKGLR offers practical advantages, including no requirement for pre-training, ease of deployment, and higher computational efficiency, which collectively improve its applicability to real-world HSI denoising tasks.  Conclusions  This study proposes an HSI denoising algorithm, LRTDKGLR, which integrates low-rank prior knowledge with a product graph model. The algorithm models HSI data as a product graph using GSP theory and applies Tucker decomposition to extract a core tensor and factor matrices, effectively capturing global spatial and spectral correlations. Simultaneously, KGLR is employed to model piecewise smoothness in both spatial and spectral dimensions, enhancing the representation of inter-dimensional relationships within the data. The denoising task is formulated as an optimization problem that combines low-rank tensor decomposition with KGLR constraints and is efficiently solved using the ALM method. Experimental results on simulated and real-world datasets demonstrate that the proposed LRTDKGLR algorithm achieves superior noise suppression while preserving structural details, validating its effectiveness for HSI denoising.
Design of Reconfigurable FeFET-MUX and Its Application in Mapping
WU Qianhuo, WANG Lunyao, ZHA Xiaojing, CHU Zhufei, XIA Yinshui
Available online  , doi: 10.11999/JEIT250263
Abstract:
  Objective  The growing demand for massive computing power and big data processing has exposed bottlenecks in conventional Von Neumann architectures, known as the “storage wall” and the “power wall”. Computing-in-Memory (CiM) offers a promising solution by integrating storage and computation, thereby reducing delays and energy consumption caused by data transfer. Emerging non-volatile memories used in CiM circuit design include Spin Transfer Torque Magnetic Random Access Memory (STT-MRAM), Phase Change Memory (PCM), Resistive Random Access Memory (ReRAM), and Ferroelectric Field-Effect Transistors (FeFETs). FeFETs have become key components in CiM designs due to their non-volatile storage capability, low power consumption, high on–off ratio, compatibility with Complementary Metal-Oxide-Semiconductor (CMOS) processes, and voltage-driven writing mechanism. Various FeFET-based CiM circuit designs have been proposed, with most focusing on array-based structures. However, the potential of FeFET-based CiM logic circuits remains underexplored. This study proposes a methodology for mapping Boolean functions onto FeFET-based CiM logic circuits by designing a reconfigurable FeFET Multiplexer (FeFET-MUX) and developing corresponding Boolean function partitioning algorithms.  Methods  The reconfigurable FeFET-MUX consists of an elementary 2-to-1 MUX, as shown in Fig. 2(a), with multiple data inputs and selection inputs, illustrated in Fig. 2(b). The sub-circuit enclosed within the dashed box in Fig. 2(b) functions as the storage element of the FeFET-MUX and is time-shared by the data pathways. To ensure correct logical function execution, at any given time, no more than one address input is permitted to write to the FeFETs, and no more than one data input is selected simultaneously. Logical functions can be expressed using Binary Decision Diagrams (BDDs). By replacing each node in the BDD with a 2-to-1 MUX, the corresponding functions can be implemented using 2-to-1 MUX circuits. This technique is also applicable to mapping with 2-to-1 FeFET-MUXs; however, its major limitation is the relatively high area overhead. In this work, instead of replacing each individual BDD node with a 2-to-1 MUX, a sub-BDD is mapped onto the proposed FeFET-MUX, reducing area consumption. To prevent logic errors caused by incorrect rewriting of stored data due to the shared structure, a BDD partitioning approach is proposed. After applying specific partitioning rules, each sub-BDD can be independently implemented using the proposed FeFET-MUX, ensuring that stored data is preserved until it is no longer needed, thereby maintaining the logical function’s correctness.The operation of the proposed FeFET-MUX follows a three-phase cycle: (1) The polarization states of the two FeFETs are programmed by applying complementary gate pulses Vg1 and Vg2; (2) During each computation cycle, the selection gate pulses are temporally modulated to select distinct input data, which are routed to the FeFET drains; (3) Finally, the output enable pulses control the transmission of the computed result to the inverter’s output for storage. The proposed BDD partitioning algorithms are presented in Algorithm 1 and Algorithm 2. The methodology proceeds as follows: First, the target BDD, constructed using the Colorado University Decision Diagram (CUDD) library, is traversed through a breadth-first search. Next, upon identifying the starting node of a sub-BDD via the subroutine “find_node_start”, the subroutine “Extend_node” iteratively evaluates candidate nodes for inclusion in the current sub-BDD. After the traversal is complete, Algorithm 1 invokes the subroutine “Out_node_check” to determine whether additional sub-BDDs need to be created.  Results and Discussions  The proposed algorithms are implemented in C++ and executed on an Ubuntu 24.04 platform with an Intel Ultra 7 processor and 32 GB of memory. The compiler used is g++, version 13.3.0. Test benchmarks are selected from open-source designs described in Verilog. Prior to mapping, the benchmarks are converted into Reduced Ordered Binary Decision Diagrams (ROBDDs) using the CUDD library. Node information is extracted and stored in data structures, and ROBDD partitioning is performed using the proposed algorithms. The experimental results show that the number of sub-BDDs is not directly determined by the number of circuit inputs or outputs but is associated with the maximum number of nodes present at the same level within the BDD. This relationship results from the constraint that each sub-BDD cannot contain multiple nodes at the same level. For example, ROBDDs such as “parity,” which contain only one sub-BDD, exhibit a maximum of one node per level. However, the reverse does not always apply. For example, the circuit “i3” has a maximum of one node per level but still requires multiple sub-BDDs due to the presence of nodes with level differences greater than one, which violate the partitioning constraint and necessitate additional sub-BDDs to ensure correct function mapping. By integrating the reconfigurable FeFET-MUX with the proposed partitioning algorithms, the number of FeFET devices required decreases by an average of 79.9% compared with conventional mapping approaches (Table 2). In addition, the methodology successfully processes large-scale benchmarks, such as “i10,” which contains over 30,000 BDD nodes, demonstrating its scalability.  Conclusion  This work presents a novel methodology for mapping Boolean functions to FeFET-based CiM logic circuits. The approach consists of two core contributions: (1) A reconfigurable FeFET-MUX circuit is designed, featuring shared FeFET components and a common output drive stage. This configuration consolidates multiple 2-to-1 MUX functions into a single circuit, significantly improving resource utilization. (2) A BDD partitioning strategy is proposed, in which the Boolean logic circuit is partitioned into sub-BDDs, each implemented by a corresponding FeFET-MUX. Experimental results based on open-source logic synthesis benchmarks demonstrate an average reduction of 79.9% in FeFET usage (Table 2) compared to conventional mapping techniques. This is particularly important because FeFET devices occupy considerably more area than conventional Metal-Oxide-Semiconductor (MOS) transistors. Reducing FeFET usage leads to substantial area savings at the circuit level. Moreover, the proposed algorithms effectively process large and complex designs, including circuits exceeding 30,000 BDD nodes, confirming their applicability to large-scale CiM logic implementations.
Cross-Layer Collaborative Resource Allocation in Maritime Wireless Communications: QoS-Aware Power Control and Knowledge-Enhanced Service Scheduling
ZHANG Zhilin, MAO Zhongyang, LU Faping, PAN Yaozong, LIU Xiguo, KANG Jiafang, YOU Yang, JIN Yin
Available online  , doi: 10.11999/JEIT250252
Abstract:
  Objective  Maritime wireless communication networks face significant challenges, including dynamic topology drift, large-scale channel fading, and cross-layer resource competition. These factors hinder the effectiveness of traditional single-layer resource allocation methods, which struggle to maintain the balance between high-quality communications and heterogeneous service demands under limited network resources. This results in degraded Quality of Service (QoS) and uneven service guarantees. To address these challenges, this study proposes a cross-layer collaborative resource allocation framework that achieves balanced enhancement of system throughput and QoS assurance through closed-loop optimization, integrating physical-layer power control with network-layer service scheduling. First, a cross-layer wireless network transmission model is established based on the coupling mechanism between physical-layer channel capacity and transport-layer TCP throughput. Second, a dual-threshold water-level adjustment mechanism, incorporating both Signal-to-Noise Ratio (SNR) and QoS metrics, is introduced into the classical water-filling framework, yielding a QoS-aware dual-threshold water-filling algorithm. This approach strategically trades controlled throughput loss for improved QoS of high-priority services. Furthermore, a conflict resolution strategy optimization filter with dual-channel feature decoupling is designed within a twin deep reinforcement learning framework to enable real-time, adaptive node-service dynamic matching. Simulation results demonstrate that the proposed framework improves average QoS scores by 9.51% and increases critical service completion by 1.3%, while maintaining system throughput degradation within 10%.  Methods  This study advances through three main components: theoretical modeling, algorithm design, and system implementation, forming a comprehensive technical system. First, leveraging the coupling relationship between physical-layer channel capacity and transport-layer Transmission Control Protocol (TCP) throughput, a cross-layer joint optimization model integrating power allocation and service scheduling is established. Through mathematical derivation, the model reveals the nonlinear mapping between wireless resources and service demands, unifying traditionally independent power control and service scheduling within a non-convex optimization structure, thus providing a theoretical foundation for algorithm development. Second, the proposed dynamic dual-threshold water-filling algorithm incorporates a dual-regulation mechanism based on SNR and QoS levels. A joint mapping function is designed to enable flexible, demand-driven power allocation, enhancing system adaptability. Finally, a twin deep reinforcement learning framework is constructed, which achieves independent modeling of node mobility patterns and service demand characteristics through a dual-channel feature decoupling mechanism. A dynamic adjustment mechanism is embedded within the strategy optimization filter, improving critical service allocation success rates while controlling system throughput loss. This approach strengthens system resilience to the dynamic, complex maritime environment.  Results and Discussions  Comparative ablation experiments demonstrate that the dynamic dual-threshold water-filling algorithm within the proposed framework achieves a 9.51% improvement in QoS score relative to conventional water-filling methods. Furthermore, the Domain Knowledge-Enhanced Siamese DRL (DKES-DRL) method exceeds the Siamese DRL approach by 3.25% (Fig. 6), albeit at the expense of a 9.3% reduction in the system’s maximum throughput (Fig. 7). The average number of completed transactions exceeds that achieved by the traditional water-filling algorithm by 1.3% (Fig. 8, Fig. 9). In addition, analysis of the effect of node density on system performance reveals that lower node density corresponds to a higher average QoS score (Fig. 10), indicating that the proposed framework maintains service quality more effectively under sparse network conditions.  Conclusions  To address the complex challenges of dynamic topology drift, multi-scale channel fading, and cross-layer resource contention in maritime wireless communication networks, this paper proposes a cross-layer collaborative joint resource allocation framework. By incorporating a closed-loop cross-layer optimization mechanism spanning the physical and network layers, the framework mitigates the imbalance between system throughput and QoS assurance that constrains traditional single-layer optimization approaches. The primary innovations of this work are reflected in three aspects: (1) Cross-layer modeling is applied to overcome the limitations of conventional hierarchical optimization, establishing a theoretical foundation for integrated power control and service scheduling. (2) A dual-dimensional water-level adjustment mechanism is proposed, extending the classical water-filling algorithm to accommodate QoS-driven resource allocation. (3) A knowledge-enhanced intelligent decision-making system is developed by integrating model-driven and data-driven methodologies within a deep reinforcement learning framework. Simulation results confirm that the proposed framework delivers robust performance in dynamic maritime channel conditions and heterogeneous traffic scenarios, demonstrating particular suitability for maritime emergency communication environments with stringent QoS requirements. Future research will focus on resolving engineering challenges associated with the practical deployment of the proposed framework.
Constructing Two Classes of Maximum Distance Separable Entanglement-Assisted Quantum Error-Correcting Codes by Using Twisted Generalized Reed-Solomon Codes
PAN Xin, GAO Jian
Available online  , doi: 10.11999/JEIT250258
Abstract:
  Objective  With the rapid development of quantum communication and computation, efficient error-correction technologies are critical for ensuring the reliability of quantum systems. Maximal Distance Separable (MDS) Entanglement-Assisted Quantum Error-Correcting Codes (EAQECCs) with flexible code lengths and larger minimum distances offer significant advantages in enhancing quantum system robustness. However, classical quantum codes face limitations in parameter flexibility and minimum distance constraints. This study addresses these challenges by leveraging Twisted Generalized Reed-Solomon (TGRS) codes to construct novel MDS EAQECCs, aiming to improve performance in complex quantum communication scenarios. In this paper, TGRS codes are innovatively applied to construct MDS EAQECCs. Different from polynomial construction methods, we determine the dimension of the Hermitian Hull through special matrix rank analysis, and successfully construct \begin{document}$ q $\end{document}-ary MDS EAQECCs with a minimum distance exceeding\begin{document}$ q+1 $\end{document}. Two construction schemes with flexible code lengths are proposed. Notably, the minimum distances of the constructed \begin{document}$ q $\end{document}-ary codes all exceed \begin{document}$ q+1 $\end{document}. Table 2 systematically summarizes the known MDS EAQECCs with a length less than \begin{document}$ {q}^{2} $\end{document} and a minimum distance exceeding \begin{document}$ q+1 $\end{document}. The novelty of the construction schemes in this paper is highlighted through parameter comparison. Reasoning shows that our schemes achieve flexible adjustment of the code length while maintaining the advantage of the minimum distance.  Methods  The proposed approach integrates TGRS codes with algebraic techniques to design MDS EAQECCs. Two families of MDS EAQECCs are constructed by using TGRS codes over finite fields. The key steps are as follows: (1) TGRS Code Construction: Utilizing twisted polynomials and hook parameters to generate TGRS codes with adjustable dimensions and minimum distances. (2) Hermitian Hull Analysis: Applying matrix rank analysis to determine the dimensions of the Hermitian Hull of the constructed codes, which is crucial for satisfying the Singleton bound in quantum codes. (3) Twisted Operation Optimization: Employing twisted operations to transform the constructed MDS EAQECCs into ME-MDS EAQECCs.  Results and Discussions  This paper constructs two families of MDS EAQECCs using TGRS codes and gives certain twisted conditions under which the codes are ME-MDS EAQECCs. Compared with other known codes, these new codes have more flexible parameters and significant advantages in terms of code length, minimum distance, and maximum entanglement state. This paper constructs \begin{document}$ q $\end{document}-ary EAQECCs with \begin{document}$ [[i(q-1), $\end{document}\begin{document}$ i(q-1)-2j-q-t+2v+2,q+j+1;q-t+2v+2]{]}_{q} $\end{document} when \begin{document}$ i $\end{document} is odd and \begin{document}$ [[i(q-1),i(q-1)- 2j-q-t+ $\end{document}\begin{document}$ 2v+3,q+j+1;q-t+2v+3]{]}_{q} $\end{document} when \begin{document}$ i $\end{document} is even. Based on Theorem 1 and Theorem 2, several types of MDS EAQECCs are obtained, and their parameters are listed in Table 1.Theorems 3-6 give the existence conditions of specific \begin{document}$ q $\end{document}-ary EAQECCs under different parameter settings. Furthermore, this paper upgrades the MDS EAQECCs to ME-MDS EAQECCs with \begin{document}$ \left[\right[i(q-1),i(q-1)-q-j,q+j+1;q+j]{]}_{q} $\end{document} by a twisted operation. Meanwhile, this paper constructs \begin{document}$ q $\end{document}-ary EAQECCs with \begin{document}$ \left[\right[i(q+1),(i-1)(q+1)-s,q+2;q-s+1]{]}_{q} $\end{document}. Moreover, this paper upgrades the MDS EAQECCs to ME-MDS EAQECCs with \begin{document}$ [[i(q+1),(i-1) $\end{document}\begin{document}$ (q+1),q+2;q+1]{]}_{q} $\end{document} by a twisted operation. Theorem 7 gives the dimension of the Hermitian Hull of this RS code after the twisted operation on its generating matrix. Similar to the first construction, a new twisted operation is applied, which upgrades the MDS EAQECCs to ME-MDS EAQECCs with specific parameters. Theorems 8, 9 give the existence conditions of specific \begin{document}$ q $\end{document}-ary EAQECCs under different parameter settings based on this construction method.  Conclusions  Two families of MDS EAQECCs are constructed using TGRS codes, and the parameters of known MDS EAQECCs are systematically summarized. Comparative analysis reveals that the EAQECCs proposed in this paper offer the following advantages: Compared with the known \begin{document}$ q $\end{document}-ary MDS EAQECCs (Table 2), the parameters of the MDS EAQECCs constructed here are new and have not been covered by previous studies; The codes not only enable flexible code-length adjustments but also achieve a minimum distance that significantly exceeds traditional theoretical bounds; Under specific twisted conditions, the constructed MDS EAQECCs are upgraded to ME-MDS EAQECCs. By introducing the twisted operation on RS codes, more flexible parameter combinations are obtained. This provides greater flexibility in the design of quantum error-correcting codes, enabling better adaptation to different quantum communication requirements. This improvement further optimizes the performance of quantum error-correcting codes, enhancing the entanglement-assisted error-correcting ability and improving the overall efficiency of quantum systems. These research results indicate that TGRS codes are important theoretical tools for constructing high-performance EAQECCs with excellent parameters. They play a pivotal part in advancing the development of quantum communication technology. Moreover, they offer a firm theoretical underpinning for the practical implementation of quantum error-correcting codes. Future research can focus on further exploring the potential of TGRS codes in constructing more advanced quantum error-correcting codes and expanding their applications in different quantum communication scenarios.
A 64 Gb/s Single-Ended Simultaneous Bi-Directional Transceiver for Die-to-Die Interfaces
WANG Zhifei, HUANG Zhiwen, YE Tianchen, YE Bingyi, LI Fangzhu, WANG Wei, YU Dunshan, GAI Weixin
Available online  , doi: 10.11999/JEIT250506
Abstract:
  Objective  Chiplet technology, which packages multiple dies with different functions and processes together, offers a cost-effective way for fabricating high-performance chips. For die-to-die data transmission, the edge density, Bit Error Rate (BER), and power consumption of the interface are crucial to the chip’s key performance metrics, such as computing power and throughput. Simultaneous Bi-Directional (SBD) signaling is an effective way to double the edge density by transmitting and receiving data on the same channel. However, with higher data rate and smaller channel pitch, channel reflection and crosstalk bring severe challenges to the design of interface circuits. This paper presents a single-ended SBD transceiver with echo and crosstalk cancellation to achieve a larger edge density and a lower BER.  Methods  The transceiver improves the per-wire data rate by utilizing the SBD signaling and denser shield-less channels. However, as both ends of the channel transmit data simultaneously, bi-directional signal coupling arises. Signal coupling, echo from impedance mismatch, and crosstalk from adjacent channels degrade the received data’s Signal-to-Noise Ratio (SNR). To decouple the bi-directional signal and cancel the echo and Near-End Crosstalk (NEXT), this paper proposes a Dynamic Voltage ThresHold generator (D-VTH). It generates the slicer’s threshold voltage according to the interfering signals needing to be subtracted. To cancel the Far-End Crosstalk (FEXT), a channel with the same capacitive and inductive coupling is designed by adjusting its width and space. FEXT is the subtraction of these two kinds of coupling, so it is canceled as expected. The source-synchronize architecture enhances the clock-data tracking performance, thereby reducing the clock-to-data jitter to improve the link’s noise margin. The synchronous clock distribution circuit includes a standing wave-based half-rate clock (CK2) distribution and a delay-controlled reset chain. The end of the CK2’s Transmission Line (TL) is terminated by a dedicated inductor, making the reflected wave have a proper amplitude and phase relative to the incident wave; thus, a standing wave can be formed, and CK2 synchronization is realized. To ensure the divided clocks (up to 1/32-rate) are synchronous, the dividers’ reset signals must be released at the same time or skewed with an integer multiple of 32 Unit Interval (UI). A reset chain is proposed to release the reset signals with controlled delay. The delay increases by 2 UI at each lane and is compensated by different stages of DFFs. After the CK2 and the divided clocks’ synchronization, the transmitter’s output and NEXT cancellation synchronization are achieved.  Results and Discussions  The test chip, including the proposed transceiver and the 3 mm on-chip channel, is fabricated in 28 nm CMOS. The shield-less data channels are routed in the M9 layer, with a channel pitch of 6.1 um. An electromagnetic field solver calculates the channel’s frequency response and the equivalent lumped model. The equivalent \begin{document}$ {C}_{\mathrm{m}}/{C}_{\mathrm{s}} $\end{document} is 0.28, and the \begin{document}$ {L}_{\mathrm{m}}/{L}_{\mathrm{s}} $\end{document} is 0.26, making FEXT 24 dB smaller than the Insertion Loss (IL) at the Nyquist frequency. In contrast, NEXT and Return Loss (RL) are much larger; they are just 7.3 dB and 8.3 dB smaller than the IL at the Nyquist frequency, respectively (Fig.12). The D-VTH filter’s coefficients are obtained from the Sign-Sign Least Mean Square (SS-LMS) adaptation algorithm, and the data is received correctly using the adapted coefficients. The bi-directional decoupling coefficient is the largest because the local transmitter’s output is the strongest compared to the echo and crosstalk. The echo cancellation coefficient is the smallest because it has to undergo additional insertion loss in the channel (Fig.13). The simulated clock-to-data tracking performance shows the transceiver’s robustness against power supply noise (Fig.15). The standing wave distribution’s simulation results show its amplitude is double that of the conventional traveling wave because of the superposition of incident and reflected waves. A slight skew of 0.6 ps is observed, caused by the residual traveling wave due to the TL’s loss (Fig.18). The measured internal eye diagrams and bathtub curves at 64 Gb/s shows the eye-opening is 0.68 UI/80 mV at 10-9 BER and 0.64 UI/77 mV at 10-12 BER, with both crosstalk cancellation and echo cancellation enabled (Fig.21). In addition, the measured BER at the optimal sampling point is less than 10-16 with all the lanes counting bit errors. The Crosstalk-Induced Jitter (CIJ) is reduced from 0.58 UI to 0.06 UI after crosstalk cancellation is enabled, representing a reduction ratio of 89.6% (Table 1). The measured power efficiency is 1.21 pJ/b, and the simulated power breakdown shows that the transmitter, receiver, D-VTH, and clock distribution account for 40%, 23%, 34%, and 3%, respectively (Fig.22). This work achieves the best per-wire data rate and per-layer edge density compared with previous works (Table 2).  Conclusions  This paper utilizes SBD signaling and denser shield-less channels to achieve a per-wire data rate of 64 Gb/s and a per-layer edge density of 10.5 Tb/s/mm. The proposed echo and crosstalk cancellation circuit ensures an extremely low BER of less than 10-16. It provides new insights for increasing the edge density of die-to-die interfaces.
A CNN-LSTM Fusion-Based Method for Detecting Hardware Trojan Bypasses
ZHOU Kang, HOU Bo, WANG Liwei, LEI Dengyun, LUO Yongzhen, HUANG Zhongkai
Available online  , doi: 10.11999/JEIT250241
Abstract:
  Objective  The globalization of Integrated Circuit (IC) design and increasing reliance on outsourcing have heightened the vulnerability of hardware supply chains to malicious modifications, such as hardware Trojans. These covert circuits may remain dormant until triggered, causing data leakage, system performance degradation, or physical damage. Detecting such threats is essential for safeguarding the security and reliability of semiconductor devices. Traditional side-channel detection methods based on power consumption or timing analysis often depend on manually designed features, which are sensitive to noise and lack generalizability across hardware platforms. Therefore, these techniques suffer from low detection accuracy and high false-positive rates under practical conditions. To address these limitations, this study proposes a deep learning-based side-channel detection method. By leveraging the ability of neural networks to automatically extract features from raw power signals, the proposed approach targets the identification of subtle anomalies associated with Trojan activation. The aim is to develop a robust, scalable detection solution applicable to real-world industrial scenarios.  Methods  The proposed detection framework integrates a hybrid deep learning architecture that combines a One-Dimensional Convolutional Neural Network (1D-CNN) with a Long Short-Term Memory (LSTM) network (Fig. 5). This architecture is designed to exploit the complementary advantages of CNNs and LSTMs for feature extraction. Specifically, the 1D-CNN component captures local spatial correlations within transient power traces, which are critical for detecting short-term fluctuations indicative of Trojan activity. The convolutional filters automatically learn edges, patterns, and shifts in signal magnitude, thereby reducing reliance on manual feature engineering. In parallel, the LSTM component is employed to model long-range temporal dependencies in the power signal sequence. Compared with conventional Recurrent Neural Networks (RNNs), LSTMs incorporate memory gates that enable selective retention or dismissal of past information, making them suitable for analyzing time-series data such as power traces. This enhances the framework’s ability to detect sequential patterns and context-dependent anomalies that may emerge over extended periods. The dataset comprises real-world transient power traces collected from fabricated Application-Specific Integrated Circuit (ASIC) chips, including both Trojan-free and Trojan-infected samples. Each power trace contains 125,000 sample points, capturing high-resolution dynamic power consumption under controlled activation scenarios. To reduce computational complexity and focus the model on signal segments most relevant to Trojan detection, a preprocessing step is applied. Specifically, windows of power data are extracted around the rising edges of the clock signal, where circuit state transitions are most likely to reveal Trojan-induced anomalies. This reduces the data dimensionality to 22,485 points per sample. To enhance the robustness of the model and mitigate overfitting, Gaussian noise is injected into the training data for data augmentation. This simulates realistic environmental and sensor-related noise conditions. The final dataset is divided into training, validation, and test sets in a 50%-25%-25% ratio, with balanced distributions of Trojan-free and Trojan-infected samples.  Results and Discussions  The experimental evaluation confirms the effectiveness of the proposed hybrid deep learning model for accurate and efficient hardware Trojan detection. By applying preprocessing to reduce input dimensionality, the training time is reduced by approximately an order of magnitude, substantially lowering computational requirements without compromising detection accuracy. The final model, trained using the RMSProp optimizer with a learning rate of 0.0005 and a batch size of 64, achieves a detection accuracy of 99.6% for the four-class classification task (Table 1). Analysis of the confusion matrix (Fig. 6) demonstrates that the model reliably distinguishes Trojan-free samples from three different types of Trojan-infected samples. Precision and recall rates exceed 99% across all classes, with minimal misclassification. The introduction of Gaussian noise during training further enhances the model’s generalization ability, ensuring stable performance on previously unseen test data. The macro-average F1-score reaches 99.6%, indicating balanced detection performance for all classes. In comparative evaluations with existing state-of-the-art methods, including Domain-Adversarial Neural Networks (DANN), Principal Component Analysis combined with LSTM (PCA-LSTM), and Siamese networks (Table 3), the proposed 1D-CNN-LSTM model consistently achieves superior accuracy and robustness. A key advantage is the model’s ability to process real-world measured power traces, rather than relying solely on simulated data. These results highlight the significance of combining spatial and temporal modeling for side-channel analysis and demonstrate the potential of deep learning techniques for hardware security applications. Nevertheless, the current experiments are conducted under ideal laboratory conditions with controlled data acquisition. Practical deployments are likely to encounter additional challenges, such as environmental fluctuations, measurement noise, and potential adversarial interference with power signals. Addressing these limitations remains an open research problem.  Conclusions  This paper proposes a deep learning-based hardware Trojan side-channel detection method that integrates a 1D-CNN-LSTM hybrid model to automatically extract and analyze features from power consumption signals. The method achieves substantial improvements in both detection efficiency and accuracy, supporting the feasibility of deep learning for hardware security applications. Future research will focus on addressing real-world challenges, including sensor noise, environmental variability, and adversarial attacks, as well as exploring semi-supervised or unsupervised learning to reduce reliance on labeled data. These findings provide a promising basis for enhancing the security and reliability of IC designs against hardware Trojan threats.
Low Switching Loss Double Trench SiC MOSFET with Integrated JFET Continuity Diode
GAO Sheng, ZHANG Xianfeng, CHEN Qiurui, CHEN Weizhong, ZHANG Hongsheng
Available online  , doi: 10.11999/JEIT250237
Abstract:
  Objective  Silicon Carbide Metal Oxide Semiconductor Field Effect Transistors (SiC MOSFETs) are considered ideal power devices for power systems due to their ultra-low on-resistance and excellent switching characteristics. However, Conventional SiC MOSFETs (CON-MOS) present considerable limitations in reverse current applications. These limitations stem primarily from their reliance on the body diode during reverse conduction, which exhibits a high reverse conduction voltage, significant reverse recovery loss, and is prone to bipolar degradation during long-term operation, adversely affecting power system stability. Furthermore, CON-MOS devices in high-frequency switching circuits suffer from substantial switching losses, reducing overall circuit efficiency. A widely adopted solution is to connect an external Schottky Barrier Diode (SBD) in parallel to enhance reverse current continuity. However, this approach increases device size and parasitic capacitance. Moreover, Schottky contacts are susceptible to large reverse leakage currents at elevated temperatures. Although SiC MOSFETs with integrated SBDs mitigate issues caused by external parallel SBDs, they still exhibit degraded blocking characteristics and thermal stability. SiC MOSFETs incorporating integrated MOS channel diodes have also been proposed to improve reverse conduction performance. Nonetheless, these devices raise reliability concerns due to increased process complexity and the presence of an ultra-thin (10 nm) oxide layer. Alternative industry structures employing polysilicon heterojunctions with 4H-SiC epitaxial layers aim to enhance reverse current continuity in SiC MOSFETs. However, these structures exhibit high reverse leakage currents and lack avalanche capability, primarily because the heterojunction barrier is insufficient to sustain the full blocking voltage. Devices integrating channel accumulation diodes have demonstrated lower reverse conduction voltage and reduced reverse recovery charge. Nevertheless, the barrier height in these designs is highly sensitive to oxide layer thickness, imposing stricter process control requirements. To address these challenges, this paper proposes an Integrated Junction Field Effect Transistor (JFET) SiC MOSFET (IJ-MOS) structure. The IJ-MOS effectively reduces reverse recovery loss, eliminates bipolar degradation, and significantly improves performance and reliability in reverse continuous current applications.  Methods  Technology Computer-Aided Design (TCAD) simulations are conducted to evaluate the performance of the proposed and conventional structures. Several critical models are included in the simulation process, such as mobility saturation under high electric fields, Auger recombination, Okuto–Crowell impact ionization, bandgap narrowing, and incomplete ionization. Furthermore, the effects of traps and fixed charges at the SiC/SiO2 interface are also considered. This study proposes an IJ-MOS structure based on the physical mechanism of energy band bending within the space charge region of the PN junction. Specifically, the IJ-MOS blocks the intermediate channel region through PN junctions formed between the Current Spreading Layer (CSL) and the P-body and P-shield layers, respectively. The blocking mechanism relies on the PN junction inducing conduction band bending within the CSL layer, thereby raising the conduction band energy and forming a barrier region. During reverse conduction, the integrated JFET provides a unipolar, low-barrier reverse conduction path, which mitigates bipolar degradation and significantly reduces reverse recovery charge. This improves device performance and reliability under reverse current conditions. Furthermore, the IJ-MOS reduces gate-drain coupling by separating the polysilicon gate and extended oxide structure, while optimising the internal electric field distribution. These design features enhance the device’s blocking voltage capability, increasing the potential of IJ-MOS for high-voltage applications.  Results and Discussions  Simulation results indicate that, compared to CON-MOS, the proposed IJ-MOS structure significantly reduces the reverse conduction voltage from 2.92 V in CON-MOS to 1.83 V (Fig. 3). The reverse recovery charge is reduced by 43.6%, and the peak reverse recovery current decreases by 31.6%, while maintaining comparable forward conduction characteristics (Fig. 7). Furthermore, due to the split gate and extended oxide structure, the IJ-MOS exhibits a lower gate-drain capacitance, effectively reducing the coupling between the gate and drain. The extended oxide layer also improves the internal electric field distribution, leading to an increase in breakdown voltage and a 60% improvement in the Baliga Figure of Merit (BFOM) (Table 2). Benefiting from the lower gate-drain capacitance, the total switching loss of IJ-MOS is reduced by 24.2% compared to CON-MOS (Fig. 8).  Conclusions  This paper proposes a novel SiC MOSFET structure evaluated through TCAD simulation. The proposed IJ-MOS reduces reverse conduction voltage and significantly lowers reverse recovery charge, thereby enhancing reverse conduction performance. Since the barrier region of the integrated JFET is lower than that of the PN junction, the JFET conducts prior to the body diode, which effectively suppresses bipolar conduction of the body diode and avoids bipolar degradation. The primary carriers in the JFET are electrons rather than both electrons and holes, meaning only electrons must be removed during the reverse recovery process, reducing reverse recovery charge. Additionally, the split gate and extended oxide structure reduce gate-drain coupling, which decreases gate-drain capacitance, switching time, and overall switching losses. These advantages make the IJ-MOS a promising candidate for high-performance power electronics applications.
SR-FDN: A Frequency-Domain Diffusion Network for Image Detail Restoration in Super-Resolution
LI Xiumei, DING Linlin, SUN Junmei, BAI Huang
Available online  , doi: 10.11999/JEIT250224
Abstract:
  Objective  Image Super-Resolution (SR) is a critical computer vision task aimed at reconstructing High-Resolution (HR) images from Low-Resolution (LR) inputs, with broad applications in fields such as medical imaging and satellite imaging. Recently, diffusion-based SR methods have attracted significant attention due to their generative capability and strong performance in restoring fine image details. Existing diffusion model-based SR approaches have demonstrated potential in recovering textures and structures, with some methods focusing on spatial domain features and others utilizing frequency domain information. Spatial domain features aid in reconstructing overall structural information, whereas frequency domain decomposition separates images into amplitude and phase components across frequencies. High-frequency components capture details, textures, and edges, whereas low-frequency components describe smooth structures. Compared to purely spatial modeling, frequency domain features improve the aggregation of dispersed high-frequency information, enhancing the representation of image textures and details. However, current frequency domain SR methods still show limitations in restoring high-frequency details, with blurring or distortion persisting in some scenarios. To address these challenges, this study proposes SR-FDN, a SR reconstruction network based on a frequency-domain diffusion model.  Methods  SR-FDN leverages the distribution mapping capability of diffusion models to improve image reconstruction. The proposed network integrates spatial and frequency domain features to enhance high-frequency detail restoration. Two constraints guide the model design: (1) The network must generate plausible HR images conditioned solely on LR inputs, which serve as the primary source of structural information, ensuring high-fidelity reconstruction. (2) The model should balance structural reconstruction with enhanced detail restoration. To achieve this, a dual-branch frequency domain attention mechanism is introduced. A portion of the features undergoes Fourier transform for frequency domain processing, where high-frequency information is emphasized through self-attention. The remaining features adjust frequency domain weights before being combined with spatial domain representations. Skip connections in the U-Net architecture preserve LR structural information while enhancing frequency domain details, improving both structural and textural reconstruction. Wavelet downsampling replaces conventional convolutional downsampling within the U-Net noise predictor, reducing spatial resolution while retaining more detailed information. In addition, a Fourier frequency domain loss function constrains amplitude and phase components of the reconstructed image, further enhancing high-frequency detail recovery. To guide the generative process, additional image priors are incorporated, enabling the diffusion model to restore textures consistent with semantic category features.  Results and Discussions  The results of SR-FDN on face datasets and general datasets for 4× and 8× SR (Table 1, Table 2, Table 3) demonstrate that the proposed method achieves strong performance across objective evaluation metrics. These results indicate that SR-FDN can effectively restore image detail information while better preserving structural and textural features. A comparison of iteration counts between SR-FDN and two diffusion-based methods (Fig. 2) shows that SR-FDN can reconstruct higher-quality images with fewer iterations. Despite the reduced number of iterations, SR-FDN maintains high-fidelity reconstruction, reflecting its ability to lower computational overhead without compromising image quality. To further verify the effectiveness of the proposed SR-FDN, visual comparisons on the FFHQ dataset (Fig. 3) and the DIV2K dataset (Fig. 4) are presented. The results show that SR-FDN offers clearer and more detailed image reconstruction, particularly in high-frequency regions such as facial features and hair textures. Ablation experiments (Table 5) and feature visualization results (Fig. 5) are also provided. These results confirm that the proposed dual-branch frequency domain design and the Fourier domain loss function significantly contribute to improved restoration of fine details.  Conclusions  This study proposes SR-FDN, a diffusion-based SR reconstruction model that integrates frequency domain information to enhance detail restoration. The SR-FDN model incorporates a dual-branch frequency domain attention mechanism, which adaptively reinforces high-frequency components, effectively addressing the limitations of conventional methods in recovering edge structures and texture details. In addition, SR-FDN employs wavelet downsampling to preserve informative features while reducing spatial resolution, and introduces a frequency domain loss function that constrains amplitude and phase information, enabling more effective fusion of frequency and spatial domain features. This design substantially enhances the model’s ability to recover high-frequency details. Extensive experiments on benchmark datasets demonstrate that SR-FDN reconstructs images with superior quality and richer details, exhibiting clear advantages in both qualitative and quantitative evaluations.
Acoustic DOA Estimation in Underwater Environments by Integrating Spatial Domain Wiener Filtering and Convolutional Neural Networks
XING Chuanxi, HUANG Tinglong, TAN Guangzhi, LI Weiqiang
Available online  , doi: 10.11999/JEIT250141
Abstract:
  Objective  Human activities in shallow-sea regions have increased with the progressive development and utilization of marine environments. However, interference from noise sources such as water currents, vessels, and marine organisms substantially reduces the Signal-to-Noise Ratio (SNR) of received signals, leading to performance degradation in conventional Direction of Arrival (DOA) estimation algorithms. These limitations affect accurate target localization and detection, which are essential for underwater acoustic applications. This paper proposes a method that combines Wiener filtering with deep learning, specifically an improved Convolutional Neural Network (CNN), to enhance the robustness and accuracy of DOA estimation in shallow-sea environments by addressing challenges such as noise interference and grid mismatch.  Methods  The proposed algorithm integrates Wiener filtering and deep learning to optimize DOA estimation. First, Wiener filtering is applied to the covariance matrix of received signals to reduce noise. Considering the complexity of underwater acoustic environments, this preprocessing step is essential for preserving spatial signal features and enhancing the effectiveness of network training. Second, an improved CNN architecture is designed to perform both classification and regression tasks, enabling accurate DOA angle estimation while correcting off-grid errors. A fractional label correction strategy is introduced to mitigate grid mismatch, allowing prediction errors to be compensated and estimated angles to align more closely with true values. The network is trained using a large-scale dataset of simulated underwater acoustic signals, and its feasibility is verified through real-world sea trial data.  Results and Discussions  The proposed method is validated using both simulation data and sea trial data, demonstrating superior performance under low SNR conditions. Simulation results show that incorporating Wiener filtering improves overall network performance by 25.20% (Fig. 4). The proposed algorithm achieves lower estimation errors in low SNR environments (Fig. 4, Fig. 5). Compared with the MUSIC and ESPRIT algorithms, the Root Mean Square Error (RMSE) is reduced by 93.42% and 92.14%, respectively, at an SNR of –20 dB. Moreover, as the SNR increases, the algorithm exhibits a gradual reduction in estimation error (Fig. 6). In sea trial experiments, the algorithm accurately identifies target azimuths and achieves lower estimation errors compared to conventional methods (Fig. 9, Table 1). These results indicate that the proposed method meets the requirements for robustness and accuracy in practical underwater acoustic applications.  Conclusions  This paper proposes a method for direction estimation in shallow-water environments that integrates Wiener filtering with a deep learning-based CNN to address the performance degradation caused by low SNR conditions in conventional algorithms. The combination of noise reduction preprocessing with CNN model training enhances both network efficiency and estimation accuracy. Simulation and sea trial experiments yield the following results: (1) Overall network performance is improved by 25.20% through Wiener filtering optimization; (2) The improved binary cross-entropy loss function enables the network to perform both classification and regression tasks for DOA estimation; (3) The algorithm demonstrates higher robustness to noise interference and grid mismatch, particularly under low SNR and limited snapshot conditions, making it suitable for practical deployment. However, the network is trained using datasets with known source numbers. Future research will focus on network designs that accommodate more diverse source conditions, including blind source scenarios and adaptive noise reduction models.
Task Segmentation and Computing Resource Allocation Method Driven by Path Prediction in Internet of Vehicles
HUO Ru, LV Kecheng, HUANG Tao
Available online  , doi: 10.11999/JEIT250135
Abstract:
  Objective  The rapid development of 5G and the Internet of Vehicles (IoV) has accelerated the demand for intelligent services and large-scale data processing. However, limited computational capacity at vehicle terminals and frequent topology changes due to high mobility reduce transmission efficiency and reliability. While current task offloading and resource allocation methods meet basic processing needs, they fall short in supporting fine-grained task segmentation and in adapting to the dynamic characteristics of vehicular networks. To address these limitations, this study proposes an optimization approach for task segmentation and computing resource allocation based on Reinforcement Learning (RL). By combining the time-series modeling capabilities of advanced Deep Learning (DL) models with the resource scheduling flexibility of a Multi-Agent Deep Reinforcement Learning (MADRL) algorithm, this method enables intelligent, prediction-driven task segmentation and allocation. It supports decoupled processing and multi-objective optimization of latency and resource use. The proposed framework enhances offloading efficiency and resource utilization in Vehicular Edge Computing (VEC) environments and provides theoretical support for real-time resource management in highly dynamic vehicular networks.  Methods  The proposed method integrates vehicle trajectory prediction, dynamic task segmentation, and adaptive resource scheduling. For trajectory prediction, historical movement data are preprocessed through trajectory mapping. A Transformer-based model captures long-range temporal dependencies to achieve accurate path prediction. Based on predicted routes and estimated dwell times in different edge zones, a segmentation strategy divides tasks into subtasks to distribute load evenly across edge servers. The resource allocation challenge under this segmentation strategy is modeled as a Markov Decision Process (MDP), with each edge server acting as an agent that interacts with its environment. The state space includes task sizes, server loads, and available resources, while the action space defines the scheduling decisions each agent may select. The reward function is designed to minimize delay and improve efficiency, promoting cooperative behavior across agents. Through continual environmental feedback, the MADRL framework adaptively adjusts resource allocation in response to real-time task demands and server workloads.  Results and Discussions  The Transformer-based path prediction model outperforms baseline models (Table 1), achieving a Mean Squared Error (MSE) of 0.0116, substantially lower than that of LSTM (0.0246) and BiLSTM (0.0323). To identify the optimal prediction horizon, the model is evaluated across different prediction steps (Fig. 5). The reward increases from 2 to 5 steps but declines beyond 5 due to increased transmission delays caused by frequent edge node handovers. Thus, 5-step prediction is adopted as the optimal setting. The proposed algorithm is assessed against three baseline strategies: binary offloading, random segmentation, and alternating convex search. Under a 5-step horizon and a five-task vehicle scenario (Fig. 6), the proposed method reduces average task latency by 54.1% compared with random segmentation and by 16.5% compared with alternating convex search. By tuning the weight coefficient (Fig. 7), the method demonstrates balanced performance across latency and resource efficiency objectives. Specifically, at a coefficient of 0.8, it achieves a convergence reward of 39.224, exceeding that of alternating convex search (35.872) by 9.4%. Scalability is evaluated under varying vehicle numbers (Fig. 8). The proposed method consistently yields the highest convergence reward across all cases. With four vehicles, it reaches a reward of 41.059, 7.3% higher than alternating convex search (38.277). With seven vehicles, it maintains a reward of 35.630, exceeding the baseline’s 32.524 by 9.6%. Overall, the proposed method exhibits superior latency reduction, higher resource efficiency, and better adaptability across different vehicle densities, validating its effectiveness and scalability for task offloading and resource allocation in VEC systems.  Conclusions  This study proposes a task segmentation and computing resource allocation method tailored for VEC scenarios. By exploiting the time-series modeling capacity of Transformer networks, the accuracy of vehicle trajectory prediction is enhanced. A dynamic task segmentation strategy is then applied to optimize workload distribution across Mobile Edge Computing (MEC) servers. The use of a MADRL-based cooperative resource allocation approach effectively reduces task latency while improving resource use. Simulation results confirm the method’s effectiveness in minimizing processing delay and enhancing system efficiency. Future research will explore the role of task prioritization in resource allocation and investigate strategies to sustain performance under increased vehicle counts and task complexity in large-scale deployments.
Multi-Objective Optimization of UAV-Assisted Wireless Power Transfer Mobile Edge Computing System
LIU Jianhua, LI Guohua, LIU Jiajia, TU Xiaoguang
Available online  , doi: 10.11999/JEIT250026
Abstract:
  Objective   With the rapid growth of data volume at the edge of the Internet of Things (IoT), traditional centralized computing architectures are inadequate to meet real-time processing requirements. Due to the limited computing and storage capabilities of IoT devices, data overflow frequently occurs when handling large volumes of information. Meanwhile, the reliance on battery power exacerbates energy shortages, further hindering continuous operation and data processing. Unmanned Aerial Vehicles (UAVs), with their flexible deployment capabilities, are increasingly being integrated into distributed computing environments to improve IoT data processing efficiency. However, the stochastic and dynamic nature of IoT user demands presents significant challenges for existing resource scheduling schemes, which struggle to effectively coordinate UAV deployments. Moreover, most studies focus on single-objective optimization, making it difficult to simultaneously balance energy harvesting, energy consumption, system latency, and data transmission rates under complex environmental conditions. To address these challenges, this paper proposes an optimization framework for a UAV-assisted, wireless-powered Mobile Edge Computing (MEC) system based on an improved Multi-Objective Deep Deterministic Policy Gradient (MODDPG) algorithm. The proposed method jointly optimizes task offloading, energy harvesting, flying energy consumption, and system latency, enhancing the overall performance and feasibility of IoT systems.  Methods   This paper designs a UAV-assisted, wireless-powered MEC system model, where the UAV follows a “fly-hover-communicate” protocol and operates in full-duplex mode during hovering to simultaneously perform data collection and energy transmission. The system model integrates a propulsion energy consumption model and a nonlinear energy harvesting model, formulating a multi-objective optimization problem based on task generation, energy state, and wireless channel conditions of IoT devices. To address this optimization problem, an improved MODDPG algorithm is proposed. A multi-dimensional reward function is constructed to maximize the data transmission rate and total harvested energy while minimizing system energy consumption and task offloading latency within a unified reinforcement learning framework. The agent continuously interacts with the dynamic environment to optimize the UAV’s flight trajectory, task offloading ratio, and energy transmission strategy. The proposed method is implemented and validated through extensive simulations conducted on the Python platform under various experimental settings.  Results and Discussions  To evaluate the performance of the proposed algorithm, comparative experiments are conducted against four other control strategies. The impact of different data collection coverage radii on the proposed strategy is analyzed (Fig. 3, Fig. 4). In terms of total data transmission rate, the \begin{document}${P_{\rm{MODDPG}}}$\end{document} strategy consistently ranks second to the maximum throughput strategy across all coverage settings, while significantly outperforming the other three strategies (Fig. 3(a)). Regarding total harvested energy, \begin{document}${P_{\rm{MODDPG}}}$\end{document} achieves the highest energy collection in all scenarios, with performance improving as the coverage radius increases (Fig. 3(b)), demonstrating its capability to enhance energy transmission coverage and system efficiency. In terms of system energy consumption, the \begin{document}${P_{{V_{\max }}}}$\end{document} strategy exhibits the highest energy usage, whereas \begin{document}${P_{\rm{MODDPG}}}$\end{document} effectively controls flight energy consumption, maintaining significantly lower energy usage compared to high-speed flying strategies (Fig. 3(c)). For average latency, \begin{document}${P_{\rm{MODDPG}}}$\end{document} achieves lower latency than traditional flight strategies and the TD3-based method in most cases, though it remains slightly higher than the maximum throughput strategy (Fig. 3(d)). In terms of service performance, \begin{document}${P_{\rm{MODDPG}}}$\end{document} demonstrates superior results across multiple metrics, including the number of uploading users, average harvested energy, average number of charged users per hovering instance, and the number of users experiencing data overflow, with particularly significant advantages in energy collection and reducing data overflow rates. (Fig. 4(a), Fig. 4(b), Fig. 4(c), Fig. 4(d)). In experiments with varying IoT device densities (Fig. 5, Fig. 6), \begin{document}${P_{\rm{MODDPG}}}$\end{document} exhibits strong environmental adaptability. As device density increases and resource competition intensifies, \begin{document}${P_{\rm{MODDPG}}}$\end{document} intelligently optimizes flight paths and resource scheduling, maintaining high data transmission rates and energy harvesting levels while effectively suppressing system latency and data overflow. Furthermore, simulation results under different task load scenarios (small and large tasks) (Fig. 7) indicate that \begin{document}${P_{\rm{MODDPG}}}$\end{document} can flexibly adjust its strategy based on task load variations, maximizing resource utilization. Even under heavy load conditions, the proposed algorithm maintains superior data rates, energy harvesting efficiency, and system stability, effectively coping with dynamic traffic variations. Notably, \begin{document}${P_{\rm{MODDPG}}}$\end{document} not only delivers outstanding optimization performance but also demonstrates significantly faster convergence during training compared to traditional methods such as Proximal Policy Optimization (PPO) and TD3. Overall,\begin{document}${P_{\rm{MODDPG}}}$\end{document} outperforms existing strategies in multi-objective optimization, system energy efficiency enhancement, task scheduling flexibility, and environmental adaptability.  Conclusions   This paper addresses the critical challenges of energy limitations, data overflow, and high latency in IoT environments by proposing an optimization framework for a UAV-assisted wireless-powered MEC system, which integrates UAV, MEC, and Wireless Power Transfer (WPT) technologies. In the proposed system, IoT devices’ data upload demands are updated in real time, with the UAV sequentially accessing each device based on the priority of its demands using a "fly-hover-communicate" protocol. An improved MODDPG algorithm, based on deep reinforcement learning, is introduced to jointly optimize data transmission rate, energy harvesting, energy consumption, and system latency. Training results demonstrate that, under all optimization scenarios, the proposed scheme consistently achieves the highest total harvested energy while simultaneously balancing the other three optimization objectives. Compared to other baseline strategies, the proposed method exhibits better adaptability by dynamically adjusting to environmental changes and varying device priorities, ultimately achieving coordinated multi-objective optimization and superior performance across different conditions. Future research will focus on extending this work to address multi-UAV cooperative task allocation, resource management, energy harvesting coordination, and multi-UAV trajectory planning.
Research on Fusion Localization Algorithm of Indoor UWB and IMU Assisted by GPR Error Calibration
MA Xinpeng, CHEN Yu, CUI Zhicheng, LI Xingguang, CUI Wei
Available online  , doi: 10.11999/JEIT241145
Abstract:
  Objective  Ultra-WideBand (UWB) ranging in confined indoor environments is prone to coarse ranging errors and Gaussian noise, which substantially degrade localization accuracy for both static and dynamic targets, even under Line-of-Sight (LOS) conditions. In addition, during indoor operations of wheeled vehicles, obstacles often obstruct the LOS between onboard UWB tags and anchors, resulting in Non-Line-of-Sight (NLOS) propagation. NLOS-induced interference results in abrupt fluctuations in range measurements, which severely compromise the estimation of the vehicle’s motion state.  Methods   To address these challenges, this study proposes a Gaussian Process Regression (GPR)-calibrated indoor UWB/Inertial Measurement Unit (IMU) fusion localization algorithm (GIU-EKF). The proposed approach offers two key features. First, a UWB ranging error model for the entire two-dimensional localization area is established by collecting UWB measurements at a limited set of reference points under LOS conditions. This model is applied to correct LOS-related ranging biases. Second, by leveraging the low short-term drift characteristics of inertial measurements, the algorithm fuses IMU and UWB data to mitigate NLOS ranging errors.  Results and Discussions  Under LOS conditions, a GPR-based error calibration model is constructed by sampling and analyzing UWB ranging deviations at known reference coordinates. This model captures the statistical association between two-dimensional spatial positions and the corresponding ranging errors. For any queried location, the model generates a set of probabilistic range estimates, with the final range value obtained by weighting nearby sample points according to their normalized likelihoods. This enables real-time suppression of LOS-related ranging errors. A threshold-based detection mechanism identifies NLOS conditions when the UWB range increment exceeds a predefined threshold. In NLOS scenarios, a subordinate Extended Kalman Filter (EKF) fuses UWB range data with short-term IMU measurements to compensate for NLOS-induced ranging errors during motion. The corrected range data are then incorporated into a primary EKF to update the vehicle’s motion state estimation. Experimental results demonstrate that the proposed GPR-based coarse error correction reduces localization errors by 64% and 58% for static and dynamic tags under LOS conditions, respectively. In three representative NLOS scenarios, the GIU-EKF algorithm maintains reliable motion state estimation for low-speed targets, achieving an average localization error of 7.5 cm. For tags moving at speeds between 0.2 m/s and 0.8 m/s, localization errors remain below 10 cm. The robustness of the proposed algorithm under extreme conditions is also validated. As shown in Section 4.3.2, the algorithm maintains stable velocity vector estimation even when the wheeled vehicle experiences alternating occlusions between single-anchor and multi-anchor configurations. Under low-speed conditions (2.2 cm/s), the localization error remains as low as 6.7 cm. Section 4.3.3 further verifies the algorithm’s convergence capability under large initial deviations. When subjected to initial heading errors between 5° and 50°, or a combined 1.5-meter position offset and 10° heading deviation, the proposed method consistently converges to the true position within a 2-meter travel distance.  Conclusions  This study presents a GPR-assisted indoor UWB/IMU fusion localization algorithm. By independently suppressing LOS and NLOS ranging errors from four UWB anchors, the proposed approach enhances localization accuracy in complex indoor environments. Under LOS conditions, a GPR-based correction mitigates coarse UWB ranging errors. In NLOS scenarios, short-term inertial estimates compensate for anomalous UWB measurements. A subordinate EKF adaptively balances observation uncertainties from the two sensing modalities, maintaining motion state observability when the tag moves slowly in confined spaces. This design avoids long-term drift accumulation, which is often observed in tightly coupled systems that treat IMU data as a strong prior, particularly when using low-cost inertial sensors. Experimental results demonstrate that the proposed algorithm achieves sub-10 cm localization accuracy under both LOS and NLOS conditions. During low-speed operations, the system maintains convergence of both velocity and position estimates. Furthermore, even with significant initial motion state biases, the algorithm consistently converges to the true trajectory. These findings indicate that the proposed method effectively meets the operational requirements of mobile robots in narrow indoor environments. However, practical application still requires further attention to two key aspects: efficient and autonomous collection of coordinate samples for GPR model training, and integration of real-time localization outputs with vehicle path planning and motion control systems.
Bit-configurable Physical Unclonable Function Circuit Based on Self-detection and Repair Method
XU Mengfan, ZHANG Yuejun, LIU Tianxiang, PAN Yu
Available online  , doi: 10.11999/JEIT250359
Abstract:
  Objective  The proliferation of Internet of Things (IoT) devices has intensified the need for robust, hardware-level security. Among hardware-based security primitives, Physical Unclonable Functions (PUFs) serve a critical role in lightweight authentication and dynamic key generation by leveraging inherent process variations to produce unique, unclonable responses. Achieving reliable PUF performance under environmental fluctuations—such as temperature and supply voltage variation—requires balancing sensitivity to process variations with environmental robustness. Conventional approaches, including circuit-level stabilization and architecture-level error correction, can improve reliability but often increase area, power, and test complexity. To overcome these drawbacks, recent work has explored voltage or bias perturbation for unstable response correction. However, entropy degradation during mode transitions in dual-mode PUFs remains a major concern, compromising both reliability and energy efficiency. This study proposes a bit-configurable bistable electric bridge-divider PUF that addresses these challenges by maintaining entropy independence between operational modes, reducing error correlation, and limiting repair and masking overhead. The proposed solution improves randomness, reliability, and energy efficiency, making it suitable for secure, cost-effective authentication in IoT edge devices operating under dynamic conditions.  Methods  Hardware overhead and testing complexity associated with conventional PUF stabilization techniques are reduced by introducing a bit-configurable bistable electric bridge-divider PUF architecture. Entropy generation is enhanced by amplifying process-induced variations through electric bridge imbalance and the exponential behavior of subthreshold current. A reconfigurable bit-cell is employed to enable seamless switching between electric bridge mode and voltage divider mode without additional layout cost; dual-mode operation is thus supported while preserving area efficiency. A voltage-skew-based self-detection and repair mechanism is integrated to dynamically identify and mitigate unstable responses, thereby improving reliability under varying environmental conditions. The PUF circuit is fully custom-designed and fabricated in the TSMC 28 nm CMOS process. Post-layout simulations confirm the robustness of the architecture, demonstrating effective self-repair capabilities and consistent performance under temperature and voltage fluctuations.  Results and Discussions  The proposed design is fabricated using the TSMC 28 nm CMOS process. The total layout area measures 3,283.3 μm2, and each PUF cell occupies 0.7888 μm2 (Fig. 11). Simulation waveforms of the self-detection, repair, and masking operations are presented in (Fig. 12). Inter-chip Hamming distance histograms and fitted curves for both electric bridge mode and voltage divider mode are shown in (Fig. 13). Autocorrelation results of the 40,960-bit output are illustrated in (Fig. 14). The randomness of the responses is evaluated using the NIST test suite provided by the U.S. National Institute of Standards and Technology, with the results summarized in (Table 1). The native Bit Error Rate (BER), measured before repair or masking, is analyzed under various temperature and supply voltage conditions (Fig. 15). By dynamically adjusting the voltage skew, precise control of the error correction rate is achieved, leading to a substantial reduction in BER across different environments (Fig. 16). A performance comparison with previously reported designs is provided in (Table 2). After applying the entropy source repair and masking mechanism, the BER converges to below 1.62 × 10-9, approaching the ideal “zero” BER.  Conclusions  A bit-configurable PUF architecture is proposed to address environmental variability and hardware constraints in IoT edge devices. A reconfigurable bit-cell is employed to support dynamic switching between electric bridge mode and voltage divider mode without incurring additional layout cost. Process-induced variations are amplified through bridge imbalance and the exponential behavior of subthreshold current, which enhances the randomness and uniqueness of the PUF responses. A voltage-skew-based self-detection and repair mechanism is integrated to identify and correct unstable responses, effectively reducing the BER under varying environmental conditions. The proposed design, fabricated using the TSMC 28 nm CMOS process, demonstrates high entropy, robustness, and low overhead in terms of area and power consumption. These characteristics make it suitable for secure and lightweight authentication and key generation in resource-constrained IoT systems.
OTFS Communication Link Construction and Three-Dimensional Constellation Encryption Design
MA Yingjie, LIU Yueheng, ZHAO Geng, ZHAO Mingjing, WANG Dan
Available online  , doi: 10.11999/JEIT250181
Abstract:
  Objective  With the advancement of Sixth-Generation (6G) communication technologies, the demand for highly reliable and secure transmission in high-speed mobile scenarios has grown significantly. Orthogonal Time Frequency Space (OTFS) modulation has emerged as a promising solution due to its superior transmission reliability in high-mobility environments. However, research on the security aspects of OTFS systems remains limited. To address this challenge, this study proposes a Three-Dimensional (3D) constellation encryption scheme for OTFS communication links. The objective is to enhance the security of OTFS systems while preserving communication performance by designing a novel 3D constellation structure and implementing physical-layer dynamic encryption based on a Coupled Map Lattice (CML) spatiotemporal chaotic system. This research provides a secure and efficient transmission strategy for future wireless communication systems.  Methods  The integration of 3D constellation encryption with OTFS modulation enhances the security of OTFS systems. First, a novel 3D constellation is designed, where 16 constellation points are uniformly distributed on the surface of a sphere centered at the origin. By optimizing the spatial distribution of constellation points, the minimum Euclidean Distance (MED) is increased and the average transmission power is reduced, thereby improving Bit Error Rate (BER) performance. Second, a dynamic encryption scheme for the 3D constellation over the OTFS link is proposed. In this scheme, pseudo-random sequences are generated using the CML system, quantized into 16-bit binary sequences, and employed to control both the selection of mapping rules and the rotational encryption of the 3D constellation. The encryption process operates as follows: first, the mapping rules are dynamically selected based on the quantized chaotic sequences; second, the rotation angles of the constellation points around the X, Y, and Z axes are determined by the same chaotic sequences. This dual encryption mechanism enhances the system’s security while maintaining reliable communication performance.  Results and Discussions  The performance of the proposed scheme is evaluated through extensive simulations. The results show that the optimized 3D constellation achieves a 22% improvement in MED compared with traditional 2D 16-quadrature amplitude modulation (16QAM). Compared with existing 3D constellations, the proposed design increases the MED by 5%, reduces average transmission power by 11%, and improves the Constellation Figure of Merit (CMF) by 10% (Table 2, Table 3). Superior communication performance is also demonstrated, with a BER improvement of approximately 0.6 dB over existing methods (Fig. 5, Fig. 6). In terms of security, the encryption scheme provides a key space of 10120, effectively resisting brute-force attacks. Key sensitivity tests confirm that even minor changes to the initial key result in completely different decrypted images (Fig. 7), verifying the robustness of the encryption mechanism. The encrypted images exhibit uniform histogram distributions (Fig. 8) and correlation coefficients near zero (Table 6), indicating strong resistance to statistical attacks. Additionally, the scheme maintains reliable performance under noisy conditions, with decrypted images preserving high quality even at low Signal-to-Noise Ratios (SNR) (Fig. 9).  Conclusions  This study combines OTFS communication links with 3D constellation encryption to realize physical-layer security for OTFS systems, enhancing both communication performance and transmission security. The proposed 3D constellation improves the MED by 5%, reduces average power by 11%, and increases the CMF by 10%, thereby enhancing BER performance. The dual dynamic encryption mechanism based on a CML system offers strong security. Simulation results show that the proposed scheme achieves a favorable balance between communication performance and security. BER performance improves by approximately 0.6 dB, while the large key space, high key sensitivity, strong resistance to statistical attacks, and good robustness collectively ensure enhanced security. The combination of OTFS modulation and 3D constellation encryption not only improves spectrum resource utilization but also strengthens channel confidentiality, enhancing resilience to interference and potential attacks. This scheme improves transmission security without compromising communication performance, offering broad application prospects in fields such as the Internet of Things, unmanned aerial vehicles, satellite communication, and ocean communication. The proposed approach is expected to provide more secure and reliable transmission solutions for future wireless communication systems.
Resilient Semantic Communication for Space-Air-Ground Networks
WANG Wenyuan, ZHOU Mingyu, WANG Chaowei, XU Jisong, ZHANG Yunze, PANG Mingliang, JIANG Fan, XU Lexi, ZHANG Zhi
Available online  , doi: 10.11999/JEIT250077
Abstract:
  Objective  Space-air-ground networks represent a fundamental evolution in wireless communication infrastructure by extending conventional terrestrial networks into airspace and outer space through the integration of satellites, unmanned aerial vehicles, and ground-based nodes. These networks have become a cornerstone technology for future sixth-generation wireless communications, providing wide-area coverage and flexible networking capabilities that overcome the limitations of geographically constrained terrestrial systems. However, image transmission over space-air-ground networks faces considerable challenges due to inherent bandwidth constraints, severe channel impairments, and highly dynamic propagation environments. Traditional Separate Source-Channel Coding (SSCC) methods, although theoretically sound, exhibit poor performance under these adverse conditions, particularly near information-theoretic limits. The rapid expansion of multimedia applications and service demands requires new approaches to information extraction and transmission that meet ultra-reliable low-latency communication requirements. Conventional image coding algorithms, such as JPEG and JPEG2000, experience significant performance degradation in satellite communication systems due to channel impairments. These limitations highlight the need for advanced semantic communication frameworks that move beyond conventional bit-level transmission and prioritize the preservation of semantic content integrity under complex wireless conditions.  Methods  A resilient semantic communication framework is proposed for image transmission in space-air-ground networks, based on enhanced information bottleneck theory. The key innovation is the development of an augmented rate-distortion function that jointly optimizes system capacity, pixel-level reconstruction fidelity, semantic feature preservation, and perceptual quality. The framework combines deep learning with information-theoretic principles to ensure reliable performance under varying channel conditions. A Gumbel-Softmax approach is integrated with variable rate networks to enable dynamic rate adaptation in response to fluctuating channel states. The system employs a weighted multi-asymmetric Gaussian distribution model to characterize the probability density functions of heterogeneous semantic features, providing accurate entropy estimation in the latent space. The architecture incorporates attention mechanisms and residual learning structures to enhance feature extraction and semantic preservation. A reconfigurable neural network architecture is introduced, featuring adaptive module selection driven by real-time signal-to-noise ratio assessments and service quality requirements. The encoder subsystem consists of semantic extractor networks that identify and isolate critical image features, joint source-channel encoders that perform integrated compression and error protection, and adaptive networks that generate binary mask vectors for intelligent symbol selection. The semantic extractor combines convolutional neural networks with fully connected layers to capture hierarchical feature representations. The joint source-channel coding architecture integrates residual convolution blocks, attention feature blocks, and residual transpose convolution blocks to optimize rate-distortion performance. The adaptive network produces dynamic masks through Gumbel-Softmax sampling, controlling the transmission of semantic symbols based on their relevance and channel state. A two-stage training strategy is implemented. First, end-to-end optimization of the joint source-channel encoders and decoders is performed using mean squared error loss. This is followed by full system training based on a composite loss function that jointly considers transmission rate, pixel-level distortion, semantic distortion, and perceptual quality.  Results and Discussions  Comprehensive experimental validation is conducted using the CIFAR-10, CIFAR-100, and Kodak24 datasets to assess the effectiveness of the proposed framework. Performance is evaluated using multiple metrics, including Peak Signal-to-Noise Ratio (PSNR) for objective image quality assessment and USC metric for overall system efficiency. Comparative analysis with conventional JPEG combined with LDPC and QAM schemes shows substantial performance gains, particularly under low Signal-to-Noise Ratio (SNR) conditions (Fig. 3, Fig. 4). The proposed resilient semantic communication framework consistently outperforms conventional methods across different compression ratios, demonstrating robust resistance to channel impairments and effectively mitigating the cliff effect observed in traditional SSCC systems. When compared with advanced deep learning-based approaches such as ADJSCC and DeepJSCC-V, the proposed method achieves significant improvements in both PSNR and transmission efficiency (Fig. 5, Fig. 6). Efficiency evaluation using USC metrics shows that the proposed framework achieves higher utility values across various compression settings, with performance advantages becoming more evident as SNR decreases (Fig. 7). Further analysis of different network configurations demonstrates the adaptability of the architecture in balancing computational complexity and transmission performance (Fig. 8, Fig. 9). Configurations with five network units consistently provide higher PSNR values compared to three- and four-unit designs. However, four-unit configurations achieve optimal efficiency under high SNR conditions, indicating effective resource allocation in response to varying channel states. Visual quality assessment using the Kodak dataset confirms improved reconstruction performance with reduced channel bandwidth requirements, supporting the practical feasibility of the proposed approach (Fig. 10). Computational complexity analysis shows that the five-unit configuration maintains complexity comparable to existing benchmarks, while three- and four-unit designs significantly reduce computational demands, supporting deployment on resource-constrained platforms (Table 4). These experimental results demonstrate that the proposed framework provides superior reconstruction quality, transmission efficiency, and adaptive capability across diverse and challenging wireless environments.  Conclusions  This study presents a resilient semantic communication framework that addresses key challenges in image transmission for space–air–ground networks. Experimental validation confirms that the proposed method achieves substantial improvements over conventional approaches in reconstruction quality, transmission efficiency, and adaptability.
SINR Adaptive Symbol Level Precoding and Position Joint Optimization Strategy for Multiple Unmanned Aerial Vehicles Anti-Jamming Communication
WEI Haoran, YAO Rugui, FAN Ye, MA Weixin, ZUO Xiaoya
Available online  , doi: 10.11999/JEIT250221
Abstract:
  Objective  Unmanned Aerial Vehicles (UAVs) are widely applied in areas such as traffic monitoring, wireless coverage, and precision agriculture due to their high mobility and deployment flexibility. In air–ground communication scenarios requiring flexible deployment, UAV mobility can be leveraged to counteract external malicious jamming. Further, the collaborative operation of multiple UAVs enables improved system performance. However, the broadcast nature of wireless communication renders multiple UAV communication systems vulnerable to jamming attacks that disrupt legitimate communication services. Addressing this challenge, this study proposes a Signal-to-Interference-plus-Noise Ratio (SINR) adaptive Symbol-Level Precoding (SLP) and position joint optimization strategy for anti-jamming communication in multi-UAV systems. The approach fully exploits UAV mobility to enhance communication robustness under different user requirements. By integrating Coordinated Multi-Point (CoMP) transmission with SLP in a Multi-User Multiple-Input Single-Output (MU-MISO) system, the strategy improves interference utilization, enhances system energy efficiency, and reduces computational complexity.  Methods  An SINR adaptive SLP and position joint anti-jamming optimization strategy for multiple UAVs is proposed by integrating CoMP and SLP technologies. To address the challenges of three-dimensional operational space and the overlap of nodes assigned to multiple sets, a ground-to-air multi-node matching mechanism based on three-dimensional K-means++ collaborative set partitioning is designed. To reduce the computational complexity of the joint optimization process, an iterative optimization algorithm based on particle reconstruction is developed. This algorithm simultaneously solves for both the UAV precoding matrix and spatial positions with low computational overhead. Additionally, an SINR adaptive SLP approach is introduced to enable optimized power allocation for multiple UAVs, considering the varying power characteristics of jamming and noise experienced by users. Simulation results demonstrate that the integration of CoMP and SLP technologies effectively enhances the communication performance of jammed users, while stable communication performance is maintained for ordinary users.  Results and Discussions  In the proposed UAV anti-jamming communication strategy, the SINR of jammed users is improved without compromising the normal communication performance of ordinary users. In the simulation results, UAV positions are marked with five-pointed stars (Figs. 6 and 7), UAV coverage areas are represented by circles, and users are indicated by “*”. A comparison of SINR variations under four schemes (Fig. 8) shows that the received SINR of jammed users increases by approximately 12~13 dB, while the SINR of ordinary users remains above the required threshold. When different SINR thresholds are applied, the received SINR of each user type varies accordingly (Fig. 9). By setting appropriate thresholds based on actual scenario requirements, different energy allocation effects can be achieved. Following optimization, the Bit Error Rate (BER) of jammed users is significantly reduced (Fig. 10). The constellation diagrams comparing the received signals under two precoding schemes (Fig. 11) indicate that the proposed SINR adaptive SLP strategy for multiple UAVs effectively improves the SINR of jammed users, while maintaining the communication quality of ordinary users. Moreover, the fitness evolution curve of the iterative optimization algorithm based on particle reconstruction (Fig. 12) shows that the algorithm approaches the global optimal solution at an early stage of iteration.  Conclusions  To address the challenge of anti-jamming communication for multi-user services supported by multiple UAVs, this study integrates CoMP transmission with SLP technology and proposes an SINR adaptive SLP and position joint optimization strategy for multi-UAV anti-jamming communication. The strategy is implemented in two stages. First, to solve the clustering problem in three-dimensional space and allow nodes to belong to multiple groups, a ground-to-air multi-node matching mechanism based on three-dimensional K-means++ collaborative set partitioning is designed. Second, an SINR adaptive SLP method is proposed to optimize UAV power allocation based on the SINR requirements of different user types. To reduce the computational complexity of jointly optimizing the precoding matrix and UAV positions, an iterative optimization algorithm based on particle reconstruction is developed. Simulation results demonstrate that the proposed strategy, by combining CoMP and SLP, effectively improves the communication performance of jammed users while maintaining reliable communication for ordinary users.
A Battery Internal-Short-Circuit Fault Diagnosis Method Combining Battery Phase Plane with Conformer-BiGRU Network
MAO Lin, ZHANG Haixin, HE Zhiwei, GAO Mingyu, DONG Zhekang
Available online  , doi: 10.11999/JEIT250313
Abstract:
  Objective  New Energy Vehicles (NEVs) have gained rapid popularity in recent years due to their environmental benefits and high efficiency. However, as their market share continues to grow, concerns regarding frequent malfunctions and safety risks have also increased. Among these issues, Internal Short Circuit (ISC) faults are particularly concerning due to their strong concealment and the potential for severe consequences. Without accurate diagnosis and timely intervention, ISC faults can result in serious safety incidents. Therefore, developing efficient and reliable diagnostic methods for ISC faults is of practical significance.  Methods  A novel ISC fault diagnosis method is proposed for battery packs by combining an improved battery phase plane approach with a Conformer-BiGRU network. First, the improved battery phase plane method is employed to extract two-dimensional features from voltage sequences, providing deeper spatial and structural information. Second, a Conformer-BiGRU network is employed to learn features from the voltage data. The network integrates a CNN branch for local feature extraction and a Transformer branch for global representation. A feature coupling unit fuses the outputs of both branches, which are then passed to a BiGRU module to classify individual cells within the battery pack and detect ISC faults.  Results and Discussions  The proposed method is evaluated using fault data collected from an experimental platform. The results demonstrate that the improved battery phase plane effectively distinguishes between normal and faulty batteries within a two-dimensional plane (Figure 6) and further confirm its capability to detect ISC faults with varying severity under different data volumes (Figure 9). Using the Conformer-BiGRU network for fault diagnosis, the method achieves classification accuracies of 94.30%, 92.70%, and 94.85% under FUDS, UDDS, and US06 operating conditions, respectively (Table 3), significantly exceeding the performance of comparative models. Additionally, the feature extraction module contributes to an overall performance improvement of approximately 2.04% (Table 4). These findings indicate that the proposed method exhibits strong robustness (Table 4 and Figure 11) and offers a promising approach for enhancing the safety of NEVs.  Conclusions  This study proposes a novel method for diagnosing ISC faults in battery packs by integrating an improved battery phase plane approach with a Conformer-BiGRU network. The main contributions are as follows: First, the improved battery phase plane method enhances the separability of different fault states in two-dimensional space by incorporating both voltage and its first-order differential distribution, addressing the limitations of conventional one-dimensional feature extraction. Second, a hybrid Conformer-BiGRU architecture is developed, in which the Conformer module captures local discharge characteristics, while the BiGRU module models temporal dependencies. These features are integrated through a feature coupling unit to achieve cross-level feature fusion. Third, an experimental ISC fault dataset with varying severity levels is established using a self-built testing platform. Experimental results demonstrate average diagnostic accuracy, recall, and F1-scores of 91.26%, 85.17%, and 88.09%, respectively, across three international driving cycles. Although laboratory testing verifies the effectiveness of the proposed method, real-world application requires targeted optimization. This includes adapting BiGRU parameters during the migration of the Improved Battery Phase Plane (IBPP) module and refining the Conformer’s local perception weights through transfer learning to enhance feature decoupling. Future research focuses on improving diagnostic performance under concurrent fault scenarios to enhance engineering robustness in complex operating conditions.
Features Extraction and Correlation Analysis of Multi-Source Data for Maritime Targets
LIU Ningbo, ZHANG Zihao, CHEN Baoxin, DONG Yunlong, LI Jia
Available online  , doi: 10.11999/JEIT250200
Abstract:
  Objective  The growing demand for maritime target detection and recognition has made multi-source information fusion a critical area of research. Different sensing modalities including radar, visible-light imaging, and infrared imaging offer complementary information that can improve detection and classification. However, the correlation among heterogeneous features extracted from these sources remains insufficiently understood. In addition, the effects of observational conditions on feature stability and discriminability needs further investigation. This study investigates the intrinsic relationships among multi-source features, evaluates their robustness under varying conditions, and provides theoretical support for effective multi-source feature fusion in maritime target detection.  Methods  Nine physically interpretable features are extracted across three categories: radar features (radial dimension, number of peaks, position distribution entropy, cross-range dimension, and relative average amplitude), visible image features (number of edges and horizontal projected width), and infrared image features (high-temperature connected component count and thermal texture energy). To ensure accurate feature extraction, data preprocessing consists of cleaning each compressed dataset. Radar data with excessively low signal-to-noise ratios and images with extensive occlusion are excluded. For dataset construction, radar echo data are visualized frame by frame, and a subset of radar, visible-light, and infrared images is manually annotated using LabelImg. Targets are classified into five types: passenger ships, dry cargo ships, container ships, tugboats, and search-and-rescue vessels. Based on these annotations, deep learning models are independently trained for each modality to automate annotation, and the results are manually validated to ensure quality. A standardized target dataset is then constructed by integrating the manually and automatically labeled data. Radar echo data are temporally aligned with visible-light and infrared images using prior time synchronization information. Features are extracted from each modality, and inter-feature correlations are analyzed. Spearman correlation coefficients are used to quantify relationships, and hypothesis testing is conducted to assess significance, revealing intrinsic associations among multi-source features.  Results and Discussions  Comparative analysis indicates that the correlation between radar echo and image features is strongly affected by feature attributes. Size-related features show stronger correlations, reflecting greater stability for multi-source data fusion, whereas structural features are more sensitive to observation conditions and exhibit weaker correlations. External factors including target motion state, ship type, and radar polarization mode also influence feature extraction and intermodal correlation. When targets are anchored, posture remains stable and motion blur is reduced, strengthening the correlation between radar structural features and image features. Larger vessels, such as container ships and passenger ships, benefit from multiple scattering centers and larger physical dimensions, which enhance feature extraction and intermodal correlation. In contrast, smaller vessels, such as tugboats and search-and-rescue boats, exhibit irregular structures and weaker radar backscatter, leading to lower correlations. The HH polarization mode, being less sensitive to background clutter, improves radar feature stability across various target types and enhances alignment with image features. Image feature stability also varies by modality: visible-light imaging is effective for extracting structural features, such as contours and edges, but is sensitive to illumination and occlusion; infrared imaging offers more stable size-related features and performs robustly in complex and low-visibility environments. These results highlight the complementary properties of multi-source features and establish a theoretical basis for their integration, supporting improved maritime target detection and classification.  Conclusions  This study demonstrates the complementary characteristics of multi-source features and their potential to improve maritime target detection and recognition. By analyzing feature correlations and stability across varying observational conditions, the results provide a theoretical foundation for refining multi-source fusion strategies. These insights support the development of more robust and reliable detection algorithms, contributing to enhanced situational awareness in maritime surveillance and defense.
Dynamic Analysis and Synchronization Control of Extremely Simple Cyclic Memristive Chaotic Neural Network
LAI Qiang, QIN Minghong
Available online  , doi: 10.11999/JEIT250212
Abstract:
  Objective  Memristors are considered promising devices for the construction of artificial synapses because their unique nonlinear and non-volatile properties effectively mimic the functions and mechanisms of biological synapses. These features have made memristors a research focus in brain-inspired science. Memristive neural networks, composed of memristive neurons or memristive synapses, constitute a class of biomimetic artificial neural networks that exhibit dynamic behaviors more closely aligned with those of biological neural systems and provide more plausible biological interpretations. Since the concept of the memristive neural network was proposed, extensive pioneering research has been conducted, revealing several critical issues that require further exploration. Although current memristive neural networks can generate complex dynamic behaviors such as chaos and multistability, these effects are often achieved at the cost of increased network complexity or the requirement for specialized memristive characteristics. Therefore, the systematic exploration of simple memristive neural networks that can produce diverse dynamic behaviors, the proposal of practical design strategies, and the development of efficient, precise control schemes remain of considerable research value.  Methods  This paper proposes a chaoticization method for an Extremely Simple Cyclic Memristive Convolutional Neural Network (ESCMCNN) that contains only unidirectional synaptic connections based on memristors. Using a three-node neural network as an example, a class of memristive cyclic neural networks with simple structures and rich dynamic behaviors is constructed. Numerical analysis tools, including bifurcation diagrams, basins of attraction, phase plane diagrams, and Lyapunov exponents, are employed to investigate the networks’ diverse bifurcation processes, multiple types of multistability, and multi-variable signal amplitude control. Electronic circuit experiments are used to validate the feasibility of the proposed networks. Finally, a novel multi-power reaching law is developed to achieve chaotic synchronization within fixed time.  Results and Discussions  For a three-node cyclic neural network initially in a periodic state, two network chaotification methods—full-synaptic memristivation and multi-node extension—are proposed using magnetically controlled memristors (Fig. 1). Phase plane diagrams illustrate the chaotic attractors generated by these networks (Fig. 2), confirming the feasibility of the proposed methods. Using network (B) as an example, numerical analysis tools are utilized to study its diverse dynamic evolution processes (Fig. 5, Fig. 6, Fig. 7), various forms of multistability (Fig. 8, Fig. 9), and multi-variable amplitude control (Fig. 10). The physical realization of network (B) is further demonstrated through circuit experiments (Fig. 11, Fig. 12). Additionally, the effectiveness of the fixed-time synchronization control strategy for network (B) is verified through numerical simulations (Fig. 13, Fig. 14).  Conclusions  This paper proposes a construction method for the ESCMCNN capable of generating rich dynamic behaviors. A series of ESCMCNNs is successfully designed based on a three-node neural network in a periodic state. The dynamic evolution of the ESCMCNN as a function of memristive parameters is investigated using numerical tools, including single- and dual-parameter bifurcation diagrams and Lyapunov exponents. Under different initial conditions, the ESCMCNN exhibits various forms of multistability, including the coexistence of point attractors with periodic attractors, and point attractors with chaotic attractors. The study further demonstrates that the oscillation amplitudes of multiple variables in the ESCMCNN are strongly dependent on the memristive coupling strength. The reliability of these numerical results is confirmed through electronic circuit experiments. In addition, a novel multi-power reaching law is proposed to achieve fixed-time synchronization of the network, and its feasibility and effectiveness are validated through simulation tests.
Resource Allocation in Reconfigurable Intelligent Surfaces Assisted NOMA Based Space-Air-Ground Integrated Network
LIANG Wei, LI Aoying, LUO Wei, LI Lixin, LIN Wensheng, LI Xu, WEI Baoguo
Available online  , doi: 10.11999/JEIT250078
Abstract:
  Objective  The exponential growth of 6G wireless communication demands has positioned the Space-Air-Ground Integrated Network (SAGIN) as a promising architecture, aiming to achieve broad coverage and adaptive networking. However, complex geographic environments, including building obstructions, frequently hinder direct communication between ground users and base stations, thereby requiring effective relay strategies to maintain reliability. Reconfigurable Intelligent Surfaces (RIS) have attracted considerable attention for their capacity to improve signal coverage through passive beamforming. This study develops an RIS-assisted SAGIN architecture incorporating aerial RIS clusters and High-Altitude Platforms (HAPS) to enable communication between ground users and a Low Earth Orbit (LEO) satellite. To enhance energy efficiency, the system further optimizes user relay selection, power allocation, and beamforming for both LEO and RIS components.  Methods  The proposed system integrates LEO satellites, HAPS, Unmanned Aerial Vehicles (UAVs) equipped with RIS, and ground users within a three-dimensional communication space. Due to environmental obstructions, ground users are unable to maintain direct links with base stations; instead, RIS functions as a passive relay node. To improve relay efficiency, users are grouped and associated with specific RIS units. The total system bandwidth is partitioned into sub-channels assigned to different user groups. A matching algorithm is designed for user selection, followed by user group association with each RIS. For LEO communications, HAPS serve as active relay nodes that decode and forward signals to ground base stations. The system considers both direct and RIS-assisted communication links. An optimization problem is formulated to maximize energy efficiency under constraints related to user Quality of Service (QoS), power allocation, and beamforming for both LEO and RIS. To solve this, the proposed Alternating Pragmatic Iterative Algorithm in SAGIN (APIA-SAGIN) decomposes the problem into three sub-tasks: user relay selection, LEO beamforming, and RIS beamforming. The Successive Convex Approximation (SCA) and SemiDefinite Relaxation (SDR) methods are employed to transform the original non-convex problem into tractable convex forms for efficient solution.g.  Results and Discussions  Simulation results confirm the effectiveness of the proposed APIA-SAGIN algorithm in optimizing the energy efficiency of the RIS-assisted SAGIN. As shown in (Fig. 5), increasing the number of RIS elements and LEO antennas markedly improves energy efficiency compared with the random phase shift algorithm. This demonstrates that the proposed algorithm enables channel-aware control by aligning RIS beamforming with the ground transmission channel and jointly optimizing LEO beamforming, RIS beamforming, and LEO-channel alignment. As illustrated in (Fig. 6), both energy efficiency and the achievable rate of the LEO link increase with transmission power. However, beyond a certain power threshold, energy consumption rises faster than the achievable rate, leading to diminishing or even negative returns in energy efficiency. (Fig. 7) shows that higher power in ground user groups leads to increased achievable rates. Nonetheless, expanding the number of RIS elements proves more effective than increasing transmission power for enhancing user throughput. As shown in (Fig. 8), a higher number of RIS elements leads to simultaneous improvements in energy efficiency and achievable rate in the ground segment. Moreover, increasing the number of ground users does not degrade energy efficiency; instead, it results in a gradual increase, suggesting efficient resource allocation. Compared with the random phase shift algorithm, the proposed approach achieves superior performance in both energy efficiency and achievable rate. These findings support its potential for practical deployment in SAGIN systems.  Conclusions  This study proposes an RIS-assisted SAGIN architecture that utilizes aerial RIS clusters and HAPS to support communication between ground users and LEO satellites. The APIA-SAGIN algorithm is developed to jointly optimize user relay selection, LEO beamforming, and RIS beamforming with the objective of maximizing system energy efficiency. Simulation results demonstrate the effectiveness and robustness of the algorithm under complex conditions. The proposed approach offers a promising direction for improving the energy efficiency and overall performance of SAGIN, providing a foundation for future research and practical implementation.
Overview of the Research on Key Technologies for AI-powered Integrated Sensing, Communication and Computing
ZHU Zhengyu, YIN Menglin, YAO Xinwei, XU Yongjun, SUN Gangcan, XU Mingliang
Available online  , doi: 10.11999/JEIT250242
Abstract:
The Integration of Sensing, Communication and Computing (ISCC) combined with Artificial Intelligence(AI) algorithms has emerged as a critical enabler of Sixth-Generation (6G) networks due to its high spectral efficiency and low hardware cost. AI-powered ISCC systems, which combine sensing, communication, computing, and intelligent algorithms, support fast data processing, real-time resource allocation, and adaptive decision-making in complex and dynamic environments. These systems are increasingly applied in intelligent vehicular networks—including Unmanned Aerial Vehicles (UAVs) and autonomous driving—as well as in radar, positioning, tracking, and beamforming. This overview outlines the development and advantages of AI-enabled ISCC systems, focusing on performance benefits, application potential, evaluation metrics, and enabling technologies. It concludes by discussing future research directions. Future 6G networks are expected to evolve beyond data transmission to form an integrated platform that unifies sensing, communication, computing, and intelligence, enabling pervasive AI services.  Significance   AI-powered ISCC marks a transformative shift in wireless communication, enabling more efficient spectrum utilization, reduced hardware cost, and improved adaptability in complex environments. This integration is central to the development of 6G networks, which aim to deliver intelligent and efficient services across applications such as autonomous vehicles, UAVs, and smart cities. The significance of this research lies in its potential to reshape the management and optimization of communication, sensing, and computing resources, advancing the realization of a ubiquitously connected and intelligent infrastructure.  Progress   Recent advances in AI—particularly in machine learning, deep learning, and reinforcement learning—have substantially improved the performance of ISCC systems. These methods enable real-time data processing, intelligent resource management, and adaptive decision-making, which are critical for future 6G requirements. Notable progress includes AI-driven waveform design, beamforming, channel estimation, and dynamic spectrum allocation, all of which enhance ISCC efficiency and reliability. Additionally, the integration of edge computing and federated learning has mitigated challenges related to latency, data privacy, and scalability, facilitating broader deployment of AI-enabled ISCC systems.  Conclusions  Research on AI-powered ISCC systems highlights the benefits of integrating AI with sensing, communication, and computing. AI algorithms improve resource efficiency, sensing precision, and real-time adaptability, making ISCC systems well suited for dynamic and complex environments. The adoption of lightweight models and distributed learning has broadened applicability to resource-limited platforms such as drones and IoT sensors. Overall, AI-enabled ISCC systems advance the realization of 6G networks, where sensing, communication, and computing are unified to support intelligent and ubiquitous services.  Prospects   The advancement of AI-powered ISCC systems depends on addressing key challenges, including data quality, model complexity, security, and real-time performance. Future research should focus on developing robust AI models capable of generalizing across diverse wireless environments. Progress in lightweight AI and edge computing will be critical for deployment in resource-constrained devices. The integration of multi-modal data and the design of secure, privacy-preserving algorithms will be essential to ensure system reliability and safety. As 6G networks evolve, AI-powered ISCC systems are expected to underpin intelligent, efficient, and secure communication infrastructures, reshaping human-technology interaction in the digital era.
Precise Hand Joint Motion Analysis Driven by Complex Physiological Information
YAN Jiaqing, LIU Gengchen, ZHOU Qingqi, XUE Weiqi, ZHOU Weiao, TIAN Yunzhi, WANG Jiaju, DONG Zhekang, LI Xiaoli
Available online  , doi: 10.11999/JEIT250033
Abstract:
  Objective  The human hand is a highly dexterous organ essential for performing complex tasks. However, dysfunction due to trauma, congenital anomalies, or disease substantially impairs daily activities. Restoring hand function remains a major challenge in rehabilitation medicine. Virtual Reality (VR) technology presents a promising approach for functional recovery by enabling hand pose reconstruction from surface ElectroMyoGraphy (sEMG) signals, thereby facilitating neural plasticity and motor relearning. Current sEMG-based hand pose estimation methods are limited by low accuracy and coarse joint resolution. This study proposes a new method to estimate the motion of 15 hand joints using eight-channel sEMG signals, offering a potential improvement in rehabilitation outcomes and quality of life for individuals with hand impairment.  Methods  The proposed method, termed All Hand joints Posture Estimation (AHPE), incorporates a continuous denoising network that combines sparse attention and multi-channel attention mechanisms to extract spatiotemporal features from sEMG signals. A dual-decoder architecture estimates both noisy hand poses and the corresponding correction ranges. These outputs are subsequently refined using a Bidirectional Long Short-Term Memory (BiLSTM) network to improve pose accuracy. Model training employs a composite loss function that integrates Mean Squared Error (MSE) and Kullback-Leibler (KL) divergence to enhance joint angle estimation and capture inter-joint dependencies. Performance is evaluated using the NinaproDB8 and NinaproDB5 datasets, which provide sEMG and hand pose data for single-finger and multi-finger movements, respectively.  Results and Discussions  The AHPE model outperforms existing methods—including CNN-Transformer, DKFN, CNN-LSTM, TEMPOnet, and RPC-Net—in estimating hand poses from multi-channel sEMG signals. In within-subject validation (Table 1), AHPE achieves a Root Mean Squared Error (RMSE) of 2.86, a coefficient of determination (R2) of 0.92, and a Mean Absolute Deviation (MAD) of 1.79° for MetaCarPophalangeal (MCP) joint rotation angle estimation. In between-subject validation (Table 2), the model maintains high accuracy with an RMSE of 3.72, an R2 of 0.88, and an MAD of 2.36°, demonstrating strong generalization. The model’s capacity to estimate complex hand gestures is further confirmed using the NinaproDB5 dataset. Estimated hand poses are visualized with the Mano Torch hand model (Fig. 4, Fig. 5). The average R2 values for finger joint extension estimation are 0.72 (thumb), 0.692 (index), 0.696 (middle), 0.689 (ring), and 0.696 (little finger). Corresponding RMSE values are 10.217°, 10.257°, 10.290°, 10.293°, and 10.303°, respectively. A grid error map (Fig. 6) highlights prediction accuracy, with red regions indicating higher errors.  Conclusions  The AHPE model offers an effective approach for estimating hand poses from sEMG signals, addressing key challenges such as signal noise, high dimensionality, and inter-individual variability. By integrating mixed attention mechanisms with a dual-decoder architecture, the model enhances both accuracy and robustness in multi-joint hand pose estimation. Results confirm the model’s capacity to reconstruct detailed hand kinematics, supporting its potential for applications in hand function rehabilitation and human-machine interaction. Future work will aim to improve robustness under real-world conditions, including sensor noise and environmental variation.
Research on Station Centrality and Cascade Failure Invulnerability of Urban Rail Transit Networks
RUI Xiaobin, LIN Weihan, JI Jiaxin, WANG Zhixiao
Available online  , doi: 10.11999/JEIT250182
Abstract:
  Objective  Research on node centrality in rail transit networks is essential for ensuring operational safety. Identifying critical stations enables the development of preventive strategies and mitigates the effects of station failures. Existing studies highlight two key determinants of station importance: static topology and dynamic passenger flow. However, most current approaches treat these factors separately, leading to biased estimations of node importance. To address this limitation, this study proposes a novel node centrality measure that integrates static topology and dynamic passenger flow. The method combines topology-based centrality—derived from PageRank and a modified K-shell algorithm—with passenger centrality, which is based on station inflow and outflow volumes. A reinforcement mechanism ensures that passenger centrality consistently amplifies topology-based centrality, balancing the influence of both components. Using cascade failure simulations and real-world data from the Shanghai Metro, the proposed method reliably identifies key stations. These findings offer practical guidance for the design and maintenance of robust metro systems, enhancing their resilience to cascading failures and improving overall safety and stability.  Methods  The proposed method integrates static topology and dynamic passenger flow to evaluate the centrality of urban rail transit stations, addressing the limitations of existing approaches in identifying key stations. It consists of three components: static topology centrality, dynamic passenger flow centrality, and an integration strategy. (1) Static topology centrality is computed using a combination of PageRank and an improved K-core method. This hybrid approach captures both connectivity and node importance based on iterative removal order, mitigating the loss of resolution caused by the long-tail degree distribution typical in transit networks. (2) Dynamic passenger flow centrality assigns separate weights to inbound and outbound flows to account for congestion effects and directional asymmetry—factors often overlooked in previous models. The weights are derived from average boarding and alighting times and adjusted for flow variations across morning peak, evening peak, and off-peak periods. (3) Integration strategy: An exponential function combines the two centrality measures, ensuring that passenger flow consistently amplifies topology-derived importance. This design improves sensitivity to dynamic changes while preserving structural significance. The integrated centrality metric enhances network resilience by supporting targeted protection of critical stations, based on both static and dynamic characteristics.  Results and Discussions  This study investigates the vulnerability of Shanghai’s urban rail transit network by simulating cascading failures and identifying key stations. Using dynamic passenger flow data from three representative weekdays across 14 subway lines—comprising 289 stations and 335 edges—a load-capacity model is applied to assess node importance based on each station’s effect on network stability during cascading failure events. The results (Fig. 1, Fig. 2, Fig. 3) demonstrate that the proposed method consistently and effectively identifies key stations, outperforming five benchmark approaches. When assessing passenger flow loss, passenger flow centrality alone proves more informative than static topology centrality alone (Fig. 3). Moreover, the influence of passenger flow centrality is more pronounced during morning and evening peak periods, highlighting the role of temporal dynamics in station vulnerability. These findings highlight the importance of incorporating dynamic passenger flow data into vulnerability assessments to better capture real-world operational risks. The ablation study (Table 1) confirms that the integrated centrality—combining static topology and dynamic flow—offers superior performance over single-factor methods. Prioritizing the protection of stations identified by this approach can substantially improve the network’s resilience to cascading failures and enhance overall system safety.  Conclusions  This study investigates station centrality in urban rail transit networks and analyzes the cascading failure effects triggered by key node disruptions using a load-capacity model. The proposed method enhances the ranking of station importance by jointly capturing network structure and usage patterns, offering practical value for the design and maintenance of safe and resilient metro systems.Experiments on the Shanghai Metro network show that the method effectively identifies critical stations, with improvements observed in network average efficiency, connectivity, and reduced passenger flow loss. The results indicate the following: (1) Accurate identification of key stations requires the integration of both static topology and dynamic passenger flow; relying on either alone limits precision. (2) Failures at key stations can induce substantial cascading failures, highlighting the need to prioritize their protection to improve system resilience. (3) Future research should focus on developing more effective strategies for integrating static and dynamic centrality measures to extend applicability across different urban transit networks.
Breakthrough in Solving NP-Complete Problems Using Electronic Probe Computers
XU Jin, YU Le, YANG Huihui, JI Siyuan, ZHANG Yu, YANG Anqi, LI Quanyou, LI Haisheng, ZHU Enqiang, SHI Xiaolong, WU Pu, SHAO Zehui, LENG Huang, LIU Xiaoqing
Available online  , doi: 10.11999/JEIT250352
Abstract:
This study presents a breakthrough in addressing NP-complete problems using a newly developed Electronic Probe Computer (EPC60). The system employs a hybrid serial–parallel computational model and performs large-scale parallel operations through seven probe operators. In benchmark tests on 3-coloring problems in graphs with 2,000 vertices, EPC60 achieves 100% accuracy, outperforming the mainstream solver Gurobi, which succeeds in only 6% of cases. Computation time is reduced from 15 days to 54 seconds. The system demonstrates high scalability and offers a general-purpose solution for complex optimization problems in areas such as supply chain management, finance, and telecommunications.  Objective   NP-complete problems pose a fundamental challenge in computer science. As problem size increases, the required computational effort grows exponentially, making it infeasible for traditional electronic computers to provide timely solutions. Alternative computational models have been proposed, with biological approaches—particularly DNA computing—demonstrating notable theoretical advances. However, DNA computing systems continue to face major limitations in practical implementation.  Methods  Computational Model: EPC is based on a non-Turing computational model in which data are multidimensional and processed in parallel. Its database comprises four types of graphs, and the probe library includes seven operators, each designed for specific graph operations. By executing parallel probe operations, EPC efficiently addresses NP-complete problems.Structural Features:EPC consists of four subsystems: a conversion system, input system, computation system, and output system. The conversion system transforms the target problem into a graph coloring problem; the input system allocates tasks to the computation system; the computation system performs parallel operations via probe computation cards; and the output system maps the solution back to the original problem format.EPC60 features a three-tier hierarchical hardware architecture comprising a control layer, optical routing layer, and probe computation layer. The control layer manages data conversion, format transformation, and task scheduling. The optical routing layer supports high-throughput data transmission, while the probe computation layer conducts large-scale parallel operations using probe computation cards.  Results and Discussions  EPC60 successfully solved 100 instances of the 3-coloring problem for graphs with 2,000 vertices, achieving a 100% success rate. In comparison, the mainstream solver Gurobi succeeded in only 6% of cases. Additionally, EPC60 rapidly solved two 3-coloring problems for graphs with 1,500 and 2,000 vertices, which Gurobi failed to resolve after 15 days of continuous computation on a high-performance workstation.Using an open-source dataset, we identified 1,000 3-colorable graphs with 1,000 vertices and 100 3-colorable graphs with 2,000 vertices. These correspond to theoretical complexities of O(1.3289n) for both cases. The test results are summarized in Table 1.Currently, EPC60 can directly solve 3-coloring problems for graphs with up to n vertices, with theoretical complexity of at least O(1.3289n).On April 15, 2023, a scientific and technological achievement appraisal meeting organized by the Chinese Institute of Electronics was held at Beijing Technology and Business University. A panel of ten senior experts conducted a comprehensive technical evaluation and Q&A session. The committee reached the following unanimous conclusions:1. The probe computer represents an original breakthrough in computational models.2. The system architecture design demonstrates significant innovation.3. The technical complexity reaches internationally leading levels.4. It provides a novel approach to solving NP-complete problems.Experts at the appraisal meeting stated, “This is a major breakthrough in computational science achieved by our country, with not only theoretical value but also broad application prospects.” In cybersecurity, EPC60 has also demonstrated remarkable potential. Supported by the National Key R&D Program of China (2019YFA0706400), Professor Xu Jin’s team developed an automated binary vulnerability mining system based on a function call graph model. Evaluation of the system using the Modbus Slave software showed over 95% vulnerability coverage, far exceeding the 75 vulnerabilities detected by conventional depth-first search algorithms. The system also discovered a previously unknown flaw, the “Unauthorized Access Vulnerability in Changyuan Shenrui PRS-7910 Data Gateway” (CNVD-2020-31406), highlighting EPC60’s efficacy in cybersecurity applications.The high efficiency of EPC60 derives from its unique computational model and hardware architecture. Given that all NP-complete problems can be polynomially reduced to one another, EPC60 provides a general-purpose solution framework. It is therefore expected to be applicable in a wide range of domains, including supply chain management, financial services, telecommunications, energy, and manufacturing.  Conclusions   The successful development of EPC offers a novel approach to solving NP-complete problems. As technological capabilities continue to evolve, EPC is expected to demonstrate strong computational performance across a broader range of application domains. Its distinctive computational model and hardware architecture also provide important insights for the design of next-generation computing systems.
Covert Communication Transmission Scheme in Low Earth Orbit Satellite Integrated Sensing and Communication Systems
ZHU Zhengyu, OUYANG Zebin, PAN Gaofeng, WANG Shuai, SUN Gangcan, CHU Zheng, HAO Fengyu
Available online  , doi: 10.11999/JEIT250208
Abstract:
  Objective  Globalization has driven growing demand for ubiquitous communication. However, traditional terrestrial networks face limitations in areas with poor infrastructure, such as deserts, oceans, and airspace. Satellite communication, with its flexibility, wide frequency coverage, and high capacity, offers a viable alternative for these environments. Yet, due to the open nature of satellite channels and long-distance transmission, signals remain vulnerable to interception, posing serious security challenges. Conventional wireless security measures, including physical-layer techniques and encryption, are limited by assumptions about computational hardness and the statistical structure of intercepted ciphertexts. Covert communication aims to reduce the probability of signal detection by eavesdroppers, providing an alternative layer of security. In Low Earth Orbit (LEO) satellite systems, high mobility leads to dynamic channel variations and misalignment of transmission beams, further complicating covert communication. To mitigate these challenges, Integrated Sensing And Communication (ISAC) technology can be embedded in satellite systems. By detecting the approximate location of potential eavesdroppers and directing radar beams to degrade their detection capability, ISAC improves communication covertness. This study proposes a covert communication strategy tailored for LEO satellite ISAC systems, offering enhanced security in satellite-based infrastructure.  Methods  This paper investigates a covert communication scheme in LEO satellite ISAC systems. First, the system model of the LEO satellite ISAC system is proposed, and the covertness constraints within the system are analyzed. Based on this, an optimization problem is formulated with the objective of maximizing the total covert communication rate for multiple users, subject to constraints including satellite power limits, radar power limits, sensing performance thresholds, and covertness requirements. This optimization problem is non-convex and highly coupled in variables, making it intractable to solve directly. To address this, relaxation variables are introduced to simplify the objective function into a solvable form. Subsequently, an alternating optimization algorithm is employed to decompose the original problem into two subproblems. One subproblem is transformed into an equivalent convex problem using methods like SemiDefinite Relaxation (SDR) and Successive Convex Approximation (SCA), which is then solved iteratively. The obtained solution from this subproblem is then used to solve the other subproblem. Once the solutions to both subproblems are determined, the original problem is effectively solved.  Results and Discussions  This paper validates the performance of the proposed covert communication transmission scheme through numerical simulations. To demonstrate the effectiveness of the proposed scheme, a comparison with a system lacking ISAC technology is introduced. First, the convergence of the proposed algorithm is verified (Fig. 2). The algorithm achieves rapid convergence under different resource allocation configurations in the system (converging at the 7th iteration). Additionally, higher covertness requirements lead to lower covert communication rates, as stricter covertness constraints reduce the allowable satellite transmit power. Increasing radar power causes a slight decline in the covert rate because Bob is subjected to minor interference from the radar beam. However, since the primary radar beam is focused on Eve, the interference level remains limited. Second, a higher number of satellite antennas improves the system’s covert communication rate (Fig. 3). This is attributed to the increased spatial dimensions available for optimizing signal transmission and reception. More antennas enable additional independent signal paths, enhancing transmission performance and flexibility. When the satellite is equipped with 81 antennas and the covertness requirement is 0.1, the total covert rate reaches 17.18 bit/(s·Hz). Compared to the system without ISAC technology, the proposed system demonstrates superior resistance to detection (Fig. 4). The radar beam introduces interference to eavesdroppers, and higher radar power degrades the eavesdropper’s detection performance. Finally, increasing the satellite transmit power improves the covert communication rate (Fig. 5). In the system without ISAC, the satellite power cannot be arbitrarily increased due to covertness constraints. However, the ISAC-enabled system leverages radar integration to grant the satellite greater flexibility in power allocation, thereby enhancing the covert rate. Notably, raising the lower bound of sensing performance slightly reduces the covert rate, as higher sensing demands require increased radar power, which introduces additional interference to Bob’s signal reception.  Conclusions  This paper investigates a covert communication transmission scheme in an LEO satellite-enabled ISAC system. An optimization problem is formulated to maximize the sum of multi-user covert communication rates, subject to constraints including covertness requirements, minimum sensing performance, radar power budget, and satellite power budget. The alternating optimization algorithm and SCA algorithm are jointly employed to design the radar and satellite beamforming, as well as the radar receiver filter. Simulation results demonstrate that, compared to a system without radar beam interference, the proposed system with radar beam interference significantly reduces the detectability of communication between Alice and Bob. Furthermore, the integration of radar enhances the satellite’s flexibility to increase transmit power, thereby improving the overall covert communication rate of the system.
Research on an EEG-based Neurofeedback System for the Auxiliary Intervention of Post-Traumatic Stress Disorder
TAN Lize, DING Peng, WANG Fan, LI Na, GONG Anmin, NAN Wenya, LI Tianwen, ZHAO Lei, FU Yunfa
Available online  , doi: 10.11999/JEIT250093
Abstract:
  Objective  The ElectroEncephaloGram (EEG)-based Neurofeedback Regulation (ENR) system is designed for real-time modulation of dysregulated stress responses to reduce symptoms of Post-Traumatic Stress Disorder (PTSD) and anxiety. This study evaluates the system’s effectiveness and applicability using a series of neurofeedback paradigms tailored for both PTSD patients and healthy participants.  Methods  Employing real-time EEG monitoring and feedback, the ENR system targets the regulation of alpha wave activity, to alleviate mental health symptoms associated with dysregulated stress responses. The system integrates MATLAB and Unity3D to support a complete workflow for EEG data acquisition, processing, storage, and visual feedback. Experimental validation includes both PTSD patients and healthy participants to assess the system’s effects on neuroplasticity and emotional regulation. Primary assessment indices include changes in alpha wave dynamics and self-reported reductions in stress and anxiety.  Results and Discussions  Compared with conventional therapeutic methods, the ENR system shows significant potential in reducing symptoms of PTSD and anxiety. During functionality tests, the system effectively captures and regulates alpha wave activity, enabling real-time and efficient neurofeedback. Dynamic adjustment of feedback thresholds and task paradigms allows participants to improve stress responses and emotional states following training. Quantitative data indicate clear enhancements in EEG pattern modulation, while qualitative assessments reflect improvements in participants’ self-reported stress and anxiety levels.  Conclusion  This study presents an effective and practical EEG-based neurofeedback regulation system that proves applicable and beneficial for both individuals with PTSD and healthy participants. The successful implementation of the system provides a new technological approach for mental health interventions and supports ongoing personalized neuroregulation strategies. Future research should explore broader applications of the system across neurological conditions to fully assess its efficacy and scalability.
A Novel Silicon Carbide (SiC) MOSFET with Diode Integration Technology
MA Chao, CHEN Weizhong, ZHANG Bo
Available online  , doi: 10.11999/JEIT250180
Abstract:
This paper proposes a novel double-trench Silicon Carbide (SiC) MOSFET that integrates a Schottky diode structure to improve reverse recovery and switching characteristics. In the proposed design, the conventional right-side trench channel is replaced by a Schottky diode, and a split-gate structure is connected to the source. The Schottky diode suppresses body diode conduction and eliminates the bipolar degradation effect. The split gate reduces the coupling area between the gate and drain, thereby lowering the feedback capacitance and gate charge. In addition, when the split gate is connected to a high potential, it attracts electrons to form an accumulation layer near the source, which increases electron density. During reverse conduction, the current flows through the Schottky diode, while the split gate enhances electron concentration and thus current density. The split-gate structure also shields the gate from the drain, reducing the Gate–Drain Charge (QGD) and improving switching performance.  Objective  Conventional Double-Trench MOSFETs (DT-MOS) typically require an external anti-parallel diode to function as a freewheeling diode in converter and inverter systems. This necessitates additional module area and increases parasitic capacitance and inductance. Utilizing the body diode as a freewheeling diode could reduce cost and save space. However, this approach presents two major challenges. First, due to the wide bandgap of SiC, the turn-on voltage of the intrinsic body diode rises significantly (approaching 3 V), which increases switching loss. This paper presents a new DT-MOS, referred to as SDT-MOS, with an integrated Schottky diode, demonstrated using TCAD SENTAURUS. In the proposed structure, the conventional right-side channel is replaced with a Schottky junction, and a source-connected split gate is embedded in the gate oxide. The SDT-MOS achieves low power consumption and reduced reverse recovery current.  Methods  Sentaurus TCAD is used to simulate and analyze the electrical performance of the proposed structure and its conventional counterpart. The simulation includes key physical models, such as mobility saturation under high electric fields, Auger recombination, Okuto–Crowell impact ionization, bandgap narrowing, and incomplete ionization. To improve simulation accuracy and align the results with experimental data, interface traps and fixed charges at the SiC/SiO2 interface are also considered.  Results and Discussions  The Miller capacitance (Crss or CGD) extracted at Vds of 400 V is 29 pF/cm2 for the SDT-MOS, representing a 61% reduction compared to the DT-MOS, which has a CGD of 74 pF/cm2. This reduction is primarily attributed to the integrated split-gate structure, which decreases the capacitive coupling between the gate and drain electrodes (Fig. 7). The total switching loss (Eon + Eoff) of the SDT-MOS is 1.58 mJ/cm2 , which is 59.3% lower than that of the DT-MOS (3.88 mJ/cm2 ), due to the improved switching characteristics enabled by the split gate (Fig. 10). In addition, the peak reverse recovery current (IRRM) and reverse recovery charge (QRR) of the SDT-MOS are 165 A/cm2 and 1.39 μC/cm2, representing reductions of 31.3% and 54%, respectively, compared to the DT-MOS (Fig. 11).  Conclusions  A novel double-trench SiC MOSFET (SDT-MOS) with an integrated Schottky diode has been numerically investigated. In this structure, the right-side channel of a conventional DT-MOS is replaced with a Schottky diode, and a split gate is connected to the source. This configuration results in improved switching and reverse recovery performance. With appropriate optimization of key design parameters, the SDT-MOS retains the fundamental characteristics of a standard MOSFET. Compared with the conventional DT-MOS, the proposed device suppresses body diode conduction, mitigates bipolar degradation, and achieves a 64.9% reduction in QGD. Switching loss is reduced by 59.3%, and QRR is reduced by 54%. These enhancements make the SDT-MOS a strong candidate for high-efficiency, high–power density applications.
Personalized Federated Learning Method Based on Collation Game and Knowledge Distillation
SUN Yanhua, SHI Yahui, LI Meng, YANG Ruizhe, SI Pengbo
Available online  , doi: 10.11999/JEIT221203
Abstract:
To overcome the limitation of the Federated Learning (FL) when the data and model of each client are all heterogenous and improve the accuracy, a personalized Federated learning algorithm with Collation game and Knowledge distillation (pFedCK) is proposed. Firstly, each client uploads its soft-predict on public dataset and download the most correlative of the k soft-predict. Then, this method apply the shapley value from collation game to measure the multi-wise influences among clients and quantify their marginal contribution to others on personalized learning performance. Lastly, each client identify it’s optimal coalition and then distill the knowledge to local model and train on private dataset. The results show that compared with the state-of-the-art algorithm, this approach can achieve superior personalized accuracy and can improve by about 10%.
The Range-angle Estimation of Target Based on Time-invariant and Spot Beam Optimization
Wei CHU, Yunqing LIU, Wenyug LIU, Xiaolong LI
Available online  , doi: 10.11999/JEIT210265
Abstract:
The application of Frequency Diverse Array and Multiple Input Multiple Output (FDA-MIMO) radar to achieve range-angle estimation of target has attracted more and more attention. The FDA can simultaneously obtain the degree of freedom of transmitting beam pattern in angle and range. However, its performance is degraded due to the periodicity and time-varying of the beam pattern. Therefore, an improved Estimating Signal Parameter via Rotational Invariance Techniques (ESPRIT) algorithm to estimate the target’s parameters based on a new waveform synthesis model of the Time Modulation and Range Compensation FDA-MIMO (TMRC-FDA-MIMO) radar is proposed. Finally, the proposed method is compared with identical frequency increment FDA-MIMO radar system, logarithmically increased frequency offset FDA-MIMO radar system and MUltiple SIgnal Classification (MUSIC) algorithm through the Cramer Rao lower bound and root mean square error of range and angle estimation, and the excellent performance of the proposed method is verified.
Satellite Navigation
Research on GRI Combination Design of eLORAN System
LIU Shiyao, ZHANG Shougang, HUA Yu
Available online  , doi: 10.11999/JEIT201066
Abstract:
To solve the problem of Group Repetition Interval (GRI) selection in the construction of the enhanced LORAN (eLORAN) system supplementary transmission station, a screening algorithm based on cross interference rate is proposed mainly from the mathematical point of view. Firstly, this method considers the requirement of second information, and on this basis, conducts a first screening by comparing the mutual Cross Rate Interference (CRI) with the adjacent Loran-C stations in the neighboring countries. Secondly, a second screening is conducted through permutation and pairwise comparison. Finally, the optimal GRI combination scheme is given by considering the requirements of data rate and system specification. Then, in view of the high-precision timing requirements for the new eLORAN system, an optimized selection is made in multiple optimal combinations. The analysis results show that the average interference rate of the optimal combination scheme obtained by this algorithm is comparable to that between the current navigation chains and can take into account the timing requirements, which can provide referential suggestions and theoretical basis for the construction of high-precision ground-based timing system.