FTT-NAS: Implementation and Review

Overview
The FTT-NAS (Fast Transformer-based Neural Architecture Search) is an advanced method for optimizing neural architectures using an efficient search strategy. This approach aims to enhance model performance while reducing computational costs, making it suitable for real-world applications in Natural Language Processing (NLP) and Computer Vision (CV).
📄 Read our full paper here: ACL Anthology
FTT-NAS Project Report
1. Paper Summary
The FTT-NAS paper introduces a novel method for Fault-Tolerant Neural Architecture Search, aiming to reduce the impact of hardware faults on the performance of neural networks. As neural networks are increasingly deployed in embedded systems, they become vulnerable to various faults, including cosmic radiation-induced noise, voltage instability, temperature fluctuations, and hardware defects.
The paper investigates two fault models:
- Feature Fault Model (MiBB) – For faults occurring in input feature data.
- Weight Fault Model (adSAF) – For faults affecting the network’s weights.
The proposed method, FTT-NAS, employs a multi-objective fault-tolerant architecture search, incorporating Fault-Tolerant Training (FTT). Experimental results show that architectures discovered through this approach, such as F-FTT-Net and W-FTT-Net, exhibit greater resilience to hardware faults compared to manually designed models like ResNet-18 and MobileNet-V2.
2. Proposed Method and Its Strengths & Weaknesses
Strengths of FTT-NAS
- Improved Accuracy Under Fault Conditions:
- F-FTT-Net maintained 86.2% accuracy under fault conditions, while MobileNet-V2 only achieved 68.1%.
- W-FTT-Net retained 69.6% accuracy under weight faults (SAF), compared to 60.8% for ResNet-18.
- Reduction in Parameters & FLOPs:
- Discovered models optimize parameters and FLOPs, making them more efficient than manually designed models.
- Example: F-FTT-Net has only 245M FLOPs, compared to MobileNet-V2’s 182M FLOPs, but offers significantly better accuracy.
- High Flexibility:
- Applicable to various hardware platforms such as FPGA, RRAM, and ASIC.
- Can be combined with other fault-reduction techniques like Redundancy and Scrubbing.
Weaknesses of FTT-NAS
- High Complexity and Processing Time:
- The architecture search process involves extensive trial and error, leading to high computational costs.
- Shared Weights Evaluation was used to speed up processing, but it still demands substantial computation.
- Limited Fault Model Combination:
- The method evaluates only one type of fault (feature or weight) at a time.
- It does not support simultaneous multi-fault robustness search.
Overall Process
SuperNet Training
- The model is first trained without fault injection using a standard approach for multiple epochs to learn shared weights.
- In this stage, error reduction is observed, and both training and validation accuracy improve over epochs.
Controller Training (NAS Search)
- A Transformer-based Controller is used to explore different architectures through Neural Architecture Search (NAS).
- The controller generates candidate architectures by selecting operations for each SuperNet cell.
- Architectures are evaluated based on clean accuracy and faulty accuracy, optimizing the search using a reward function.
Fault Injection & Final Training
- The best architecture discovered by the Controller is applied to SuperNet.
- The final training process includes fault injection to test model robustness.
- A balanced error function is used to account for both clean and faulty data.
Reliability Calculation
- After training, reliability is computed by comparing model performance in fault-free and fault-induced conditions.
- Reliability is measured for accuracy and F1-score, showing how well the model withstands errors.
Training Results
Controller Training Results
(Results of training the controller on various architectures.)
Finalization & Retraining Results
(Results after finalizing and refining the best architecture.)
Challenges Faced & Limitations
- Some errors were not fully resolved due to:
- Insufficient training data.
- Limited number of training steps.
- Small search space (due to lower computational resources compared to the original paper’s authors).
- However, during higher training stages, the learning loss significantly decreased in the finalize and retrain phase compared to the initial training phase, showing noticeable improvement.
Final Code Improvements & Optimizations
1. Data Processing & Preprocessing Enhancements
- Applied Normalization & Data Augmentation for improved generalization.
2. SuperNet Optimization & Structural Improvements
- Added ResidualBlock, SEBlock, and Attention Mechanisms to improve learning stability.
3. NASCell Optimization & Processing Fixes
- Fixed Tensor mismatches and output inconsistencies.
4. Shared Weight Training Enhancements
- Boosted accuracy using Label Smoothing and AdamW optimizer.
Conclusion
The project aimed to enhance fault tolerance in neural architecture search. While resource limitations restricted full replication of the original study, multiple optimizations significantly improved model stability, accuracy, and robustness. The final FTT-NAS model demonstrated superior fault tolerance compared to traditional architectures, paving the way for future enhancements in fault-resilient deep learning models.
2024
1 minute read
Overview
In this project, we fine-tuned the Wav2Vec2 model to perform sentiment analysis based on both voice features and text transcripts from the Shemo da...
1 minute read
Overview
In this project, we developed an Android application to estimate the location of cellular network cells using Received Signal Strength Indicator (RS...
less than 1 minute read
Overview
In this project, we explore multimodal sentiment analysis, which involves analyzing both text and image data together. Our goal is to predict sentim...
Back to top ↑
2023
1 minute read
Project Iridium
less than 1 minute read
My Internship at the NLP Lab :
Back to top ↑
2021
less than 1 minute read
Its My Heroku project :
less than 1 minute read
My sonic pi project :
. این پست من مربوط به پوروژه سونیک پای بنده است
Back to top ↑
2020
1 minute read
My works and wishes
Success secret
مصاحبه با جناب اقا پارسا:
ایشون بسیار ادم سخت کوشی بودن و گفتن که حتی در دوران دانشجویی شون درس هم میدادن
1 به موار...
less than 1 minute read
Its my favorite university
This is oxford:
In England
The best unniversity.
this university is first inthe world.
this university have best descover...
less than 1 minute read
Its My Hackathon project :
Back to top ↑