1. Pipeline
The research methodology involves several key stages, from dataset creation to performance
evaluation. The study began by creating a diverse dataset by combining data from the
AFLOW and Materials Project databases, which contain comprehensive material properties.
A meticulous feature selection process followed, identifying the most significant
variables influencing the work function based on both domain knowledge and statistical
analysis.
To combat bias and overfitting, data augmentation was used to expand the dataset artificially.
This involved applying transformations and perturbations to the original data, creating
a more diverse set for training machine learning (ML) and deep learning (DL) models.
These models were trained using cross-validation to predict the work function of materials,
a critical parameter for evaluating FinFET performance.
The predicted work function values from the trained models were implemented into TCAD
Sentaurus simulations for a 14 nm FinFET architecture (Fig. 1). This allowed for realistic simulations of the device’s performance, bridging the
gap between theoretical and material properties and real-world device behavior. From
the I-V profile, the Subthreshold Swing (SS), ON-Current (Ion), and OFF-Current (Ioff)
were extracted to evaluate FinFET performance. This data, combined with the predicted
work function, provided a comprehensive assessment of the device’s suitability for
14 nm technology node applications.
Fig. 1. Workflow of ML/DL implementation in Advanced 14 nm Technology node. The dataset
construction followed by data augmentation, and then fed to ML/DL models for work
function inference. Finally, inference data implemented in TCAD.
2. Model Performance
Before modeling, exploratory data analysis (EDA) was performed to understand the dataset’s
properties. A Pearson correlation analysis showed no strong linear relationships between
features, with all coefficients below 0.49. However, a Variance Inflation Factor (VIF)
analysis revealed high scores for ‘sites’ and ‘volume,’ indicating multicollinearity
that the initial correlation analysis missed. A pair plot analysis further identified
non-linear relationships and cluster formations, specifically between ‘bandgap’ and
‘work function,’ and ‘Energy Above Hull’ (EAH) and ‘bandgap,’ suggesting underlying
groupings in the data beyond simple linear associations.
Table 1. Exploratory data analysis.
|
Count
|
7500
|
|
Pearson Correlation (all parameters)
|
< 0.49
|
|
Sites (VIF)
|
16.49
|
|
Energy above hull (VIF)
|
1.1
|
|
Volume (VIF)
|
13.11
|
|
Density (VIF)
|
3.95
|
|
Band gap (VIF)
|
1.59
|
|
Work function (VIF)
|
1.31
|
|
Dielectric constant (VIF)
|
1.23
|
As for ML analysis, it starts with Linear Regression, Lasso, Ridge, and Gradient Boosting
regression. After that explored with Ensemble Algorithms like Extreme Boost (XGB)
Regression, Random Forest (RF) Regression, and Decision Tree regression. As for DL
algorithm, Neural Network Regression. As for benchmarking ML and DL algorithms are
compared with Genetic algorithm. All algorithms’ performance is depicted in Table 2.
Table 2. Model performance.
|
Model
|
RMSE
|
MAE
|
R2 |
|
Linear Regression
|
0.88
|
0.64
|
0.13
|
|
Ridge Regression
|
0.88
|
0.64
|
0.13
|
|
Lasso Regression
|
0.94
|
0.69
|
0.13
|
|
Elastic Net Regression
|
0.88
|
0.65
|
0.13
|
|
Gradient Boost Regression
|
0.29
|
0.25
|
0.91
|
|
XGB Regression
|
0.31
|
0.27
|
0.88
|
|
Decision Tree Regression
|
0.31
|
0.26
|
0.89
|
|
Random Forest Regression
|
0.32
|
0.27
|
0.88
|
|
KNN Regression
|
0.32
|
0.26
|
0.89
|
|
Support Vector Regression
|
0.39
|
0.31
|
0.82
|
|
Neural Network Regression
|
0.14
|
0.14
|
0.84
|
|
Genetic Algorithm
|
1.01
|
0.71
|
−0.03
|
From Table 2, we noticed that comparatively simpler regression algorithms like linear regression,
Ridge, Lasso, and Elastic Net Regressor performed poorly in this regard. Their R2 scores are 0.13 with high Root Mean Square Error (RMSE) and Mean Absolute Error (MAE),
which is not acceptable. As a result, these four algorithms are not suitable for this
dataset. The reason for this poor performance can be the inherent behavior of the
algorithms. As these algorithms approximate their prediction in a linear manner and
the dataset is not linear, the prediction is wrong with a high value of RMSE and MAE.
As RMSE and MAE increase, R2 score will decrease. Fig. 2 exhibits the summary of all models’ performance.
Fig. 2. Visualization of predicted and true values of work function with several ML
algorithms. (a) Gradient Boost Regression. (b) XGBoost Regression. (c) Decision Trree
Regression. (d) Random Forest Regression. (e) Neural Network Regression. (f) Support
Vector Regression. R2 score is also mentioned in the images.
Boosting algorithms like Gradient Boost and Extreme Gradient Boost (XGB) perform well
in this context. Gradient Boost regression has a lower value of RMSE and MAE. In the
case of R2 score, the value is high, which implies the model is getting overfitted. On the other
hand, XGB has RMSE and MAE values within a tolerable limit. R2 score of 0.88 ensures the model is not biased or overfitted. XGB identifies parameter
‘density’ as the most important feature (Fig. 3(a)).
Decision Tree regressor and Random Forest regressor are regarded as ensemble algorithms.
These two algorithms have RMSE and MAE of low value, and the difference is 0.01 eV.
R2 scores are in an excellent range of 0.88 to 0.89, which implies these two algorithms
are not biased or overfitted. In the case of feature importance, both of the models
show ‘density’ feature as the most important feature to predict work function (Figs. 3(b) and 3(c)).
Fig. 3. Feature Importance from Several ML models. (a) XGB. (b) Decision Tree Regressor.
(c) Random Forest Regressor.
K nearest neighbor algorithm shows similar performance to ensemble algorithms. KNN
performs 0.01 eV more than ensemble algorithms in R2 scores. Support vector regressor with RBF kernel performs the same in the case of
RMSE and MAE. In case of R2 scores, the value is lower little lower than KNN.
Neural network Regression has the lowest value of RMSE and MAE. The reason for this
lowest value is neural nets. As each neural network is trained with Adam optimizer
and back propagation, the model predicts work function with the least error and a
high R2 score. Fig. 4 depicts the performance of neural network regression. Fig. 4(a) exhibits Predicted work function Vs True Work function with 0.84 R2 score. Figs. 4(b) and 4(c) depict Loss Vs Epoch and Mean Squared Logarithmic Error Vs Epoch. Both of the cases,
train and test sets aligned, which implies the error is decreasing with time and learning
is progressing with the least error.
Fig. 4. Neural network Regression performance. (a) Predicted work function Vs True
work function. (b) Loss vs Epoch for train and test sets. Both sets aligned at the
end (c) Mean Squared Logarithmic Error vs Epoch for Train and Test sets.
Genetic algorithm is initiated with four parents and each iteration, there will be
six hundred generations with 5% mutation rate with random mutation. Parent selection
type is ‘sss’. Crossover is the process where two parent chromosomes exchange genetic
material to produce new, unique offspring chromosomes. This introduces new combinations
of genes into the population, combining the best traits of both parents. For our case
it is “single point” (Fig. 5). Generation Vs Fitness is a steep curve so as is Generation Vs New Solution rate
(Figs. 6(a) and 6(b)). The performance is the worst in this research. RMSE and MAE values are the highest
of all the models, and R2 score is the lowest of all the algorithms.
Fig. 5. Some genetic evolution in the genetic algorithm.
Fig. 6. (a) Generation vs fitness in Genetic Algorithm. (b) Generation vs new solution
rate.
Summary of all ML/DL models, simple linear regression models can be eliminated because
of poor performance. All the ensemble algorithms, boosting algorithms and neural network
regressor have high R2 score values without any sign of bias and overfitting. But in the case of MAE, RMSE,
Neural Network regressor has the lowest values for all. MAE, RMSE play pivotal role
in bias and overfitting. So low value of MAE, RMSE indicate less possibility of bias
and overfitting. Considering all the parameters, neural network regressor is the best
model for this research (Fig. 7).
Fig. 7. Work function (eV) of different models.
Neural Network Regression (NNR) is selected to predict the work function. NNR model
predicts monoclinic Hafnium Oxide (Hf2O) with the work function of 4.66 eV. Implying NNR data in TCAD Sentaurus SProcess
along with other process parameters like 10 nm fin thickness, 6 nm thick TiCl gate,
4e18 cm3 doping concentration, 1 nm thick HfO2 gate oxide. 14 nm technology FinFET is simulated with Sentaurus Process (Fig. 8(a)). The Characteristics of the FinFET are analyzed with Sentaurus SDevice (Figs. 8(b) and 8(c)). With 4.6 eV (data predicted from NNR) On current is 2.5 e–6 A and Off current is
1 e-10 A. On current in different switching condition with varying work function for
our 14 nm technology FinFET is depicted in Fig. 8(d). From Fig. 8(d), for 4.6 eV work function for logic 1 On current is 1.5 e–6 A. From Sprocess and
SDevice simulation, it is prominent that 14 nm FinFET is working well with NNR material
data, and the research pipeline is working well.
Fig. 8. FinFET TCAD simulation. (a) Process simulation of 14 nm FinFET. (b) I-V characteristics
of 14 nm FinFET. (c) Logarithmic I-V. (d) Work function vs on current in different
logic state.