Next Article in Journal
Human β-Defensins in Diagnosis of Head and Neck Cancers
Previous Article in Journal
DRAM1 Promotes Lysosomal Delivery of Mycobacterium marinum in Macrophages
Previous Article in Special Issue
SGAEMDA: Predicting miRNA-Disease Associations Based on Stacked Graph Autoencoder
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

iProm-Sigma54: A CNN Base Prediction Tool for σ54 Promoters

1
Department of Electronics and Information Engineering, Jeonbuk National University, Jeonju 54896, Republic of Korea
2
School of Pharmacy, Jeonbuk National University, Jeonju 54896, Republic of Korea
3
School of International Engineering and Science, Jeonbuk National University, Jeonju 54896, Republic of Korea
4
Advanced Electronics and Information Research Center, Jeonbuk National University, Jeonju 54896, Republic of Korea
*
Authors to whom correspondence should be addressed.
Cells 2023, 12(6), 829; https://doi.org/10.3390/cells12060829
Submission received: 9 December 2022 / Revised: 23 February 2023 / Accepted: 23 February 2023 / Published: 7 March 2023
(This article belongs to the Special Issue Advances of Deep Learning in Cell Biology)

Abstract

:
The sigma ( σ ) factor of RNA holoenzymes is essential for identifying and binding to promoter regions during gene transcription in prokaryotes. σ 54 promoters carried out various ancillary methods and environmentally responsive procedures; therefore, it is crucial to accurately identify σ 54 promoter sequences to comprehend the underlying process of gene regulation. Herein, we come up with a convolutional neural network (CNN) based prediction tool named “iProm-Sigma54” for the prediction of σ 54 promoters. The CNN consists of two one-dimensional convolutional layers, which are followed by max pooling layers and dropout layers. A one-hot encoding scheme was used to extract the input matrix. To determine the prediction performance of iProm-Sigma54, we employed four assessment metrics and five-fold cross-validation; performance was measured using a benchmark and test dataset. According to the findings of this comparison, iProm-Sigma54 outperformed existing methodologies for identifying σ 54 promoters. Additionally, a publicly accessible web server was constructed.

1. Introduction

The process of translating genetic data encoded in DNA into a matching RNA molecule is called transcription. RNA polymerase, an enzyme, reads the DNA code during transcription to create messenger RNA (mRNA), a single-stranded RNA molecule that retains the genetic material to the ribosome for protein synthesis. With the exception of the change of uracil for thymine in the RNA molecule, the mRNA molecule is a complimentary copy of the DNA sequence. Transcription is the initial step in gene expression and an important mechanism in regulating cellular activity [1]. A promoter is a DNA sequence that regulates transcription initiation throughout the transcription process. The promoter is the site on the DNA where the RNA polymerase enzyme attaches and begins the transcription process. The promoter sequence serves as a binding site for RNA polymerase and other regulatory proteins that govern transcription start, rate, and termination. The existence of unique promoter sequences enables gene expression regulation, allowing cells to react to various stimuli and govern the creation of certain RNA and protein products. Different kinds of promoters may result in varying degrees of gene expression and can decide whether a gene is tissue-specific or inducible. As a result, the promoter is an essential component in the control of gene expression. The sigma σ factor of RNA holoenzymes is essential for recognition and binding of promoter regions during gene transcription in prokaryotes [2]. As a result, prokaryotic promoters are classified corresponding to ( σ ) factors. Seven known σ factors exist, such as σ 19 , σ 24 , σ 28 , σ 32 , σ 38 , σ 54 , and, σ 70 and every ( σ ) factor serves a distinct purpose. The σ 54 factor can identify the consensus sequences TGCATTA and CTTGGCACTGA, which are around −12 base pairs (bp) and −24 bp upstream of the transcription start site (TSS), respectively. σ 54 promoters are involved in a variety of supporting activities and ecologically responsible practices [2], such as the extraction of chemotaxis transducers, construction of motility organs [3], nitrogen fixation [4], arginine catabolism [5], alginate biosynthesis [6], and flagella [7]. Therefore, understanding the procedure of gene transcription and establishing gene expression networks requires knowledge of σ 54 promoters. As a result, it is crucial to identify binding promoters for a specific σ factor for future research into gene regulation and functional genomics [8]. Since promoters are essential for gene transcription, precise identification of promoter locations has become essential for understanding gene expression, interpreting patterns, and building genetic regulatory networks. Several biological investigations have been carried out to determine promoters, including mutational analyses [9,10] and immunoprecipitation tests [11]. As the investigational discovery of promoters is costly and time taking, the computerized identification of σ factor promoters has emerged as an important topic in bioinformatics.
Promoter sequences may be given computationally as a classification problem, that is, as a way to identify whether a query sequence is a promoter of a particular σ factor, based on its feature characteristics. Promoter predictors may be divided into three types, namely the signal-based approach, the content-based approach, and the GpG-based approach, according to the methodology that was used to analyze them. The non-element parts of the sequence are disregarded by signal-based predictors in favor of the promoter elements that are directly connected to the RNA polymerase binding site. As a direct consequence of this, the accuracy of the forecast was poor and unsatisfying. The following are some examples of signal-based predictors: “PromoterScan” [12], which classified promoter sequences from non-promoter sequences using the extracted features of the TATA-box and a weighted matrix of transcription factor binding sites with a linear discriminator; “Promoter2.0” [13], which extracted the features from various boxes such as the TATA-Box, CAAT-Box, and GC-Box and then passed them to artificial neural networks. Ref. [14] utilized of TATA-Box and a classifier called a relevance vector machine (RVM). The content-based predictors rely on counting the frequency of k-mers by iteratively running a window of k-length length across the sequence. However, these approaches do not take into account the spatial information that is included in the sequences of base pairs. The following are some examples of content-based predictors: “PromFind” [15], which used the k-mer frequency to perform the hexamer promoter prediction; “PromoterInspector” [16], which recognized the regions containing promoters based on a common genomic context of polymerase II promoters by scanning for specific features defined as variable length motifs; and “MCPromoter1.1” [17]. In conclusion, GpG-based predictors made use of the position of GpG islands since the promoter region or the first exon region in human genes often includes GpG islands [18,19,20]. However, since only sixty percent of the promoters include GpG islands, the accuracy of predictions made using this form of predictor has never been higher than sixty percent. In recent years, techniques based on sequencing have been used in the process of promoter prediction. In order to accurately anticipate enhancer-promoter interactions [21] applied a variety of feature extraction algorithms. These tactics allowed them to obtain the most relevant sequence information. A sequence-based predictor, referred to as [22] proposed “PromoterPredict” in order to predict the strength of Escherichia coli promoters based on a dynamic multiple regression approach in which the sequences were represented as position weight matrices. This was done in order to determine whether or not a given promoter will be effective in activating transcription (PWM). In order to differentiate between promoter and non-promoter sequences, ref. [23] leveraged the variations in DNA sequence stability that exist between promoter and non-promoter sequences.
Machine Learning (ML) is an Artificial Intelligence (AI) application that allows autonomous learning and improvement from experience without explicit programming or knowledge of the learning environment. The primary objective is to design and build computational tools that can autonomously learn without human intervention. By definition, it must be able to adapt its behavior based on previous outcomes. ML has become one of the most important techniques for several genomics research objectives, including, the description and interpretation of huge genomic datasets, the Annotation of several sequence elements in genomics, and the prediction of genetic variation’s effect on DNA/RNA sequences. Diverse computational strategies for various research challenges have yielded excellent results in recent years [24,25,26,27]. A prediction based on sequences called “iPro70-PseZNC” [28] was developed to detect σ 70 promoters in prokaryotes. Samples of DNA sequences in the predictor were formed using a unique pseudo-nucleotide composition termed “PseZNC” which incorporates the six local DNA structural features and multi-window Z-curve composition. To extract the generic properties of bacterial promoters, a unique variable-window Z-curve approach was proposed to identify promoter sequences from two prokaryotic species: Escherichia coli and Bacillus subtilis [29]. A prediction tool, “iPromoter-FSEn” proposed a feature subspace-based ensemble classifier to extract features from DNA sequences and identify the σ 70 promoters [30]. To discover about σ 70 promoters in prokaryotes, “70ProPred” presented a position-specific trinucleotide propensity based on a single-stranded characteristic model [31]. An artificial neural network (ANN)-based tool, i.e., “SAPPHIRE” predicted σ 70 promoters in Pseudomonas sp. The sequence comparison to the 35 and 10 boxes of σ 70 promoters discovered experimentally in P. aeruginosa and P. putida is evaluated using “SAPPHIRE” [32]. A tool named “iPromoter-2L2.0” [33] proposed a two-stage model to identify promoter sequences and classify their type, in the first stage model predicts the given DNA sequence as a promoter and model the stage classifies the predicted promoter sequence as one of its class from σ 24 , σ 28 , σ 32 , σ 38 , σ 54 , and σ 70 . Kmer and PseKNC were used in “iPromoter-2L2.0” to extract discriminative features, while SVM was used as a classifier. “MULTiPly” [34], a new multilayer computational technique for predicting promoters and σ class from σ 24 , σ 28 , σ 32 , σ 38 , σ 54 , and σ 70 , was presented. “MULTiPly” considered both local information such as k-tuple nucleotide composition and dinucleotide-based autocovariance, as well as global information based on bi-profile Bayes and k-nearest neighbor feature extraction techniques. “iPromoter-BnCNN” [35] was presented for the accurate recognition and classification of six types of promoters: σ 24 , σ 28 , σ 32 , σ 38 , σ 54 , and σ 70 . It is a CNN-based model with three layers that uses parallel branching to incorporate local features related to the monomer nucleotide sequence: trimer nucleotide sequence, dimer structural features, and trimer structural attributes. “pcPromoter-CNN” [36] is a two-layer model that identifies the given sequence as a promoter or not a promoter and then classifies its σ type from σ 24 , σ 28 , σ 32 , σ 38 , σ 54 , and σ 70 . Recently a CNN-based tool “PromoterLCNN” [37] presented a two-layer model for bacterial promoter identification and classification from σ 28 , σ 32 , σ 38 , σ 54 , and σ 70 .
Recently, computational methods for identifying and classifying sigma promoters have made great progress. However, we conclude that the current approaches need the following improvements:
  • Most previous studies predicted promoter sequences from σ 70 . Classifying the anticipated promoter sequences from σ 54 is uncommon.
  • Not every study has produced a user-friendly and publicly accessible web server, making it difficult for most experimental scientists to use in practice.
  • The false-positive predictions of the above-mentioned studies are remarkable because of the imbalanced dataset.
  • The advancement of high-throughput whole-genome sequencing and the integration of verified promoter sequences has resulted in the development of databases such as “Pro54DB” [38], a database of σ 54 promoters. Therefore, there is a need for a computational model to identify σ 54 promoters, because databases play a vital role in development of computational tools.
To overcome these limitations, we propose a CNN-based model “iProm-Sigma54” to identify σ 54 promoter sequences. First, we built a benchmark dataset and performed a one-hot feature-encoding scheme to extract the feature representation vector. We used a variety of assessment criteria that are often applied in bioinformatics to evaluate the model’s performance. The Matthews correlation coefficient (MCC) and the accuracy, sensitivity, and specificity were computed. To further examine the model, we used five-fold cross-validation assessment measures. Additionally, the receiver operating characteristic (ROC) curve was computed. And at last, a web server was created in accordance with the proposed model. Figure 1 shows the flow diagram of the proposed model.

2. Benchmark Dataset

For the creation and evaluation of computational approaches, an empirically validated benchmark dataset is crucial. This research utilized the recently updated σ 54 promoter sequences from the “Pro54DB” (i.e., σ 54 promoter database), that contains prokaryotic promoters that have been experimentally confirmed; “Pro54DB” was created to collect data on σ 54 promoters. The present edition includes 210 experimentally proven σ 54 promoters with 297 regulated genes in 43 species, which were collected from 133 research articles. Each positive sequence in the dataset is 81 bp long. A negative dataset of promoters utilized by “iPromoter-2L2.0”, “MULTiPly”, “iPromoter-BnCNN”, “pcPromoter-CNN”, and “PromoterLCNN” was used for the same. Furthermore, this study utilized CD-Hit [39] with the threshold value set to 0.8 to minimize duplication and reduce homologous bias. Consequently, the benchmark dataset Ds is mathematically expressed in Equation (1):
D s = S e q + S e q
where S e q + represents the σ 54 promoter sequences and S e q depicts the negative sequences. We split the benchmark dataset Ds into further two sets called training and testing datasets and utilized 80% of the data for model training and five-fold cross-validation. Whereas 20% of the dataset was used to test the proposed model. Table 1 shows the parameters of the benchmark dataset.

3. Feature Encoding Scheme

A DNA sequence has four nucleotides (A, T, C, and G), and it must be translated into a numerical representation in order to execute computer operations. Thus, this study utilized a one-hot feature encoding scheme. Several recent cutting edge bioinformatics techniques have used this technique e [40,41,42,43]. Representation of each nucleotide for A, C, G, and T, characterized as follows:
A ( 1 , 0 , 0 , 0 )
C ( 0 , 1 , 0 , 0 )
T ( 0 , 0 , 1 , 0 )
G ( 0 , 0 , 0 , 1 )
Consequently, an (81,4) two-dimensional matrix can be used to represent each sample sequence.

4. Proposed Methodology

A 1D Convolutional Neural Network (1D-CNN) is a form of convolutional neural network used to handle one-dimensional signals such as time series data, audio signals, DNA/RNA sequence and text data expressed as word or character sequences. The convolutional layer of a 1D-CNN works on 1D filters rather than 2D filters as in standard CNNs for picture data. The filters move through the input sequence, calculating dot products between the filter weights and the local sequence components to generate a feature map that identifies certain patterns or features in the input sequence. 1D-CNNs, like standard CNNs, employ pooling layers to lower the spatial dimensions of the feature maps and generate more robust features. The pooling procedure, however, is performed along the time dimension of the feature maps rather than the spatial dimensions in a 1D-CNN. Finally, similar to standard CNNs, the extracted characteristics are input into one or more fully connected layers to create a prediction or classification. The 1D structure of the input data is used in these applications to efficiently learn meaningful representations and generate predictions. The most notable benefit of a CNN is that it does not need any previous feature information from data; instead, a model base on CNN can extract features directly from data. This study employed a CNN-based model to extract characteristics from a DNA sequence. CNN has demonstrated outstanding achievements in natural language processing, image processing [44,45] , and computational biology [46,47]. Several recent studies have performed hyperparameter tuning to improve the model for bioinformatics challenges. A grid search was performed on the training dataset to determine the optimal model, and six hyperparameters were tuned during the CNN learning process. Grid search is a hyperparameter tuning method used in AI to find the best set of hyperparameters for a model that optimize its performance on a given task. It is called grid search because it exhaustively searches over a specified hyperparameter grid, which is a predefined range of values for each hyperparameter. Table 2 lists the calibration ranges for each hyperparameter.

CNN Architecture

After selecting the optimal model from the grid search, we performed a one-hot feature extraction technique on the dataset and input it to the CNN. The CNN consists of two one-dimensional convolutional layers (i.e., Conv1D), followed by max-pooling layers and dropout layers; The filter sizes of Conv1D are 128 and 258 whereas the kernel sizes of 5 and 7, respectively. The pooling size in both max-pooling layers was 4, with strides of 2. Following the second dropout layer, a flattened layer is used. Following this, we utilized a dense layer using 64 nodes, followed by a dropout layer. For all three dropout layers, the dropout value was 0.5. For all Conv1Ds and dense layers, we utilized the rectified linear unit (ReLU) activation function, following equation defines the ReLU:
R e L U ( p ) = m a x ( 0 , p )
The ReLU activation function is a simple threshold function used in neural networks, where negative values are set to zero and positive values are unchanged. This function is computationally efficient, introduces non-linearity, and is less prone to the vanishing gradient problem compared to other activation functions. We utilized a dense layer as an output layer with a single node and sigmoid activation function. The neuron’s output will always range from 0 to 1 when its activation function is a sigmoid function and based on that score classify the input sequence as a promoter or non-promoter. The following are mathematical ways to express the sigmoid function:
S i g m o i d ( p ) = 1 1 + e p
where e is the base of the natural logarithm and p is the input value. The function has an “S” shaped curve, which is why it is often referred to as the “sigmoid” function. To prevent the model overfitting, we employed bias regularization and L2 regularization with a value of 0.0001 in the convolution and dense layers. Bias regularization is a technique in machine learning that aims to prevent overfitting by penalizing the magnitude of the bias values in a model. L2 regularization, also known as weight decay, is a commonly used method for regularizing the weights in a model. It involves adding a penalty term to the loss function. The penalty term encourages the model to have smaller weights, which can prevent overfitting and lead to a more generalized model. We utilized the “Keras class weights” technique to handle the imbalanced data problem. The Keras framework was used to develop and train the “iProm-Sigma54” model, which uses binary cross-entropy as its loss function.
H p ( q ) = 1 N p o s + N n e g [ i = 1 N p o s log ( p ( y i ) ) + i = 1 N n e g log ( 1 p ( y i ) ) ]
The binary cross-entropy function is given in Equation (4). Adam was utilized as an optimizer. There were 70 epochs in the batch, with a batch size of 32. Furthermore, an early stopping was imposed with a patience value of 30 to stop model training when the prediction accuracy doesn’t improve on the validation set. The proposed CNN model’s design is shown in Figure 2.

5. Results and Discussion

5.1. Evaluation Metrics

Five-fold cross-validation was used to assess the proposed model’s classification performance. This study employed four distinct measures to evaluate the effectiveness of “iProm-Sigma54”, which has previously been used in other state-of-the-art approaches. The specificity (Sp), sensitivity (Sn), accuracy (Acc), and MCC were the measurements. Sp computes the percentage of true negatives among all negatives (false positive rate), i.e., how well the model accurately recognizes negative cases. Whereas Sn measures the percentage of true positives among all positives (true positive rate), i.e., how well the model correctly identify positive cases and Acc measures the overall percentage of correct calculations (both true positive and true negative) among all cases. MCC is a measure that takes into account true and false positives and negatives and provides a balanced measure of the accuracy of binary classifications. It ranges from −1 to 1, with 1 indicating perfect prediction, 0 indicating random prediction, and −1 indicating perfect inverse prediction. The following are the mathematical expressions for these metrics
S p = P N P N + P P
S n = P P P P + F N
A c c = P P + P N P P + P N + F P + F N
M C C = P P × P N F P × F N ( P P + F P ) ( P P + F N ) ( P N + F P ) ( P N + F N )
In the above equations from Equations (5)–(8), PP denotes the number of true positives whereas, PN stands for true negatives, FP and FN denote the number of false positives, and false negatives, respectively.

5.2. Results and Comparison

Using five-fold cross-validation, the “iProm-Sigma54” prediction performance was assessed. The parameters utilized in this study were the same as those used to choose the best model, and ROC curves were also considered. During cross-validation, iProm-Sigma54 obtained an Acc of 95.45%, Sn of 96.53%, Sp of 90.64%, MCC of 0.848, and AUROC of 0.95; suggesting that the proposed predictor is capable of properly recognizing whether a query sequence is of the σ 54 promoter. Similar performance measures were performed using “iPro54-PseKNC”, “iPromoter-2L”, “iPromoter-BnCNN”, and “PromoterLCNN”, which are considered cutting-edge approaches for diagnosing and classifying σ promoters. Compared to results obtained using state-of-the-art techniques, “iProm-Sigma54” exhibited improved performance. A noticeable improvement by “iProm-Sigma54”, in terms of sensitivity and specificity, assures that the proposed model reduces the false positive values. A notable improvement of 5.8% in the value of MCC is evidenced that the proposed method accurately distinguishing between the promoter and non-promoter classes. Table 3 shows the iProm-Sigma54 performance and the comparison results with state-of-the-art methods. Figure 3a depicts the ROC curve for predicting the σ 54 promoter on the cross validation dataset; the curve evidently indicates a significant area under the curve.
A test dataset was used to determine the authenticity of iProm-Sigma54. it achieved an Acc of 98.40%, Sn of 95.12%, Sp of 97.19%, MCC of 0.9113 and AUROC of 0.97 on the test dataset. Table 4 shows the comparison results on test dataset. Figure 3b illustrate ROC curve for test dataset; the curve evidently indicates a significant area under the curve
For a more thorough examination, we investigated the effect of alternating nucleotides at each location on the performance score. Computational mutation scanning was conducted for each promoter sequence in the test set to determine whether changing each base of the input subsequence affected the results. Sequences with a length of 81 bp were considered and nucleotides at any position in the sequence were mutated individually. In addition, the absolute differences between the original and mutated sequence predictions were calculated and stored. The average expected score for all mutations was determined for all the sequences. Finally, the heat map shown in Figure 4 was generated using the average predicted score. As seen in the heat map, mutations in the middle of the sequence have a greater impact on the “iProm-Sigma54” CNN estimation than mutations in the beginning and ending nucleotides of the sequence, respectively. The results revealed that nucleotide changes had little impact on the outcome of “iProm-Sigma54” CNN recognition.
The learned motifs of the kernels of the first Conv1D can be considered position weight matrices (PWMs). Visualization of motifs offers a biologist further insight into the promoter regions. Figure 5 shows the instances of sequence motifs found using the proposed methodology.

6. Webserver

To make the suggested tool more accessible to the scientific community, The high-performance “iProm-Sigma54” tool is hosted on a web server that is accessible at http://nsclbio.jbnu.ac.kr/tools/iProm-Sigma54/. This approach has been used by several scholars [48,49,50]. Researchers and experts in medicine and bioinformatics can use “iProm-Sigma54” a simple-to-use tool. The guidelines to use the webserver are as follows:
  • It accepts input using two different methods: direct sequence input and file uploading with sequences with up to one thousand sequences in a FASTA format. To upload a file format must be “.fa”.
  • Set the threshold value ranges from 0–1.
  • Click on “submit Sequence” button for the prediction results
A web server snippet is shown in Figure 6, where Figure 6a displays writing query sequences whereas the Figure 6b shows the prediction results.

7. Conclusions

The study of the σ 54 promoter and its role in gene regulation is an important area of research in the field of bioinformatics, as it can provide insights into the mechanisms by which bacteria respond to their environment and the ways in which they are able to adapt to changing conditions. This information can be used to develop new strategies for controlling bacterial populations and for preventing the spread of bacterial infections. Therefore, precise identification of σ 54 promoter sequences is essential for understanding the underlying mechanisms of gene regulation. This study established the “iProm-Sigma54” model for this effect, which is based on a convolution neural network; a grid search algorithm was used to build a CNN-based predictor. A one-hot encoding scheme was used to generate the input matrix for the CNN. This study used five-fold cross-validation training and evaluated the results using a test dataset. The proposed method outperforms the most competitive methods in the literature. Motif and heatmap analyses were performed to provide more biological insights into the model. Finally, we made an online web server that would give convenience to other experimental scientists. In addition, this study provides a training model and dataset at GitHub: github.com/Shujaatmalik/iProm-Sigma54.

Author Contributions

Conceptualization, M.S., H.K., H.T. and K.T.C.; methodology, M.S.; software, M.S. and H.T.; validation, M.S., H.T. and K.T.C.; investigation, M.S., H.T. and K.T.C.; writing original draft preparation: M.S.; writing, review and editing, M.S., H.T., H.K. and K.T.C.; supervision H.T. and K.T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported by research funds for newly appointed professors of Jeonbuk National University in 2022.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study provides a training model and dataset at GitHub: github.com/ Shujaatmalik/iProm-Sigma54.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Ortiz-Merino, R.A.; Kuanyshev, N.; Byrne, K.P.; Varela, J.A.; Morrissey, J.P.; Porro, D.; Wolfe, K.H.; Branduardi, P. Transcriptional response to lactic acid stress in the hybrid yeast Zygosaccharomyces parabailii. Appl. Environ. Microbiol. 2018, 84, e02294-17. [Google Scholar] [CrossRef] [Green Version]
  2. Barrios, H.; Valderrama, B.; Morett, E. Compilation and analysis of σ54-dependent promoter sequences. Nucleic Acids Res. 1999, 27, 4305–4313. [Google Scholar] [CrossRef]
  3. Wigneshweraraj, S.; Bose, D.; Burrows, P.C.; Joly, N.; Schumacher, J.; Rappas, M.; Pape, T.; Zhang, X.; Stockley, P.; Severinov, K.; et al. Modus operandi of the bacterial RNA polymerase containing the σ54 promoter-specificity factor. Mol. Microbiol. 2008, 68, 538–546. [Google Scholar] [CrossRef]
  4. Kustu, S.; Santero, E.; Keener, J.; Popham, D.; Weiss, D. Expression of sigma 54 (ntrA)-dependent genes is probably united by a common mechanism. Microbiol. Rev. 1989, 53, 367–376. [Google Scholar] [CrossRef]
  5. Gardan, R.; Rapoport, G.; Débarbouillé, M. Expression of therocDEFOperon Involved in Arginine Catabolism inBacillus subtilis. J. Mol. Biol. 1995, 249, 843–856. [Google Scholar] [CrossRef]
  6. Zielinski, N.A.; Maharaj, R.; Roychoudhury, S.; Danganan, C.; Hendrickson, W.; Chakrabarty, A. Alginate synthesis in Pseudomonas aeruginosa: Environmental regulation of the algC promoter. J. Bacteriol. 1992, 174, 7680–7688. [Google Scholar] [CrossRef] [Green Version]
  7. Matsumine, H.; Yamamura, Y.; Hattori, N.; Kobayashi, T.; Kitada, T.; Yoritaka, A.; Mizuno, Y. A microdeletion of D6S305 in a family of autosomal recessive juvenile parkinsonism (PARK2). Genomics 1998, 49, 143–146. [Google Scholar] [CrossRef]
  8. Touzain, F.; Schbath, S.; Debled-Rennesson, I.; Aigle, B.; Kucherov, G.; Leblond, P. SIGffRid: A tool to search for sigma factor binding sites in bacterial genomes using comparative approach and biologically driven statistics. BMC Bioinform. 2008, 9, 73. [Google Scholar] [CrossRef]
  9. Kim, J.w.; Zeller, K.I.; Wang, Y.; Jegga, A.G.; Aronow, B.J.; O’Donnell, K.A.; Dang, C.V. Evaluation of myc E-box phylogenetic footprints in glycolytic genes by chromatin immunoprecipitation assays. Mol. Cell. Biol. 2004, 24, 5923–5936. [Google Scholar] [CrossRef] [Green Version]
  10. Dahl, J.A.; Collas, P. A rapid micro chromatin immunoprecipitation assay (ChIP). Nat. Protoc. 2008, 3, 1032–1045. [Google Scholar] [CrossRef]
  11. Lin, H.; Deng, E.Z.; Ding, H.; Chen, W.; Chou, K.C. iPro54-PseKNC: A sequence-based predictor for identifying sigma-54 promoters in prokaryote with pseudo k-tuple nucleotide composition. Nucleic Acids Res. 2014, 42, 12961–12972. [Google Scholar] [CrossRef] [Green Version]
  12. Prestridge, D.S. Predicting Pol II promoter sequences using transcription factor binding sites. J. Mol. Biol. 1995, 249, 923–932. [Google Scholar] [CrossRef]
  13. Knudsen, S. Promoter2.0: For the recognition of PolII promoter sequences. Bioinformatics 1999, 15, 356–361. [Google Scholar] [CrossRef] [Green Version]
  14. Down, T.A.; Hubbard, T.J. Computational detection and location of transcription start sites in mammalian genomic DNA. Genome Res. 2002, 12, 458–461. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Hutchinson, G. The prediction of vertebrate promoter regions using differential hexamer frequency analysis. Bioinformatics 1996, 12, 391–398. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Scherf, M.; Klingenhoff, A.; Werner, T. Highly specific localization of promoter regions in large genomic sequences by PromoterInspector: A novel context analysis approach. J. Mol. Biol. 2000, 297, 599–606. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Ohler, U.; Harbeck, S.; Niemann, H.; Reese, M.G. Interpolated markov chains for eukaryotic promoter recognition. Bioinformatics 1999, 15, 362–369. [Google Scholar] [CrossRef] [Green Version]
  18. Ioshikhes, I.P.; Zhang, M.Q. Large-scale human promoter mapping using CpG islands. Nat. Genet. 2000, 26, 61–63. [Google Scholar] [CrossRef]
  19. Davuluri, R.V.; Grosse, I.; Zhang, M.Q. Computational identification of promoters and first exons in the human genome. Nat. Genet. 2001, 29, 412–417. [Google Scholar] [CrossRef]
  20. Ponger, L.; Mouchiroud, D. CpGProD: Identifying CpG islands associated with transcription start sites in large genomic mammalian sequences. Bioinformatics 2002, 18, 631–633. [Google Scholar] [CrossRef] [Green Version]
  21. Yang, Y.; Zhang, R.; Singh, S.; Ma, J. Exploiting sequence-based features for predicting enhancer–promoter interactions. Bioinformatics 2017, 33, i252–i260. [Google Scholar] [CrossRef] [Green Version]
  22. Bharanikumar, R.; Premkumar, K.A.R.; Palaniappan, A. PromoterPredict: Sequence-based modelling of Escherichia coli σ70 promoter strength yields logarithmic dependence between promoter strength and sequence. PeerJ 2018, 6, e5862. [Google Scholar] [CrossRef] [Green Version]
  23. Kanhere, A.; Bansal, M. A novel method for prokaryotic promoter prediction based on DNA stability. BMC Bioinform. 2005, 6, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Khan, A.; Ilyas, T.; Umraiz, M.; Mannan, Z.I.; Kim, H. Ced-net: Crops and weeds segmentation for smart farming using a small cascaded encoder-decoder architecture. Electronics 2020, 9, 1602. [Google Scholar] [CrossRef]
  25. Shah, A.A.; Malik, H.A.M.; Mohammad, A.; Khan, Y.D.; Alourani, A. Machine learning techniques for identification of carcinogenic mutations, which cause breast adenocarcinoma. Sci. Rep. 2022, 12, 11738. [Google Scholar] [CrossRef]
  26. Shujaat, M.; Aslam, N.; Noreen, I.; Ehsan, M.K.; Qureshi, M.A.; Ali, A.; Naz, N.; Qadeer, I. Intelligent and Integrated Framework for Exudate Detection in Retinal Fundus Images. Intell. Autom. Soft Comput. 2021, 30, 663–672. [Google Scholar] [CrossRef]
  27. Zeng, L.; Liu, Y.; Yu, Z.G.; Liu, Y. iEnhancer-DLRA: Identification of enhancers and their strengths by a self-attention fusion strategy for local and global features. Briefings Funct. Genom. 2022, 21, 399–407. [Google Scholar] [CrossRef]
  28. Lin, H.; Liang, Z.Y.; Tang, H.; Chen, W. Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 16, 1316–1321. [Google Scholar] [CrossRef] [PubMed]
  29. Song, K. Recognition of prokaryotic promoters based on a novel variable-window Z-curve method. Nucleic Acids Res. 2012, 40, 963–971. [Google Scholar] [CrossRef] [Green Version]
  30. Rahman, M.S.; Aktar, U.; Jani, M.R.; Shatabda, S. iPromoter-FSEn: Identification of bacterial σ70 promoter sequences using feature subspace based ensemble classifier. Genomics 2019, 111, 1160–1166. [Google Scholar] [CrossRef]
  31. He, W.; Jia, C.; Duan, Y.; Zou, Q. 70ProPred: A predictor for discovering sigma70 promoters based on combining multiple features. BMC Syst. Biol. 2018, 12, 44. [Google Scholar] [CrossRef]
  32. Coppens, L.; Lavigne, R. SAPPHIRE: A neural network based classifier for σ70 promoter prediction in Pseudomonas. BMC Bioinform. 2020, 21, 415. [Google Scholar] [CrossRef] [PubMed]
  33. Liu, B.; Li, K. iPromoter-2L2. 0: Identifying promoters and their types by combining smoothing cutting window algorithm and sequence-based features. Mol. Ther.-Nucleic Acids 2019, 18, 80–87. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Zhang, M.; Li, F.; Marquez-Lago, T.T.; Leier, A.; Fan, C.; Kwoh, C.K.; Chou, K.C.; Song, J.; Jia, C. MULTiPly: A novel multi-layer predictor for discovering general and specific types of promoters. Bioinformatics 2019, 35, 2957–2965. [Google Scholar] [CrossRef]
  35. Amin, R.; Rahman, C.R.; Ahmed, S.; Sifat, M.H.R.; Liton, M.N.K.; Rahman, M.M.; Khan, M.Z.H.; Shatabda, S. iPromoter-BnCNN: A novel branched CNN-based predictor for identifying and classifying sigma promoters. Bioinformatics 2020, 36, 4869–4875. [Google Scholar] [CrossRef] [PubMed]
  36. Shujaat, M.; Wahab, A.; Tayara, H.; Chong, K.T. pcPromoter-CNN: A CNN-based prediction and classification of promoters. Genes 2020, 11, 1529. [Google Scholar] [CrossRef] [PubMed]
  37. Hernández, D.; Jara, N.; Araya, M.; Durán, R.E.; Buil-Aranda, C. PromoterLCNN: A Light CNN-Based Promoter Prediction and Classification Model. Genes 2022, 13, 1126. [Google Scholar] [CrossRef]
  38. Liang, Z.Y.; Lai, H.Y.; Yang, H.; Zhang, C.J.; Yang, H.; Wei, H.H.; Chen, X.X.; Zhao, Y.W.; Su, Z.D.; Li, W.C.; et al. Pro54DB: A database for experimentally verified sigma-54 promoters. Bioinformatics 2017, 33, 467–469. [Google Scholar] [CrossRef] [Green Version]
  39. Fu, L.; Niu, B.; Zhu, Z.; Wu, S.; Li, W. CD-HIT: Accelerated for clustering the next-generation sequencing data. Bioinformatics 2012, 28, 3150–3152. [Google Scholar] [CrossRef]
  40. Alam, W.; Tayara, H.; Chong, K.T. i4mC-Deep: An intelligent predictor of n4-methylcytosine sites using a deep learning approach with chemical properties. Genes 2021, 12, 1117. [Google Scholar] [CrossRef]
  41. Kim, J.; Shujaat, M.; Tayara, H. iProm-Zea: A two-layer model to identify plant promoters and their types using convolutional neural network. Genomics 2022, 114, 110384. [Google Scholar] [CrossRef] [PubMed]
  42. Shujaat, M.; Jin, J.S.; Tayara, H.; Chong, K.T. iProm-phage: A two-layer model to identify phage promoters and their types using a convolutional neural network. Front. Microbiol. 2022, 13, 1061122. [Google Scholar] [CrossRef] [PubMed]
  43. Oubounyt, M.; Louadi, Z.; Tayara, H.; Chong, K.T. DeePromoter: Robust promoter predictor using deep learning. Front. Genet. 2019, 10, 286. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Ilyas, T.; Khan, A.; Umraiz, M.; Kim, H. Seek: A framework of superpixel learning with cnn features for unsupervised segmentation. Electronics 2020, 9, 383. [Google Scholar] [CrossRef] [Green Version]
  45. Rashid, R.; Akram, M.U.; Hassan, T. Fully convolutional neural network for lungs segmentation from chest X-rays. In Proceedings of the International Conference Image Analysis and Recognition, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 71–80. [Google Scholar]
  46. Shah, A.A.; Alturise, F.; Alkhalifah, T.; Khan, Y.D. Deep Learning Approaches for Detection of Breast Adenocarcinoma Causing Carcinogenic Mutations. Int. J. Mol. Sci. 2022, 23, 11539. [Google Scholar] [CrossRef]
  47. Chipofya, M.; Tayara, H.; Chong, K.T. Drug Therapeutic-Use Class Prediction and Repurposing Using Graph Convolutional Networks. Pharmaceutics 2021, 13, 1906. [Google Scholar] [CrossRef]
  48. Chipofya, M.; Tayara, H.; Chong, K.T. Deep probabilistic learning model for prediction of ionic liquids toxicity. Int. J. Mol. Sci. 2022, 23, 5258. [Google Scholar] [CrossRef]
  49. Chantsalnyam, T.; Lim, D.Y.; Tayara, H.; Chong, K.T. ncRDeep: Non-coding RNA classification with convolutional neural network. Comput. Biol. Chem. 2020, 88, 107364. [Google Scholar] [CrossRef]
  50. Nazari, I.; Tayara, H.; Chong, K.T. Branch point selection in RNA splicing using deep learning. IEEE Access 2018, 7, 1800–1807. [Google Scholar] [CrossRef]
Figure 1. iProm-Sigma54 model flow diagram. We utilized σ 54 promoter sequences from the “Pro54DB” database and performed one-hot feature encoding scheme, after that we input encoded sequence to a optimal CNN model and later performed performance evaluation and finally made web server.
Figure 1. iProm-Sigma54 model flow diagram. We utilized σ 54 promoter sequences from the “Pro54DB” database and performed one-hot feature encoding scheme, after that we input encoded sequence to a optimal CNN model and later performed performance evaluation and finally made web server.
Cells 12 00829 g001
Figure 2. iProm-Sigma54 CNN Model Architecture. The CNN model consists of two one-dimensional convolutional layers, followed by max-pooling layers and dropout layers.
Figure 2. iProm-Sigma54 CNN Model Architecture. The CNN model consists of two one-dimensional convolutional layers, followed by max-pooling layers and dropout layers.
Cells 12 00829 g002
Figure 3. Receiver Operating Characteristic (ROC) curves The AUC ranges from 0.5 (for a random classifier) to 1 (for a perfect classifier): (a): ROC Curve on 5 fold (b): ROC Curve on Test Dataset.
Figure 3. Receiver Operating Characteristic (ROC) curves The AUC ranges from 0.5 (for a random classifier) to 1 (for a perfect classifier): (a): ROC Curve on 5 fold (b): ROC Curve on Test Dataset.
Cells 12 00829 g003
Figure 4. Computational mutation scanning was conducted for each promoter sequence in the test set to determine whether changing each base of the input subsequence affected. Red indicates a high frequency of mutations, while blue indicates a low frequency of mutations.
Figure 4. Computational mutation scanning was conducted for each promoter sequence in the test set to determine whether changing each base of the input subsequence affected. Red indicates a high frequency of mutations, while blue indicates a low frequency of mutations.
Cells 12 00829 g004
Figure 5. Motif Analysis: Learned motifs of the kernels of the first Conv1D can be considered position weight matrices (PWMs).
Figure 5. Motif Analysis: Learned motifs of the kernels of the first Conv1D can be considered position weight matrices (PWMs).
Cells 12 00829 g005
Figure 6. Webserver of iProm-Sigma54: (a) Adding Sequences for Prediction to Web server. The Query Sequence should be in Fasta format and must contain only A, C, G, T. (b): Predictor Output for Query Sequences. Table shows output for each query sequence.
Figure 6. Webserver of iProm-Sigma54: (a) Adding Sequences for Prediction to Web server. The Query Sequence should be in Fasta format and must contain only A, C, G, T. (b): Predictor Output for Query Sequences. Table shows output for each query sequence.
Cells 12 00829 g006
Table 1. Parameters of benchmark and test dataset.
Table 1. Parameters of benchmark and test dataset.
ClassBenchmark DatasetTest DatasetSequence Length
σ 54 Promoter1684281 bp
Non-Promoter228850081 bp
Table 2. Ranges of Parameters in Hyperparameter tuning.
Table 2. Ranges of Parameters in Hyperparameter tuning.
ParametersRange
Number of Conv1D[2, 3, 4, 5]
Filters Size in Conv1D[8, 12, 16, 22, 32, 42, 64, 128]
Kernel Sizes in Conv1D[2, 3, 4, 5, 6, 7, 8, 10, 12, 14]
Max-pooling Pool Size[2, 4, 6]
Max-pooling Stride length[2, 4]
Values of Dropout[0.2, 0.25, 0.3, 0.35, 0.4, 0.45, 0.5]
Neurons of Dense Layer[8, 16, 32, 64, 80, 100]
Table 3. σ 54 Promoter and non-promoter identification comparison using 5-Fold on benchmark.
Table 3. σ 54 Promoter and non-promoter identification comparison using 5-Fold on benchmark.
ModelAcc (%)Sn (%)Sp (%)MCC
iPromoter-2L94.0453.1999.570.65
iPro54-PseKNC78.5786.9670.190.58
iPromoter-BnCNN99.374.499.80.78
PromoterLCNN99.468.099.90.80
iProm-Sigma5495.4596.5390.640.858
Table 4. σ 54 Promoter and non-promoter identification comparison on a test dataset.
Table 4. σ 54 Promoter and non-promoter identification comparison on a test dataset.
ModelAcc (%)Sn (%)Sp (%)MCC
iPromoter-2L81.2392.2763.570.483
iPro54-PseKNC78.5297.5676.950.436
iPromoter-BnCNN92.9857.1495.230.516
iProm-Sigma5498.4095.1297.190.9113
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shujaat, M.; Kim, H.; Tayara, H.; Chong, K.T. iProm-Sigma54: A CNN Base Prediction Tool for σ54 Promoters. Cells 2023, 12, 829. https://doi.org/10.3390/cells12060829

AMA Style

Shujaat M, Kim H, Tayara H, Chong KT. iProm-Sigma54: A CNN Base Prediction Tool for σ54 Promoters. Cells. 2023; 12(6):829. https://doi.org/10.3390/cells12060829

Chicago/Turabian Style

Shujaat, Muhammad, Hoonjoo Kim, Hilal Tayara, and Kil To Chong. 2023. "iProm-Sigma54: A CNN Base Prediction Tool for σ54 Promoters" Cells 12, no. 6: 829. https://doi.org/10.3390/cells12060829

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop