Abstract

Medical diagnosis is always a time and a sensitive approach to proper medical treatment. Automation systems have been developed to improve these issues. In the process of automation, images are processed and sent to the remote brain for processing and decision making. It is noted that the image is written for compaction to reduce processing and computational costs. Images require large storage and transmission resources to perform their operations. A good strategy for pictures compression can help minimize these requirements. The question of compressing data on accuracy is always a challenge. Therefore, to optimize imaging, it is necessary to reduce inconsistencies in medical imaging. So this document introduces a new image compression scheme called the GenPSOWVQ method that uses a recurrent neural network with wavelet VQ. The codebook is built using a combination of fragments and genetic algorithms. The newly developed image compression model attains precise compression while maintaining image accuracy with lower computational costs when encoding clinical images. The proposed method was tested using real-time medical imaging using PSNR, MSE, SSIM, NMSE, SNR, and CR indicators. Experimental results show that the proposed GenPSOWVQ method yields higher PSNR SSIMM values for a given compression ratio than the existing methods. In addition, the proposed GenPSOWVQ method yields lower values of MSE, RMSE, and SNR for a given compression ratio than the existing methods.

1. Introduction

Modern learning reports that the employment of imaging has increased significantly in the previous two decades [1]. In the United States, for example, there was a study by Smith-Brinman et al. [2]. It was reported in 2008 that CCTV images doubled and MRI images doubled. This growth is the effect of a quick overhaul of hardware as well as software to support imaging systems. In addition, the transition from analog to digital provides fast, stress-free, and accurate image development. All of these factors have supplied significantly to the group of images. Therefore, the byte of these images is received every year worldwide. Imaging technology is implemented by X-ray, ultrasound, MRI/fMRI (functional magnetic resonance image), nuclear medicine, PET (positron emission tomography), CT (computerized tomography) [3], and DXA (dual energy X-ray absorptiometry). The necessity of saving, distributing, and downloading these images in their own appearance has given to the picture archiving along with communication system (PACS) a medical imaging skill that presents low-cost storage as well as access to useful images from many sources. Unluckily, the huge pixel size of these images severely limits the storage capacity [46]. For example, an emblematic NMR matrix dimension is 512 × 512 with 16 b [7].

A healthcare association that creates hundreds of images a week will need numerous megabytes of storage space. In a year, the total storage needed increases to a few gigabytes. This growth results in various terabit storage volumes per year. Moreover, a 56 kbps MRI band will take 12–15 minutes [8]. A lot of methods have been developed to reduce the size of these images, except that the major desirable novelty is the hybrid image production technique, in which the wavelet transform (DWT) along with vector quantization (VQ) as well as the loss [9] is preferred the most. In numerical analysis and function, DDD is a variation of spectacles in which the spectacles are precisely modeled and designed to process signals with a certain amount of time [10, 11].

In contrast to the discrete cosine transform (DCT) derived from the function of the base cosine, DWT is able to collect a few functions that meet the needs of multidimensional reaction analysis [12]. The origin of the image size reduction algorithm is to get an X input as well as produce an X demonstration that must need less bits for the representation. This went after recovery techniques that work on a compact X representation to produce a recovery , e.g., DWT offers a set of wavelet coefficients at various sizes showing various numerical types.

To propose the quantity for every set using the VQ technique, the bit value can significantly condense, devoid of affecting the excellence. Nevertheless, quantification leads to defeat of signal superiority. Consistent with Sannon, the dimensions of the selected vector have a significant influence on the quantitative efficiency. Larger vectors produce better quality than smaller vectors. For this reason, the purpose of this work is to develop a hybrid approach derived from discrete wavelet transforms and vector quantization to reduce the size of medical images with elevated compression rates and quality.

The main contributions of this paper are summarized as follows:(i)Recurrent neural network is developed to compress the medical images.(ii)A novel hybridized approach GENPSO is developed for optimizing the neural network parameters.(iii)Using GENPSO, the hidden layer of recurrent neural network is compressed in more effective way to increase the compression ratio.(iv)High compression ratio and better performance were achieved compared to the other existing approaches using wavelet vector quantization.(v)The proposed method provides good results for real-time medical dataset.

The manuscript of this document is organized as follows: Section 2 discusses some of the relevant contemporary literature. Section 3 presents a detailed description of the proposed architecture. Section 4 presents the experimental results, which include the general operating results of comparing the performance of GENPSO and other compression methods previously published. Conclusion and future work are provided in Section 5.

2. Background

Juliet et al. [13] discussed a new approach, including Ripplet conversion, to ensure better image quality and higher compression levels. In this work, to achieve high compression, the image is presented in various weights and instructions. Cyriac et al. [14] developed a lossless compression technique to address the issue of expanding compression. In this article, the newly developed method is to reduce the image size sufficiently as well as speed up the hardware execution of the software in real time. Brahimi et al. [15] illustrated a new coding method in which both symbols along with images are compressed together with a codec. The major point of this method is to insert a rotten stray signal into the corrupted image, and the resulting image is employed for image size reduction.

Arif et al. [16] introduced an effective approach for minimizing fluoroscopic imaging. The retrieved ROI of this approach is minimized through a grouping of run length as well as hepatic coding. The final outputs show that the newly developed method offers a compression ratio of 400% compared to other traditional approaches. Das et al. [17] described a lossless medical imaging watermark (MIW) procedure that relies on the area of interest concept. The major purpose of this technique is to offer clarification to many issues related to the dissemination of medical data. This article uses seven various methods to display as well as compare outputs to demonstrate that the newly developed approach is easy and obvious in ensuring the security of the medical database.

Lucas et al. [18] introduced 3D lossless compression techniques 3-D MRP for high-volume medical imaging equipment. The presented technique relies on the minimum velocity forecasting mechanism (MRP). This article concludes that the demonstrated method can enhance the probability of error of the MRP as well as achieve higher image size reduction efficiency than ASS along with other standards for the depth of the medical signal. Špelič et al. [19] proposed new algorithms developed as voxel compression algorithms for 3D CT image compression for the broadcasting of graphics information acquired from CT scanners. This work illustrated the Hounsfield size being employed first for partial medicine, and then image size reduction is applied. In this work, a modeling method is employed to evaluate the effectiveness of the newly developed techniques.

Anusuya et al. [20] developed a new, lossless code employing a codec to reduce the size of 3D brain images. In this paper, MRI models are employed to analyze the effectiveness of the newly developed solution, and this paper focuses on sinking utilization time using parallel calculations. Xio et al. [21] introduced integrated tablets for compact image compression without data loss. The proposed technique integrates with integer maps for inefficient compaction. Amri et al. [22] introduced two lossless image size reduction techniques. In this work, demonstrated algorithms are employed to minimize image dimension and code algorithms for lossless image size reduction. It can be seen that the newly developed method can maintain image feature for elevated image size reduction levels and also provide different enhancements to the standard JPEG image size reduction.

Ramesh et al. [23] described a method for predicting wavelengths to minimize medical images. In this method, the forecast equations for each subline depend on the correlation analysis. Evaluation outputs show that the newly developed method offers a superior level of compression matched up to the standard SPHIT and JPEG2000. Ibrahim et al. [24] introduced two novel lossless image size reduction techniques based on logarithmic calculations. The proposed method can offer better image feature compared to conventional DWT. Avramovic et al. [25] developed a new loss-free image size reduction technique using context-based Internet converters. The proposed method relies on the concept of projection to eliminate the need for transparency in the image and to effectively compress the image without losing data. The conclusion is that the newly developed method can attain the similar results as high-class images like previous standard algorithms. Bairagi [26] accounted the idea of symmetry for medical imaging. The method shown here is lossless and can effectively remove unnecessary information from the image.

Zuo et al. [27] introduced an enhanced technique to IMIC-ROI medical imaging for efficient and effective medical imaging. The proposed technique is based on the concept of interest area (ROI) as well as non-ROI area. Srinivasan et al. [28] described encoders for compact electromagnetism (EEG) matrix. The method illustrated first two phases, the lost code layer (SPHIT) and the remaining coded layer. Taquet et al. [29] reported on hierarchical-oriented forecasting methods to allow scaling without loss and near-loss of medical imaging. It can be seen that the newly developed method is best employed for lossless image size reduction as it may offer superior or equivalent PSNR for higher bit rates compared to the JPEG 2000 standard.

Zanaty and Ibrahim [30] introduced a medical image compression technique based on combining region growing and wavelets algorithms. The authors have used a region growing algorithm where an image is portioned into two parts and the capability of the algorithm is justified.

3. Materials and Methods

3.1. Recurrent Neural Network

The recurrent neural network [31] is one of the neural networks which employs precedent data in a closed loop. It can be repeated through merely transmitting the network output or intermediate state as input. For instance, the input image of k-1 might be integrated into a network in kth step. This kind of network is suitable during short-term dependencies as it depends on other and does not perform better for long-term dependencies observed in compression. For this kind of network, there are issues in the training procedure, so throughout reproduction it can “disappear” or “explode.” To avoid this limitation, this work uses a new approach called GenPSO and proposed vector quantization required to incorporate with neural network to manage the input and output image data. So in this work the codebook is generated using a novel hybrid GenPSO approach. Then this codebook is used in vector quantization for compressing the number of hidden neurons for increasing the compression ratio [32].

3.2. Updating Weights in Recurrent Neural Network Using GenPSO

In this step, the weight generated from the recurrent neural network is updated through the GenPSO method. The initial constitutive population of a chromosome has an evenly distributed random number. The chromosome represents the weight of the recurrent neural network. A crossover here can be defined as a crossover on a node. It is necessary to construct two springing from the parent node, with each node being retrieved from a hidden layer, and multiple outputs, randomly taken with equal probability. This node is considered a transit node [33]. The value of all input weights for that particular node is changed with the other master. This change can also be considered as a node change, in which all the incoming weights are for randomly selected active nodes. Finally, the PSO process is applied. The initial PSO population was determined by GA decision. Therefore, it can overcome the problem of slow convergence of the GPS network [34].

The algorithm for weights evolution by GenPSo in RNN is as follows.

The following steps are performed to determine the best weight value.(i)For population of the XiMi, a solution, i = 1, …, μ, is started on area M with Rn.(ii)Two parents are randomly selected with a parent population distribution and two births will be generated by a crossover operator.(iii)ereditary mutation of the offspring is performed.(iv)Repeat step (ii) until the number of offspring μo is μ. Otherwise, go to step (v).(v)Any parental decision, i = 1,…, μ and progeny Xo, o = 1,…, μ, are estimated in light of objective function (X).(vi)For mixed population Xm, m = 1,…, 2 μ, both the parent population and the population formed. They are randomly mixed so that parents and children mix up properly.(vii)Each solution of Xm, m = 1,…, 2 μ is evaluated with 10% of the solutions of the other randomly selected solutions from a mixed population.(viii)The highest earnings decision is reserved for the next generation of parents.(ix)If the best chromosome difference for the N number of successive generations is less than the specified brightness, stop the process and the best chromosome of the last generation is the ideal weight. Otherwise, proceed to the next step.(x)Beginning particle population and location and velocity are derived from the genetic algorithm.(xi)Estimate the fitness value of all particles depending on the fitness function based on the objective of the optimization problem.(xii)Compare the values of the fit of each particle and its holes. If the current value is better than pbest, then set the current fitness value to its pbest. Otherwise, the pbest will remain the same.(xiii)Determine the current best fitness value among all fitness values of all particles in the population. Then, compare this value to Gebel and set the current value to Gebel if the current value is better than Gebet. Otherwise, leave it blank.(xiv)Update the position and speed of each piece.(xv)Stop if the best solution is found that matches the predefined minimum error or reaches the maximum number of repetitions. Otherwise, repeat the procedure from steps (x) to (xv).

3.3. Proposed Image Compression Using GenPSO Wavelet Vector Quantization Recurrent Neural Network Approach

The proposed technique for quantifying vectors using backward neural networks is summarized in the following steps:

Step 1. Apply wavelet transform to the image to produce transformed image.

Step 2. Get a pixel value (0 to 255) of the matrix.

Step 3. Now apply these values to the forward neural network. This network must have 64 input nodes since the detector is 8 × 8. Train the recurrent neural network using the training algorithm as explained in Section 3.2.

Step 4. Weight and tilt, updated by GPS, are fed to a hidden layer that can contain 2, 4, 8, 16, 32, and 64 hidden nodes. Encode these values with GenPSOWVQ

Step 5. Select the 8 × 8 printer in order after completing the training.

Step 6. Digital bits are now converted into real values.

Step 7. Finally, the call phase shows depressed images (output images) at the output of the nervous system’s output layer. The conversion of a pixel to a true value and a true value to a pixel occurs during the compression and compression process.

4. Experimental Results

4.1. Dataset Used

Datasets of images which include chest radiographs, unenhanced brain CTs, mammograms, and abdominal CTs are used for experimenting this work. The performance of the proposed Int OPMICM method is evaluated using real-time medical images in terms of PSNR, MSE, SSIM, NMSE, SNR, and CR values [31]. These results are compared with those obtained by the existing algorithms, namely, FFNN, VQFFNN, optimized FFNN, and optimized VQFFNN. Figure 1 depicts the sample images from the real-time medical database [34].

4.2. Performance Metric

For the assessment of the compression approaches, this paper employs six metrics, namely, PSNR, MSE, SSIM, NMSE, SNR, and CR.

4.2.1. Peak Signal-to-Noise Ratio

The peak signal-to-noise ratio (PSNR) is used to evaluate the quality between the compressed image and the original image. The PSNR formula is defined as follows:where H and W are the height and width of the image, respectively, and f(x, y) and (x, y) are the grey levels located at coordinate (x, y) of the original image and compressed image, respectively.

4.2.2. Structural Similarity Index

The structural similarity index is a method for measuring the similarity between the compressed image and the original image.where is the compressed image, Y is the original image, μ is the mean, and σ2 is the variance.

4.2.3. Mean Square Error

The mean square error (MSE) is used to evaluate the difference between a compressed image and the original image. The MSE can be calculated bywhere is the compressed image and Y is the original image.

4.2.4. Root Mean Square Error

The root mean square error (RMSE) is a frequently used measure of the difference between compressed image values and the original image values.where is the compressed image and Y is the original image.

4.3. Experimental Analysis
4.3.1. Experiment No. 1: Analysis of the Proposed GenPSOWVQ Compression Approach

In this experiment, we will evaluate the contribution of the proposed GenPSOWVQ compression approach. To evaluate the performance of this scheme, the PSNR, MSE, SSIM, NMSE, and SNR against 25% CR ratio are employed. This is shown in equations (1)–(4) correspondingly. This experiment will introduce the concept of GenPSOWVQ compression. It will be performed by estimating the performance of this scheme against a 25% CR ratio. Table 1 lists the PSNR, MSE, SSIM, NMSE, and SNR measures of GenPSOWVQ as well as the existing approaches.

The proposed GenPSOWVQ approach’s performance is compared with traditional compression approaches through varying metrics, and the result is illustrated in Figure 2. The performance ratio of GenPSOWVQ has higher value than with traditional compression approaches. So it is proved that GenPSOWVQ is superior to traditional compression approaches.

4.3.2. Experiment No. 2: Performance of Best Chromosome Particles of GenPSO with Population Size of 50

The designed 2-2-1 feed forward architecture and the error value were developed according to the GenPSO method set out above. The population size was 50, and the termination rule is if the best error in the chromosome in the 50 successive generations is less than 0.00001. It is described in Figure 3.

The proposed GenPSO approach’s performance is compared with varying population size, and the result is illustrated in Figure 3. The performance ratio of GenPSO has higher value than with traditional compression approaches. So it is proved that GenPSOWVQ is superior to traditional compression approaches.

4.3.3. Experiment No. 3: Performance of Best Chromosome Particles of GenPSO with Population Size of 100

The architecture of the best chromosome particle [35] distribution 2–2–1 was developed using the GenPSO method described above. The population size was 100, and the termination rule is if the 100 best chromosome errors continue to be less than 0.00001. It is described in Figure 4.

The proposed GenPSO approach's performance is compared with varying population size of 100, and the result is illustrated in Figure 4. The performance ratio of GenPSO has higher value than with traditional compression approaches. So it is proved that GenPSOWVQ is superior to traditional compression approaches.

4.3.4. Experiment No. 4: Performance of Error of GenPSO with Population Size of 50

The designed 2-2-1 feed forward architecture and the error value were developed according to the GenPSO method set out above. The population size was 50, and the termination rule is if the best error in the chromosome in the 50 successive generations is less than 0.00001. It is described in Figure 5.

The proposed GenPSO approach's performance is compared with varying population size of 50, and the result is illustrated in Figure 5. The performance ratio of GenPSO has higher value than with traditional compression approaches. So it is proved that GenPSOWVQ is superior to traditional compression approaches [36]. The proposed GenPSO achieves high performance against a large population size. This shows that it is superior to traditional compression approaches.

5. Conclusion

This article presents an image compression method for medical image compression using recurrent neural network. This recurrent neural network used a new image compression scheme called the GenPSOWVQ method which has been introduced using GenPSO with wavelet VQ. The codebook is built using a combination of fragments and genetic algorithms. The proposed method was tested using real-time medical imaging using PSNR, MSE, SSIM, RMSE, SNR, and CR indicators. Experimental results show that the proposed GenPSOWVQ method yields higher PSNR SSIMM values for a given compression ratio than the existing methods. In addition, the proposed GenPSOWVQ method yields lower values of MSE, RMSE, and SNR for a given compression ratio than the existing methods. This work can be further expanded to achieve the goal of creating optimal rules. The work can be extended for security in case of image compression. Like the future work for this work, a new code of image compression is proposed using ambiguous logic. This research work has to be tested on large datasets and on a dynamic huge sample that is not tested and storage optimization that is not considered. The article presents a method for image compression using a recurrent neural network known as GenPSO. This method is built using a combination of genetic algorithms and fragments. The test results indicated that the proposed method achieves higher PSNR SSIMM values than the existing methods.

Data Availability

The data that support the findings of this study are available upon request from the corresponding author.

Conflicts of Interest

The authors of this manuscript declare that they do not have any conflicts of interest.

Acknowledgments

The authors extend their gratitude to the deanship of scientific research at King Khalid University for funding this work through research group program under Grant no. R.G.P. 2/48/42.