Научная статья на тему 'Acute ischemic stroke lesion segmentation in non-contrast CT images using 3D convolutional neural networks'

Acute ischemic stroke lesion segmentation in non-contrast CT images using 3D convolutional neural networks Текст научной статьи по специальности «Физика»

CC BY
184
25
i Надоели баннеры? Вы всегда можете отключить рекламу.
Журнал
Компьютерная оптика
Scopus
ВАК
RSCI
ESCI
Область наук
Ключевые слова
ischemic stroke / brain / non-contrast CT / segmentation / CNN / 3D U-Net

Аннотация научной статьи по физике, автор научной работы — Dobshik A.V., Verbitskiy S.K., Pestunov I.A., Sherman K.M., Sinyavskiy Yu.N.

In this paper, an automatic algorithm aimed at volumetric segmentation of acute ischemic stroke lesion in non-contrast computed tomography brain 3D images is proposed. Our deep-learning approach is based on the popular 3D U-Net convolutional neural network architecture, which was modified by adding the squeeze-and-excitation blocks and residual connections. Robust pre-processing methods were implemented to improve the segmentation accuracy. Moreover, a special patches sampling strategy was used to address the large size of medical images and class imbalance and to stabilize neural network training. All experiments were performed using five-fold cross-validation on the dataset containing non-contrast computed tomography volumetric brain scans of 81 patients diagnosed with acute ischemic stroke. Two radiology experts manually segmented images independently and then verified the labeling results for inconsistencies. The quantitative results of the proposed algorithm and obtained segmentation were measured by the Dice similarity coefficient, sensitivity, specificity and precision metrics. The suggested pipeline provides a Dice improvement of 12.0 %, sensitivity of 10.2 % and precision 10.0 % over the baseline and achieves an average Dice of 62.8 ± 3.3 %, sensitivity of 69.9 ± 3.9 %, specificity of 99.7 ± 0.2 % and precision of 61.9 ± 3.6 %, showing promising segmentation results.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «Acute ischemic stroke lesion segmentation in non-contrast CT images using 3D convolutional neural networks»

Acute ischemic stroke lesion segmentation in non-contrast CT images using 3D convolutional neural networks

A.V. Dobshik1, S.K. Verbitskiy1, I.A. Pestunov12, K.M. Sherman 3, Yu.N. Sinyavskiy2, A.A. Tulupov 3, V.B. Berikov1,4 1 Novosibirsk State University, 630090, Russia, Novosibirsk, Pirogova street 1;

2 Federal Research Center for Information and Computational Technologies, 630090, Russia, Novosibirsk, Academician M.A. Lavrentiev avenue 6;

3International Tomography Center SB RAS, 630090, Russia, Novosibirsk, Institutskaya str. 3a;

4 Sobolev Institute of Mathematics SB RAS, 630090, Russia, Novosibirsk, Academician Koptyuga avenue 4

Abstract

In this paper, an automatic algorithm aimed at volumetric segmentation of acute ischemic stroke lesion in non-contrast computed tomography brain 3D images is proposed. Our deep-learning approach is based on the popular 3D U-Net convolutional neural network architecture, which was modified by adding the squeeze-and-excitation blocks and residual connections. Robust pre-processing methods were implemented to improve the segmentation accuracy. Moreover, a special patches sampling strategy was used to address the large size of medical images and class imbalance and to stabilize neural network training. All experiments were performed using fivefold cross-validation on the dataset containing non-contrast computed tomography volumetric brain scans of 81 patients diagnosed with acute ischemic stroke. Two radiology experts manually segmented images independently and then verified the labeling results for inconsistencies. The quantitative results of the proposed algorithm and obtained segmentation were measured by the Dice similarity coefficient, sensitivity, specificity and precision metrics. The suggested pipeline provides a Dice improvement of 12.0 %, sensitivity of 10.2 % and precision 10.0 % over the baseline and achieves an average Dice of 62.8 + 3.3 %, sensitivity of 69.9 + 3.9 %, specificity of 99.7 + 0.2 % and precision of 61.9 + 3.6 %, showing promising segmentation results.

Keywords: ischemic stroke, brain, non-contrast CT, segmentation, CNN, 3D U-Net.

Citation: Dobshik AV, Verbitskiy SK, Pestunov IA, Sherman KM, Sinyavskiy YuN, Tulupov AA, Berikov VB. Acute ischemic stroke lesion segmentation in non-contrast CT images using 3D convolutional neural networks. Computer Optics 2023; 47(5): 770-777. DOI: 10.18287/2412-6179-C0-1233.

Introduction

Acute cerebral circulation disorder or stroke is a disease with high rates of morbidity and mortality worldwide. According to the American Heart Association, the most common type of stroke is ischemic [1]. Early diagnosis of stroke is crucial for treatment choice [2, 3], because tissue changes in the ischemic penumbra may be reversible, especially in the early stages [4, 5]. The choice of diagnostic methods in each specific case strongly depends not only on its applicability (availability, contraindications, patient's condition, etc.), but also on the time of symptoms onset [6]. Any delay in medical care increases the risk of severe consequences and death.

Neuroimaging is fundamental to most modern methods of differential diagnosis of acute stroke [5, 7 -12]. For some imaging procedures, contrast injection is required, which is related to the risk of complications and a number of contraindications. The analysis of Computed Tomography (CT) and Magnetic Resonance Imaging (MRI) scans is an integral part of the examination guidelines for patients with signs and symptoms of acute stroke [7 - 9].

CT is the most common diagnostic tool for acute ischemic stroke (AIS) due to its availability (a large number of screening centers, low cost, no contraindications and low body burden) and short screening duration [7, 8, 13]. Non-contrast CT (NCCT) was first used in the evaluation of AIS patients in 1995 to detect intracranial hemorrhage (hemorrhagic stroke) and select a treatment strategy at an early time window (within 3 hours of symptoms onset) [15, 16] and proved to be efficient [13, 14, 17]. When diagnosing stroke, NCCT demonstrates relatively high specificity and low sensitivity, however, it allows detecting blood clots in cerebral vessels and intracranial hemorrhage (hemorrhagic stroke) which are absolute contraindications for some strategies of AIS treatment [9, 17].

Interpretation of NCCT scans is associated with certain difficulties, since early AIS changes look like areas of slightly reduced density, which the human eye due to various factors cannot always distinguish from the normal tissues [3, 7, 19]. In addition, images often show artefacts (caused by patient movements or imaging camera) that may look like strokes [8]. Therefore, developing of automated procedures for localization and assessment of AIS tissues volume based on NCCT scans

[19 - 26] (including the procedures based on convolutional neural networks (CNNs) [19 - 22]) is an urgent task. Randomized controlled trials [19, 27 - 29] have shown that CNN-based methods are comparable to radiologists in terms of sensitivity, specificity, overall accuracy, AUC, ROC, etc. It allows using CNNs as auxiliary tool in clinical practice. In addition, automated processing based on 3D CNNs allows volumetric analysis taking into account the spatial context and the rapid detection of small ischemic foci.

Over recent years in the medical image segmentation task CNNs have achieved state-of-the-art results [30, 31]. It is due to the convolution operation and its weights capable of learning complicated structures and patterns at multiple scales in the data. The U-shaped encoderdecoder CNN architecture 2D U-Net [32] has a wide application [33 - 35]. In [33], a self-adapting framework where 2D U-Net is adopted to segment various organs is proposed. For the analysis of volumetric images, 3D CNNs were introduced [36 - 38]. 3D U-Net [36] shows better performance in comparison with 2D U-Net. In [39], a slightly modified 3D U-Net is utilized for brain tumor segmentation in MRI scans. In [40] DeepMedic CNN architecture [38] is used for stroke lesions segmentation and post-processing techniques are investigated. In [33], cascaded 3D U-Net is proposed, which overcomes the disadvantage of 3D U-Net for datasets with large image sizes, but it requires training two neural networks. U-Net is often improved for a specific task using various architecture choices (e. g., Squeeze-and-Excitation blocks [34, 41 - 43], attention mechanisms, and computer vision transformers [44]).

In this work, we present a neural network algorithm for the volumetric segmentation of acute ischemic stroke lesions on NCCT brain images. We use a 3D CNN based on 3D U-Net architecture [36], which we modify with residual connections and relatively new Squeeze-and-Excitation blocks [41]. To achieve better results, we implement robust pre-processing techniques, a particular patch extraction strategy, and a weighted loss function, which are aimed at mitigating the class imbalance. We perform five-fold cross-validation and compare the results of experiments by measuring Dice, sensitivity, specificity, and precision metrics.

In the rest of the paper, we describe the used dataset and the method (including data pre-processing techniques and patches sampling strategy), the neural network architecture, and training and testing procedures. In the end, we summarize our work and discuss future plans.

1. Materials

The dataset used for our study contains volumetric non-contrast CT head scans of 81 patients diagnosed with acute ischemic stroke. The CT images were made by the Philips Ingenuity CT scanner and stored in medical DICOM format. The data were acquired from the International Tomography Center SB RAS. All

volumetric images have the same resolution of 512*512, but a different number of slices varying from 306 to 505, depending on the patient. For all images, slice thickness and spacing between slices are 1 and 0.5 mm, respectively. Also, the DICOM attribute pixel spacing, that is, the physical distance in mm between pixel centers, ranges from 0.38 to 0.5 for different volumes. For each volumetric image, corresponding manual segmentation is available. All segmentations were performed by two radiology experts (specialists in magnetic resonance imaging and X-ray computed tomography with 9 - 13 years experience, PhD in Radiology and Radiation Therapy) using 3D Slicer [45]. It is worth mentioning that the number of voxels corresponding to the area affected by acute ischemic stroke is 0.8 % of the total number of voxels in our dataset, that is, the data under study is highly imbalanced.

2. Methods 2.1. Data pre-processing

One crucial step in all deep-learning approaches is data pre-processing, which ensures data consistency. Training CNN directly on data without pre-processing leads to poor performance, as will be shown below. First, the intensities of CT images were thresholded between 0 and 80 Hounsfield units [46]. This transformation eliminates most of the artifacts and high-intensity tissues and remains visible such important parts of the brain as soft tissue, white matter, and gray matter. Second, the skull and coil were removed on each slice, leaving only the brain area on the CT image. This transformation was performed using connected component analysis. In particular, extraction of the largest connected component, then assigning zero values to pixels with the highest intensity, which correspond to the region of the skull, and re-extraction of the largest connected component. Third, each volume was normalized. We apply min-max normalization:

X — X

_ ^ ^ min

where Xmm and Xmax are the minimum and the maximum intensity values (X) of 3D image, respectively. Such a normalization rescales values to [0, 1]. We also conducted an experiment applying standardization before min-max normalization:

X=Xzl,

s

where | is the mean and s is the standard deviation of the brain region. Standardization was applied to each volumetric CT image of each patient independently. Also, the non-brain region was set to 0. Such a technique gives comparative intensity values in the brain area while invariant to the size of the background part. Fourth, the volumes were cropped to the non-zero region to dispose of a large uninformative background area. An example of a pre-processed image is shown in Fig. 1.

Fig. 1. (a) CT image after the first pre-processing step, (b) after preprocessing

2.2. Patch-based approach

The main difficulty related to 3D medical images is their large size, while computing resources are always limited. For example, a CT scan from our dataset has on average 512x512x405 = 106168320 voxels (1.04 GB of memory for 32-bit voxels), where 405 is the average number of slices for a patient. It is worth noting that if the dataset is anisotropic, that is, the resolution of different axes of an image is not the same, resizing methods are not advisable. Different interpolation techniques can deform the physical object and remove small details.

An optimal solution is to extract small parts, so-called patches, of 3D images and use them as input to a neural network. We use the uniform sampler, which extracts patches randomly from the whole volume with uniform distribution, and the weighted sampler, which extracts patches from different parts of the volume with a given non-uniform distribution. In our case, the weighted sampler extracts a patch with its center in the area of the healthy brain tissue with a probability of 0.5 and with its center in the area affected by the stroke with a probability of 0.5. The probability of the background as a patch's center is set to 0. During one training epoch, the same number of patches set to 32 are extracted for each patient. The patch size is set to 128x128x128.

2.3. Neural network architecture

Our deep learning algorithm builds on the encoderdecoder 3D U-Net [36] convolutional neural network, which we specially modify for our objective. The architecture of the neural network is shown in Figure 1. The input size is 128x128x128. Each convolutional block of encoder and decoder consists of two 3D convolutions with the size of kernel 3x3x3 and stride of 1x1x1, where each of them is followed by a normalization layer and activation function. We adopt LeakyReLU as an activation function since it showed better results in our experiments in comparison with more commonly used ReLU.

It is known that batch normalization [47] is strongly contingent on the current batch statistics during training. The statistics can include some noise, depending on the input examples. Therefore, it requires a sufficiently large batch size and also a large size of the training set. However, due to the high memory usage of 3D convolutions when employing large patches, we are

limited to a maximum batch size of 2. Thus, we apply instance normalization, which solves this issue and shows better performance in our task than batch normalization.

Trilinear upsampling is used instead of the more traditional transposed convolution operator in the decoder part. In our research, we observe similar results, but the upsampling operator has no trainable parameters, so the occupied memory can be reduced.

Residual connections mitigate the problem of vanishing and exploding gradients when training deep neural networks. It was first proposed in ResNet [48]. We integrate this technique in each convolution block of the contracting and expansive paths. The input is processed by 3D convolution with a kernel size of 1x1x1, so the element-wise addition is possible. The architecture diagram of the convolutional block of the encoder with residual connection is illustrated in Fig. 3. The parameters in brackets after Conv3D are the number of input channels, output channels, and kernel size. The decoder block is similar except for trilinear upsampling that halves the number of channels instead of 3D convolution that doubles the number of channels, therefore we leave aside the decoder block image.

We insert the Squeeze-and-Excitation (SE) mechanism [41] in our CNN as it shows strong performance in many computer vision tasks. It squeezes the global spatial information into a channel descriptor, captures inter-dependencies of all channels, and then recalibrates the feature maps to accentuate relevant channels. In our case, it can help to learn where the affected area is located by strengthening the informative features and suppressing the weak ones.

Channel-wise global average pooling is applied to the input tensor TeRCxDxHxW of the SE block. Then the obtained vector UeRC is processed by the excitation mechanism:

S = ct(w2 •8(W1 -U)), S eRC

where Wx e R(C/r)xC is a linear layer reducing the number of dimensions of the vector U, 8 is the ReLU activation function, W2 e RCx(C/r) is a linear layer increasing the number of dimensions of the vector U, ct is the sigmoid function, and r is a parameter. In the end, a channel-wise multiplication between each element si of the vector S and the input tensor T is performed. SE module is integrated into each convolutional block. Its particular location is shown in Fig. 2.

2.4. Training procedure

The obtained segmentation maps from the last CNN layer are transformed by the sigmoid function to get the probabilities of classes. Then the neural network weights are optimized using the sum of binary cross-entropy (BCE) and soft Dice [49] loss functions. We also perform experiments with weighted BCE loss since it addresses the class imbalance problem:

N

L=Yw' (-y logp- - ( - y ) log ( - p> )) ■

1=1

N

2 Y=iP>y>+e

i—

W' =

ZN 2 -s-^N 2

, =P +Y '^

No y + Ni (1 - y ) N '

+ e

where N0 is the number of background pixels of all training images, N1 is the number of pixels related to the

affected area, N=N0+N1, yi and pt denote the ground truth and the confidence score of the model for pixel i, respectively.

2.5. Training, implementation and testing details

We train our networks using the following setting: the batch size of 2, Adam optimizer [50] with the initial learning rate of 1E-4, which is reduced by factor 2 every 5000 iterations until the learning rate is 1.25E-5, L2 weight decay of 1E-5. The reduction ratio r of SE is 16. Each model was trained for a total of 40000 iterations.

D>

Down Conv Block MaxPool 2x2x2 UpSample Up Conv Block [> Conv3D 1x1x1 —► Concatenation

256

Fig. 2. CNN architecture

Input BxCxDxHxW

Conv3D (C, C, 3)

+

lnstanceNorm3D

i

LeakyReLU Conv3D (C, 2C, 1)

+ +

Conv3D (C, 2C, 3) lnstanceNorm3D

+

lnstanceNorm3D

+

-'

LeakyReLU

T

Fig. 3. Convolutional block of the encoder

The proposed algorithm was implemented in Python 3.7 using Py Torch 1.12.0 machine learning framework. We also employ TorchIO 0.18.80 [51] Python library for data pre-processing. All trainings were conducted on NVIDIA Tesla T4 GPU with 16 GB of memory.

All patients were randomly split into five parts, so all experiments were performed using five-fold cross-validation. To evaluate our results, we measure Dice similarity coefficient (DSC), sensitivity, specificity, and precision metrics, which are most significant in medical image segmentation. The use of these metrics, in addition to the main metric DSC, allows us to evaluate aspects of the behavior of segmentation methods in conditions of unbalanced samples. Results are presented as mean + standard deviation. DSC is calculated similarly to the soft

Dice loss but is not subtracted from 1 and thresholded values are used instead of confidence scores. To get the predicted binary segmentation mask, the obtained probabilities were binarized according to the threshold of 0.5. On the test set, patches were extracted across a whole volume with an overlap of 25 % over a grid. The predictions in the overlapping areas were averaged.

3. Experiments and results

The quantitative comparison of our methods is presented in Tab. 1. The pre-processing of the first experiment included only thresholding of the Hounsfield units and the min-max normalization. Also, patches were derived using the uniform sampler. The results of the second experiment show that adding the robust preprocessing techniques, including selecting the brain tissue and cutting to the non-zero region, increases DSC by 0.3 %. The restricted patch extraction method performed by the weighted sampler during training improves performance by 3.6 % of DSC and 4.9 % of sensitivity. The main proposed modifications of 3D U-Net are residual connections and SE. While the integration of the residual connections improves the DSC by 2.7 %, the consistent inserting of the SE modules gives a 3.8 % of DSC increase, also showing the promising sensitivity value of 68.9 + 5.9 %. The final algorithm configuration maintains the above-described improvements, while also standardization before min-max normalization and training using the weighted loss function are performed. Such a configuration gives the best results, 62.8 + 3.3 %

of DSC and 69.9 + 3.9 % of sensitivity and improves DSC by 12.0 %, sensitivity by 10.2 %, and precision by 10.0 % over the baseline.

We also study how the patch overlapping size impacts the segmentation accuracy when performing testing. Tab. 2 shows the quantitative assessment with 25 %, 50 %, and 75 % path overlapping using the final algorithm configuration. Although the overlapping of

Method

75 % gives the best results relying on DSC, it reduces sensitivity, and the running time of the algorithm per one patient is increased by 10 times.

We tried to insert in the training process various augmentation techniques, such as random affine transformation, random flip, random elastic deformation, random gamma intensity transformation, but it worsened the performance of our algorithm.

Tab. 1. Cross-validation results

DSC (%) Sensitivity (%) Specificity (%) Precision (%)

No pre-processing (baseline) 50.8 + 3.7 59.7 + 9.0 99.8 + 0.1 51.9 + 5.7

Pre-proccesing 51.1 + 2.3 51.8 + 3.5 99.8 + 0.1 61.7 + 3.8

Pre-processing + weighted sampler 54.7 + 2.5 56.7 + 7.6 99.8 + 0.2 61.6 + 6.9

Pre-processing + weighted sampler 57.4 + 1.6 60.6 + 6.4 99.8 + 0.1 61.5 + 4.9

+residual connections

Pre-processing + weighted sampler 61.2 + 6.4 68.9 + 5.9 99.7 + 0.2 61.9 + 4.4

+ residual connections+ SE

Pre-processing + standardization 62.8 + 3.3 69.9 + 3.9 99.7 + 0.2 61.9 + 3.6

+weighted sampler +residual connections

+ SE +weighted loss

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Tab. 2. The comparison of the patch overlap sizes

Overlap Time per DSC Sensitivity (%) Specificity (%) Precision (%)

size, % patient, s (%)_

25 45 62.8 + 3.3 69.9 + 3.9 99.7 + 0.2 61.9 + 3.6

50 90 62.2 + 2.2 69.3 + 4.7 99.7 + 0.2 62.9 + 3.6

75 440 63.0 + 2.5 69.3 + 4.9 99.7 + 0.2 63.8 + 3.0

CT axial plane

Manual segmentation U-Net w/o pre-processing Our proposed model

Fig. 4. Qualitative segmentation results

Qualitative results are presented in Fig. 3. Several axial plane slices of the 3D images are from the validation set.

Conclusion

In this work, we presented and evaluated an automatic algorithm for the segmentation of acute ischemic stroke lesion in non-contrast computed tomography brain images. Our deep learning approach is based on the 3D U-Net convolutional neural network. As far as we know, no previous research has investigated the combination of 3D U-Net, Squeeze-and-Excitation blocks, residual connections, special patch extraction

technique, and weighted loss function for solving the problem of acute ischemic stroke lesion segmentation on NCCT images. We have experimentally shown that the suggested pipeline for NCCT images segmentation, as compared to other state-of-the-art methods, improved segmentation performance by implementing robust pre processing techniques, Squeeze-and- Excitation blocks, and residual connections. Also, special patch extraction technique and weighted loss function increased the segmentation accuracy under the strong imbalance of classes. Our proposed model showed an average Dice of 62.8 + 3.3 %, sensitivity of 69.9 + 3.9 %, specificity of 99.7+0.2 %, and precision of 61.9 + 3.6 %. High specificity values are caused by the dominance of the negative class (non-affected tissue) over the positive class (pathological area) in the sample. The final improvement of the suggested pipeline over the baseline amounted to 12.0 % of DSC, 10.2 % of sensitivity, and 10.0 % of precision.

The method can be used by radiologists in delineating between damaged and healthy brain tissue and deciding on further treatment. In particular, it can help correct inaccuracies in their stroke area predictions. Moreover, the method can assist doctors with the large flow of patients by selecting cases with affected brain areas from normal ones. Thus, doctors first of all examine patients in need of emergency care.

It is also important to note that one of the main constraints of our objective is the small amount of data in

our dataset. It is with this we associate the unusual obtained values of the specificity metric, and with the class imbalance problem. In the future, we plan to explore other CNN architectures and increase the dataset.

Acknowledgments

The work was partly supported by RFBR grant No. 19-29-01175, and by the State Contract of the Sobolev Institute of Mathematics, Project No. FWNF-2022-0015.

References

[1] Tsao CW, Ada AW, Almarzooq ZI, et al. Heart disease and stroke statistics - 2022 update: A report from the American Heart Association. Circulation 2022; 145(8): e153-e639. DOI: 10.1161/CIR.0000000000001052.

[2] Shafaat O, Bernstock JD, Shafaat A, Yedavalli VS, Elsayed G, Gupta S, Sotoudeh E, Sair HI, Yousem DM, Sotoudehg H. Leveraging artificial intelligence in ischemic stroke imaging. J Neuroradiology 2022; 49(4): 343-351. DOI: 10.1016/j.neurad.2021.05.001.

[3] Fartakov EI, Tarkova AR, Kretov EI, Lomivorotov VV, Zykov IS. Possibilities and prospects of hypothermia in the treatment of ischemic stroke [In Russian]. Patologiya Krovoobrashcheniya i Kardiokhirurgiya 2019; 23(1): S18-S25. DOI: 10.21688/1681-3472-2019-1S-S18-S25.

[4] Bal S, Bhatia R, Menon BK, Shobha N, Puetz V, Dzialowski I, Modi J, Goyal M, Hill MD, Smith EE, Demchuk AM. Time dependence of reliability of noncontrast computed tomography in comparison to computed tomography angiography source image in acute ischemic stroke. Int J Stroke 2012; 10: 55-60. DOI: 10.1111/j.1747-4949.2012.00859.x.

[5] Sui B, Gao P. Imaging evaluation of acute ischemic stroke. J Int Med Res 2020; 48(1): 0300060518802530. DOI: 10.1177/0300060518802530.

[6] Simonsen CZ, Leslie-Mazwi TM, Thomalla G. Which imaging approach should be used for stroke of unknown time of onset? Stroke 2021; 52(1): 373-380. DOI: 10.1161/STROKEAHA.120.032020.

[7] Jadhav AP, Desai SM, Liebeskind DS, Wechsler LR. Neuroimaging of acute stroke Neurol Clin 2020; 38(1): 185-199. DOI: 10.1016/j.ncl.2019.09.004.

[8] Potter CA, Vagal AS, Goyal M, Nunez DB, Leslie-Mazwi TM, Lev MH. CT for treatment selection in acute ischemic stroke: A code stroke primer. Radiographics 2019; 39(6): 1717-1738. DOI: 10.1148/rg.2019190142.

[9] Kamalian S, Lev MH. Stroke imaging. Radiol Clin N Am 2019; 57(4): 717-732. DOI: 10.1016/j.rcl.2019.02.001.

[10] Goyal M, Ospel JM, Menon B, Almekhlafi M, Jayaraman M, Fiehler J, Psychogios M, Chapot R, van der Lugt A, Liu J, Yang P, Agid R, Hacke W, Walker M, Fischer U, Asdaghi N, McTaggart R, Srivastava P, Nogueira RG, Moret J, Saver JL, Hill MD, Dippel D, Fisher M. Challenging the ischemic core concept in acute ischemic stroke imaging. Stroke 2020; 51(10): 3147-3155. DOI: 10.1161/STROKEAHA.120.030620.

[11] Mangesius S, Janjic T, Steiger R, Haider L, Rehwald R, Knoflach M, Widmann G, Gizewski E, Grams A. Dual-energy computed tomography in acute ischemic stroke: State-of-the-art. Eur Radiol 2021; 31(6): 4138-4147. DOI: 10.1007/s00330-020-07543-9.

[12] Song K, Guan M, Li W, Jing Z, Xie X, Shi C, Liang J, Qiao H, Huang L. Acute ischemic stroke patients with

diffusion-weighted imaging - Alberta Stroke Program Early Computed Tomography Score < 5 can benefit from endovascular treatment: A single-center experience and literature review. Neuroradiology 2019; 61(4): 451-459. DOI: 10.1007/s00234-019-02177-1.

[13] Provost C, Soudant M, Legrand L, Hassen WB, Xie Y, Soize S, Bourcier R, Benzakoun J, Edjlali M, Boulouis G, Raoult H, Guillemin F, Naggara O, Bracard S, Oppenheim C. Magnetic resonance imaging or computed tomography before treatment in acute ischemic stroke. Stroke 2019; 50(3): 659-664. DOI: 10.1161/STROKEAHA.118.023882.

[14] Puig J, Shankar J, Liebeskind D, Terceno M, Nael K, Demchuk AM, Menon B, Dowlatshahi D, Leiva-Salinas C, Wintermark M, Thomalla G, Silva Y, Serena J, Pedraza S, Essig M. From "Time is brain" to "Imaging is brain": A paradigm shift in the management of acute ischemic stroke. J Neuroimaging 2020; 30(5): 562-571. DOI: 10.1111/jon. 12693.

[15] Pulli B, Heit JJ, Wintermark M. Computed tomography-based imaging algorithms for patient selection in acute ischemic stroke. Neuroimag Clin N Am 2021; 31(2): 235250. DOI: 10.1016/j.nic.2020.12.002.

[16] National Institute of Neurological Disorders and Stroke rt-PA Stroke Study Group. Tissue plasminogen activator for acute ischemic stroke. New Engl J Med 1995; 333(24): 1581-1587. DOI: 10.1056/NEJM199512143332401.

[17] Zhang X-H, Liang H-M. Systematic review with network meta-analysis: Diagnostic values of ultrasonography, computed tomography, and magnetic resonance imaging in patients with ischemic stroke. Medicine (Baltimore) 2019; 98(30): e16360. DOI: 10.1097/MD.0000000000016360.

[18] Brazzelli M, Sandercock PAG, Chappell FM, Celani MG, Righetti E, Arestis N, Wardla JM, Deeks JJ. Magnetic resonance imaging versus computed tomography for detection of acute vascular lesions in patients presenting with stroke symptoms. Cochrane DB Syst Rev 2009; 4: CD007424. DOI: 10.1002/14651858.CD007424.pub2.

[19] Prokhorikhin A, Baystrukov V, Boykov A, Malaev D, Tarkova A, Shayakhmetova S, Grishkov A, Kazancev A, Shigaev V, Kokh V, Avetisyan M, Umerenkov D, Kretov E. Neural network-based system of acute stroke non-contrast computed tomography diagnostics: A comparative study. Russ Electron J Radiol 2020; 10(3): 36-45. DOI: 10.21569/2222-7415-2020-10-3-36-45.

[20] Muhd Suberi AA, Zakaria WNW, Tomari R, Nazari A, Hj Mohd MN, Nik Fuad NF. Deep transfer learning application for automated ischemic classification in posterior fossa CT images. Int J Adv Comp Sci Appl 2019; 10: 459-465. DOI: 10.14569/ijacsa.2019.0100859.

[21] Nishio M, Koyasu S, Noguchi S, Kiguchi T, Nakatsu K, Akasaka T, Yamada H, Itoh K. Automatic detection of acute ischemic stroke using non-contrast computed tomography and two-stage deep learning model. Comput Meth Prog Bio 2020; 196: 105711. DOI: 10.1016/j.cmpb.2020.105711.

[22] Lisowska A, O'Neil A, Dilys V, Daykin M, Beveridge E, Muir K, McLaughlin S, Poole I. Context-aware convolutional neural networks for stroke sign detection in non-contrast CT scans. Comm Com Inf Sc 2017; 723: 494505. DOI: 10.1007/978-3-319-60964-5_43.

[23] Kuang H, Najm M, Chakraborty D, Maraj N, Sohn SI, Goyal M, Hill MD, Demchuk AM, Menon BK, Qiu W. Automated ASPECTS on noncontrast CT scans in patients with acute ischemic stroke using machine learning. Am J Neuroradiol 2019; 40(1): 33-38. DOI: 10.3174/ajnr.A5889.

[24] Reboujas Filho PP, Sarmento RM, Holanda GB, de Alencar Lima D. New approach to detect and classify

stroke in skull CT images via analysis of brain tissue densities. Comput Meth Prog Bio 2017; 148: 27-43. DOI: 10.1016/j.cmpb.2017.06.011.

[25] Peixoto SA, Rebouças Filho PP. Neurologist-level classification of stroke using a Structural Co-Occurrence Matrix based on the frequency domain. Comput Electr Eng 2018; 71: 398-407. DOI: 10.1016/j.compeleceng.2018.07.051.

[26] Kuang H, Qiu W, Najm M, Dowlatshahi D, Mikulik R, Poppe AY, Puig J, Castellanos M, Sohn SI, Ahn SH, Calleja A, Jin A, Asil T, Asdaghi N, Field TS, Coutts S, Hill MD, Demchuk AM, Goyal M, Menon BK, INTERRSeCT Collaborators. Validation of an automated ASPECTS method on non-contrast computed tomography scans of acute ischemic stroke patients. Int J Stroke 2020; 15(5): 528-534. DOI: 10.1177/1747493019895702.

[27] Sirsat MS, Fermé E, Câmara J. Machine learning for brain stroke: A review. J Stroke Cerebrovasc 2020; 29(10): 105162. DOI: 10.1016/j.jstrokecerebrovasdis.2020.105162.

[28] Inamdar MA, Raghavendra U, Gudigar A, Chakole Y, Hegde A, Menon GR, Barua P, Palmer EE, Cheong KH, Chan WY, Ciaccio EJ, Acharya UR. A review on computer aided diagnosis of acute brain stroke. Sensors 2021; 21(24): 8507. DOI: 10.3390/s21248507.

[29] Murray NM, Unberath M, Hager GD, Hui FK. Artificial intelligence to diagnose ischemic stroke and identify large vessel occlusions: A systematic review. J Neurointerv Surg 2020; 12(2): 156-164. DOI: 10.1136/neurintsurg-2019-015135.

[30] Isensee F, Kickingereder P, Wick W, Bendszus M, MaierHein KH. Brain tumor segmentation and radiomics survival prediction: Contribution to the BRATS 2017 challenge. In Book: Crimi A, Bakas S, Kuijf H, Menze B, Reyes M, eds. Brainlesion: glioma, multiple sclerosis, stroke and traumatic brain injuries. Cham: Springer International Publishing AG; 2018: 287-297. DOI: 10.1007/978-3-319-75238-9_25.

[31] Kamnitsas K, Chen L, Ledig C, Rueckert D, Glocker B. Multi-scale 3D convolutional neural networks for lesion segmentation in brain MRI. Proc MICCAI Brain Lesion Workshop 2015; 13: 46.

[32] Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In Book: Navab N, Hornegger J, Wells WM, Frangi AF, eds. Medical image computing and computer-assisted intervention - MICCAI 2015. Cham, Switzerland: Springer International Publishing; 2015: 234-241. DOI: 10.1007/978-3-319-24574-4_28.

[33] Isensee F, Jaeger PF, Kohl SA, Petersen J, Maier-Hein KH. nnUNet: A self-configuring method for deep learning-based biomedical image segmentation. Nat Methods 2021; 18(2): 203-211. DOI: 10.1038/s41592-020-01008-z.

[34] Woo I, Lee A, Jung SC, Lee H, Kim N, Cho SJ, Kim D, Lee J, Sunwoo L, Kang DW. Fully automatic segmentation of acute ischemic lesions on diffusion-weighted imaging using convolutional neural networks: Comparison with conventional algorithms. Korean J Radiol 2019; 20(8): 1275-1284. DOI: 10.3348/kjr.2018.0615.

[35] Clèrigues A, Valverde S, Bernal J, Freixenet J, Oliver A, Lladô X. Acute ischemic stroke lesion core segmentation in CT perfusion images using fully convolutional neural networks. Comput Biol Med 2019; 115: 103487. DOI: 10.1016/j.compbiomed.2019.103487.

[36] Çiçek Ö, Abdulkadir A, Lienkamp SS, Brox T, Ronneberger O. 3D U-Net: learning dense volumetric segmentation from sparse annotation. In Book: Ourselin S, Joskowicz L, Sabuncu MR, Unal G, Wells W, eds.

Medical image computing and computer-assisted intervention - MICCAI 2016. Pt II. Cham: Springer International Publishing AG; 2016: 424-432. DOI: 10.1007/978-3-319-46723-8_49.

[37] Milletari F, Navab N, Ahmadi SA. V-net: Fully convolutional neural networks for volumetric medical image segmentation. 2016 Fourth Int Conf 3D Vision 2016; 565-571 DOI: 10.1109/3DV.2016.79.

[38] Kamnitsas K, Ledig C, Newcombe VF, Simpson JP, Kane AD, Menon DK, Rueckert D, Glocker B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med Image Anal 2017; 36: 61-78. DOI: 10.1016/j.media.2016.10.004.

[39] Isensee F, Kickingereder P, Wick W, Bendszus M, MaierHein KH. No new-net. In Book: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, eds. Brainlesion: glioma, multiple sclerosis, stroke and traumatic brain injuries. Pt II. Cham: Springer Nature Switzerland AG; 2019: 234-244. DOI: 10.1007/978-3-030-11726-9_21.

[40] Tuladhar A, Schimert S, Rajashekar D, Kniep HC, Fiehler J, Forkert ND. Automatic segmentation of stroke lesions in non-contrast computed tomography datasets with convolutional neural networks. IEEE Access 2020; 8: 94871-94879. DOI: 10.1109/ACCESS.2020.2995632.

[41] Li Y., Fan Y. DeepSEED: 3D Squeeze-and-Excitation encoder-decoder convolutional neural networks for pulmonary nodule detection. Proc IEEE Int Symp Biomed Imaging 2020: 1866-1869. DOI: 10.1109/ISBI45749.2020.9098317.

[42] Abramova V, Clerigues A, Quiles A, Figueredo DG, Silva Y, Pedraza S, Oliver A, Lladö X. Hemorrhagic stroke lesion segmentation using a 3D U-Net with squeeze-and-excitation blocks. Comput Med Imag Grap 2021; 90: 101908. DOI: 10.1016/j.compmedimag.2021.101908.

[43] Roy AG, Navab N, Wachinger C. Concurrent spatial and channel 'squeeze & excitation' in fully convolutional networks. In Book: Frangi AF, Schnabel JA, Davatzikos C, Alberola-Löpez C, Fichtinger G, eds. Medical image computing and computer assisted intervention - MICCAI 2018. Pt I. Cham: Springer Nature Switzerland AG; 2018: 421-429. DOI: 10.1007/978-3-030-00928-1_48.

[44] Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, Roth RH, Xu D. UNETR: Transformers for 3d medical image segmentation. IEEE Wint Conf Appl 2022; 1748-1758. DOI: 10.1109/WACV51458.2022.00181.

[45] Fedorov A, Beichel R, Kalpathy-Cramer J, Finet J, Fillion-Robin J-C, Pujol S, Bauer C, Jennings D, Fennessy FM, Sonka M, Buatti J, Aylward SR, Miller JV, Pieper S, Kikinis R. 3D Slicer as an image computing platform for the quantitative imaging network. Magn Reson Imaging 2012; 30(9): 1323-1341. DOI: 10.1016/j.mri.2012.05.001.

[46] Lev MH, Gonzalez RG. CT angiography and CT perfusion imaging. In Book: Toga AW, Mazziotta JC, eds. Brain mapping: The methods. 2nd ed. Amsterdam: Academic Press; 2002: 427-484. DOI: 10.1016/B978-012693019-1/50019-8.

[47] Ioffe S, Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. Int Conf Mach Learn 2015; 37: 448-456.

[48] He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Proc CVPR IEEE 2016: 770-778. DOI: 10.1109/CVPR.2016.90.

[49] Sudre CH, Li W, Vercauteren T, Ourselin S, Jorge Cardoso M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Book: Cardoso MJ, Arbel T, Carneiro G, Syeda-Mahmood T, Tavares JMRS, Moradi M, Bradley A, Greenspan H, Papa

JP, Madabhushi A, Nascimento JC, Cardoso JS, Belagiannis V, Lu Z, eds. Deep learning in medical image analysis and multimodal learning for clinical decision support. Cham: Springer International Publishing AG; 2017: 240-248. DOI: 10.1007/978-3-319-67558-9_28.

[50] Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv Preprint. 2017. Source:

<https://arxiv.org/abs/1412.6980>. DOI:

10.48550/arXiv.1412.6980.

[51] Pérez-García F, Sparks R, Ourselin S. TorchIO: a Python library for efficient loading, preprocessing, augmentation and patch-based sampling of medical images in deep learning. Comput Meth Prog Bio 2021; 208: 106236. DOI: 10.1016/j.cmpb.2021.106236.

Authors' information

Anna Vadimovna Dobshik (b. 1999), bachelor of Applied Mathematics and Informatics, graduated from Novosibirsk State University, Mechanics and Mathematics department in 2021. Works as a machine learning engineer for DeepSound AI. Research interests are computer vision, medical image segmentation. E-mail: a.dobshik@alumni.nsu.ru .

Sergey Konstantinovich Verbitskiy (b. 1999), bachelor of Applied Mathematics and Informatics, graduated from Novosibirsk State University in 2021. Works as a machine learning engineer at StyleDNA. Research interests are image segmentation, signal processing, speech recognition, object detection, and video action recognition. E-mail: s.verbitskii@alumni.nsu.ru .

Igor Alekseevich Pestunov (b. 1955), graduated from Novosibirsk State University in 1977. PhD in Physics and Mathematics (1998). He is a key researcher at the Federal Research Center for Information and Computational Technologies. His current research interests include clustering, pattern recognition, and image analysis. E-mail: pestunov@ict.nsc.ru .

Kseniya Mihailovna Sherman (b.1989), graduated from Novosibirsk State University in 2012. Residency in Traumatology and Orthopedics; specialization in Radiology until 2017. Currently works as a radiologist at the International Tomography Center. Scientific interests: introduction of new technologies into medical practice. E-mail: Ksh1420@yandex.ru .

Yuriy Nikolaevich Sinyavskiy (b. 1983), graduated from Novosibirsk State University in 2005. PhD in Technical Sciences (2021). Currently he works as the researcher at the Federal Research Center for Information and Computational Technologies. Research interests are image segmentation, geoinformatics, and software development. E-mail: yorikmail@gmail.com .

Andrey Alexandrovich Tulupov (b. 1981), graduated from Moscow State University in 2005. PhD in Medical Sciences (2006). Currently he works as the chief researcher at the International Tomography Center of the Russian Academy of Sciences. Research interests are magnetic resonance imaging and computed tomography. E-mail: taa@tomo.nsc.ru .

Vladimir Borisovich Berikov (b. 1964), graduated from Novosibirsk State University in 1986. Received candidate's degree in 1996 and doctoral degree in 2007. Works at the Sobolev Institute of Mathematics SB RAS, acts as a head of laboratory and chief researcher. Research interests: mathematical methods of data analysis and their application in image processing and medical data. E-mail: berikov@math.nsc.ru .

Received October 01, 2022. The final version - April 04, 2023.

i Надоели баннеры? Вы всегда можете отключить рекламу.