- Research
- Open access
- Published:
Artificial intelligence’s impact on drug delivery in healthcare supply chain management: data, techniques, analysis, and managerial implications
Journal of Big Data volume 11, Article number: 177 (2024)
Abstract
Healthcare supply chain management’s (HSCM) significance to economic and societal development is huge. In today’s very competitive market, supply chains have seen significant changes in the last several years. There is a need for technology that can handle the increasing complexity of today’s dynamic supply chain activities. Both machine learning (ML) and the quick dissemination of information have the potential to revolutionize the supply chain. ML has spawned a slew of useful supply chain applications in recent years, HSCM has received comparatively less attention. In this study, we applied three ML algorithms such as gradient boosting (GB), histogram gradient boosting (HGB), and cat boosting (CB) with data preprocessing tools to predict whether the medicines are delivered on time or not in the HSCM. The data preprocessing tools are used to manage datasets and increase the performance of ML algorithms. There are three methods of feature selection that are applied in this study such as Pearson correlation, chi-square test, and principal component analysis to select the best features to push in the ML algorithms. The main results show the CB is the best algorithm with the highest accuracy, precision, recall, and f1 score with values respectively. The three ML algorithms are compared with other ML algorithms to show the robustness of the applied ML algorithms. We made a sensitivity analysis to show the chaining in learning rate (LR) and compute the accuracy of the ML algorithms. We show the CB is not sensitive to values between 0.1 and 1.
Introduction
Increased knowledge and digital technology have a profound impact on healthcare, one of the most essential service industries. Overcrowding, lack of availability, and inaccessibility are all problems that have disrupted patient treatment in healthcare systems. Unpredictable calamities, such as earthquakes, floods, fires, and pandemic breakouts, amplify these interruptions and cause permanent harm to the healthcare system and many other essential sectors [1, 2]. These unforeseen occurrences throw healthcare systems off-kilter and cause bottlenecks. Different sections of healthcare supply chains (HSCs) should perform their crucial responsibilities successfully to address these difficulties, especially during emergencies [3, 4].
HSCs are made up of a network of suppliers, manufacturers, hospitals, blood banks, and pharmacies all working together to meet patient needs and provide optimal care. HSCs often have vendors of medical equipment and gadgets at the top of the food chain, and patients at the bottom [5,6,7]. Any error or disturbance at any level of HSCs may endanger lives and have far-reaching consequences. Complexity, diversity, service types, unpredictability, and objectives are what set HSCs apart from other supply chains (SCs). Big data analytics, AI, Blockchain, and cloud computing are just a few of the digital technologies that have greatly benefited HSCs during the last decade. Many new business possibilities arise as a result of the use and development of these innovations in healthcare systems, including the development of new business models for enhancing performance and generating anticipated values [8, 9]. Several studies applied blockchain and big data for drug SCM such as hospital waste management systems [10], pharmaceutical cold chains [11], evaluating service supply chain performance [12], and reduction maps for COVID-19 vaccine supply chain [13].
Machine learning (ML) algorithms can be applied in HSCM in various case studies. In this study, we apply the ML algorithms to predict the delivered medicines on time or not. The ML is a subset of the AI [14, 15]. To enable a system to automatically learn and develop from experience without having to be explicitly coded, ML is a vital component of artificial intelligence (AI) [16, 17]. ML methods are data-driven, meaning they actively look for patterns and adapt as they learn more. Opportunities for evaluating, categorizing, and forecasting on-time or delayed medication delivery are greatly enhanced by ML [18, 19].
ML uses the historical data in HSCM for patients in organizations to analyze some features such as ecological factors, and disease states to forecast the best delivery method for organizations and ensure the patient receives the drug as fast as possible. ML can be connected with sensors in the smart system to monitor the real-time delivery of products and can change the delivery in real-time to reduce costs and time delivery. In the traffic ways, ML can change the ways of delivering drugs to reach patients quickly. In the long ways, ML can process large historical data to predict the drug delivery is on time or not, so it can change the delivery system to reduce cost and faster reach to patients and organizations.
To predict the long-term viability of HSCs, Azadi et al. [20] developed a network data envelopment analysis (NDEA) framework and a deep learning technique. To foretell HSCs’ long-term viability, they used an NDEA model and a deep learning strategy. Bounded connection values may be optimized with their help. The DEA abilities are used to determine the threshold for every of these constrained connections to improve the performance of decision making units (DMUs). Each DMU’s dual-role connections function is also specified. Their primary findings indicate that the highest-scoring HSCs are those that consume the fewest resources while producing the highest quality goods and the fewest problems.
An ML-based methodology for choosing vendors’ incoterms (contracts) for direct drop-shipping in a worldwide omnichannel pharmaceutical supply chain was suggested by Detwal et al. [21]. They brought to light the most important considerations when deciding on an incoterm to use with a pharmaceutical vendor. Their results demonstrate that, for a wide range of input characteristics, the suggested model can reliably forecast a vendor incoterm (contract).
Critical success factors (CSFs) for artificial intelligence adoption in healthcare service delivery were the subject of Kumar et al.‘s research [22]. They utilize a rough version of Step-wise Assessment Ratio Analysis (SWARA) to rank HSC CSFs for the application of AI. According to their findings, technical, institutional/environmental, human, and organizational elements are the most significant in determining whether or not AI is used in HSC in developing nations.
Anil Kumar et al. [23]investigated what makes HSCs resilient and how to build resilient omnichannel HSCs (OHSCRs) of the future. A thorough literature analysis and in-depth interviews with subject-matter experts yielded the components that contribute to HSCs’ resilience. In the second stage, the cutting-edge building blocks of OHSCR were created using a machine-learning technique called K-means clustering. In the last stage, we spoke about what this all means and what directions we think future research should go in. The researchers concluded that the healthcare industry should evaluate OHSCR with a focus on six key building blocks. Table 1 shows the related works.
The main contributions of this study are in summary:
-
I.
We gathered the data in HSCM to predict whether the medicine was delivered on time or not under different ML models.
-
II.
We apply feature selection methods such as Pearson correlation, chi-square test, and principal component test and show the ML model’s performance under different feature selection methods.
-
III.
The data preprocessing methods are applied in the dataset such as dealing with missing values and data normalization. Normalization is used to manage the outliers in the data and put all data in a specific range. The encoding data is used to encode the text categorical data. We compared the three models before and after applying the normalization method, we show higher accuracy after applying the normalization method.
-
IV.
We applied the three ML algorithms such as gradient boosting, histogram gradient boosting, and cat boosting in HSCM.
-
V.
We compare the applied ML with other ML algorithms. Our models show higher accuracy and performance compared with other models. The cat boosting obtained a higher accuracy equal to 98.8%.
-
VI.
We conducted a sensitivity analysis to show the changing of parameters and then computed the performance measurement. We showed the cat-boosting algorithm is not sensitive under different learning rates.
The rest of this paper is organized as: in Sect. 2 we introduce the experimental dataset. In Sect. 3, we introduce the three ML algorithms. In Sect. 4, we introduce the experimental setup. The results of the applied three ML algorithms are presented in Sect. 5 with the comparative and sensitivity analysis. In Sect. 6 we introduce the managerial implication from this study. In Sect. 7, we introduce the conclusions of this study.
Experimental dataset
This work uses the dataset to predict whether the medicine is delivered on time or not in HSCM. This study uses this dataset from the GitHub website and the details of this dataset are shown in Table 2. The dataset can be downloaded from [36]. This dataset includes 10,324 samples and 33 columns of information about drugs from pharmaceutical companies. The samples were collected from 2006 to 2016 delivering medicines to different countries. This study tends to predict whether the medicine is delivered on time or not. So we created a column named target by subtracting the delivered to client date and scheduled delivery date. If the delivered data is before the scheduled date, we put 1 in the target column. If the delivery date is after the scheduled date, we put 0 in the target column. So, we have a target column consisting of 0 (not delivered on time) and 1 (delivered on time). The column data includes day, month, and year. Figure 1 shows the number of orders by the target column, where x refers to the target column (0 not delivered on time) and (1 delivered on time), and y refers to the number of orders. The number of orders delivered on time is 4000 orders and not delivered on time is 6324. We analyze the dataset by computing the number of orders delivered on time and not in each country as shown in Fig. 2. We show that South Africa has the largest number of orders with 1013 orders not delivered on time followed by Vietnam with 666 orders and Cot d’Ivoire, and Afghanistan has the least orders not delivered on time followed by Liberia. Nigeria and Cot d’Ivoire are the largest countries that have orders delivered on time with 699 and 483 orders, and Libya and Benin have the least orders that are delivered on time. Figure 3 shows the number of dosage forms in the dataset. We show that Tablet has 34.2% of the dataset (3532 orders), followed by Tablet FDC 26.6% (2749 orders) and test kit 15.3% (1572 orders). So the Tablet is important to be delivered on time in the HSCM due to has largest number of orders in the dataset.
Figure 4 shows the relation between the date delivered to the client and the date scheduled. We observe the difference between the two dates. So, we proposed this study to predict the medicines are delivered on time or not to aid organizations in healthcare. Figure 5 shows the number of orders in each shipment mode. This study has four types of shipment modes such as air, truck, air charter, and ocean. The air shipment mode has the largest number of orders (6113 orders), followed by truck (2830 orders), air charter (650 orders), and ocean (371). So, the transportation cost is increased in the air shipment mode due to having the largest number of orders. Figure 6 shows the top 20 order names in our dataset. We observed that Efavirenz has the largest number of orders (755 orders). Table 3 shows the ten transactions of the dataset with 33 columns. Table 4 shows the summary statistics of the dataset. From Table 4, there are inconsistent values in line item value, pick price, weight, fright cost, and line item insurance. So, we drop all the missing values for our dataset.
Figure 7 shows the correlation map between features of our dataset. There are three values in correlation such as positive relationships (strong, medium, weak), no relationships, and negative relationships. There is a correlation between nine features and the target class. The line measurement, pack price, and unit price have a negative relationship with the target class. ID feature has a medium relationship with the target class. The line item quality, line item value, weight, freight cost, and line item insurance have positive relationships with the target class but weak relationships.
Figure 8 shows the outlier in the country, freight cost, line item insurance, and line item value. To mitigate the impact of the outliers, normalizing the data is recommended. Since the dataset is small, it’s preferable not to eliminate any rows unless necessary.
The dataset has many missing values, so first, we drop all missing values. Then, we perform data preprocessing. Then, we perform the feature selection to select the best feature. Then, we applied the ML algorithms.
Methods
This section offers ML models to predict the medicines delivered on time or not in HSCM. This study uses boosting ML models. Boosting is a subset of Ensemble Learning that combines many low-performing predictors to produce a single high-performing model. This is effective because successive models learn from the errors of their predecessors. We use the three types of GB, CB, and HGB. The following detailed of using models:
Gradient boosting classifier (GB)
Decision trees (categorization trees for problematizing patterns and regression trees for approximating functions) are the backbone of GB, an ensemble of machine-learning algorithms. Each succeeding tree in an ensemble is reliant on the trees that came before it. That’s because the second tree works to minimize the former’s mistakes as shown in Fig. 9. During model training as shown in Fig. 10, this technique reduces the loss function that represents the ensemble’s misclassification rate. Importantly, by going in the opposite direction as the negative gradient, GB reaches global convergence. A strong committee may be established after training, even if the basic classifier was rather poor. The last step is to compile the predictions made from all of the decision trees to acquire the outcomes of new data samples [37, 38].
Cat boosting classifier (CB)
The Train Using Auto ML program implements CB, a supervised machine-learning technique that uses decision trees for regression and categorization. CB’s two primary characteristics are its ability to handle categorical data and its reliance on CB. Iteratively building multiple decision trees is the core of the GB procedure. As more trees are added, the quality of the final product increases. CB is a speedier alternative to the traditional gradient boost approach.
With CB, you may skip the tedious pre-processing step of converting category text variables to numbers, one-hot-encodings, and so on, a common requirement of other decision tree-based approaches. Without the need for any preprocessing, this approach may be fed a mixture of category and non-categorical explanatory factors. As part of the method, it does preprocessing. CB employs an approach to encoding categorical characteristics known as ordered encoding. To find a number to stand in for the categorical characteristic, ordered encoding takes into account the desired statistics from all rows before the data point in question [39, 40].
CB is distinct in its usage of symmetric trees, for example. This implies that the same split condition is used by all decision nodes across all depths.
Histogram Gradient boosting classifier (HGB)
Furthermore, the histogram-based technique is a powerful tool for training models using GB. This approach uses tiny bins to categorize the spectrum of continuous information utilized by decision trees. Hossain and Deb proposed a value of 255 for the total amount of bins. Histograms showing the distribution of characteristic values are constructed using these bins. The amount of data examples and the total gradients for every bin may be calculated, among other elementary statistics. Using these measures, we may zero in on the sweet spots for splitting apart the base-level learners throughout their instruction. As the histogram-based technique does not need to scan all ranges of characteristics for split point evaluation during the training stage of a decision tree, it may greatly minimize the computing cost. Although the learning phase is more resistant to noise, the histogram-based technique may also aid in greater generalization. This research employs a histogram-based technique to build the categorization tree structure because of its benefits in computing efficiency and learning effectiveness [41, 42].
Experimental setting
In this section, we offer the feature selection methods to select the best features in our dataset to obtain the highest accuracy, the performance measurement, and the implementation settings of our experiments.
Feature selection
We applied three feature selection methods in this study and obtained the ML model accuracy under three feature selection methods.
Pearson Correlation (PC)
Correlation-based Feature Selection (PC) calculates inter-correlation values to compute the correlation between features. This method can obtain links and relationships between the features and target features [43, 44]. A linear relationship between an input X and an output Y may be measured using a statistic called PC. It may be positive or negative depending on the strength of the relationship being measured. Therefore, a value of 0 indicates the absence of a linear relationship between features and target features. The PC is computed by dividing the covariance of input X and output Y at the input by the standard deviation of input X times and Y at the output.
Cov refers to the covariance and std refers to the standard deviation.
Chi-Square Test
The chi-square test is useful for resolving the feature-selection issue since it examines the correlation between features [45, 46]. To determine whether or not two occurrence features are independent, researchers applied the chi-square test. Chi-Square calculates the deviation between the predicted output (E) and the actual output (O).
The number of possible, arbitrary values, or “degrees of freedom,” is defined as the largest possible set of independent variables. It may be written as the number of observations minus the number of constraints that are not based on the observations themselves.
Principal component analysis (PCA)
Often used to decrease the dimensionality of huge datasets, PCA works by reducing a large number of parameters and dimensions by preserving the amount of information in the dataset. Accuracy suffers as a data set’s amount of variables is reduced, but the idea is decreasing dimensionality without loss in information. Because ML techniques can more quickly and easily analyze data points when working with smaller data sets there are fewer factors to consider when exploring and visualizing such sets [47, 48].
Implantation settings
We implemented this study by using Python 3.10 with the library Sklearn. We trained the three algorithms by various values in hyperparameters. Table 5 shows the hyperparameter values for three algorithms. We divide the dataset into training and testing. The training value is 85% of the dataset and the test value is 15% of the dataset.
Performance measurement
In this study, we suggest four performance measurements such as accuracy, precision, recall, and f1 score to evaluate three applied ML algorithms to select the best one. The equations of four performance measurements are detailed:
Results and analysis
This section offers the results of the applied three ML algorithms in the HSCM dataset to predict whether the medicines are delivered on time or not.
Data preprocessing results (missing values and feature selection)
In this part, we introduce the data preprocessing tools to obtain good data to push it into the ML models. The data preprocessing contains results of missing values, and results, of feature selection.
We split the missing values into serval processes. First, we dropped all missing values of the dataset, we obtained 1526 records from the 10,324 records. Then we applied the feature selection tools. Table 6 shows the results of ML algorithms after dropping all missing values and applying feature selection tools.
In PC, there are three numbers of features selected including 3,6,9. We applied the ML algorithms in the PC with three features were selected. Of the 3 features, the GB has the highest accuracy, precision, recall, and f1 score. In 6 features, the CB has the highest accuracy, precision, and f1 score, but the HGB has the highest recall score. Of the 9 features, the HGB and CB have the highest accuracy, the CB has the highest precision, and the GB has the highest recall and f1 score.
In the chi-square test, in 3 features, the CB has the highest accuracy, and precision and the GB has the highest recall score and f1 score. In 6 features, CB has the highest accuracy, precision, recall, and f1 score. Of the 9 features, the GB has the highest accuracy and recall, the HGB has the highest precision and f1 score, and CB has the highest accuracy with the GB.
In the PCA, in the 3 features, the CB has the highest accuracy, recall, and f1 score, and the HGB has the highest precision and f1 score. Of the 6 features, the HGB has the highest accuracy, precision, recall, and f1 score. In the 9 features, the CB has the highest accuracy, precision, recall, and f1 score.
Overall, the CB has the highest accuracy, precision, and score in 9 features with the PCA feature selection. We conclude the CB is the best algorithm in the applied HSCM dataset with the PCA feature selection.
Enhancement model accuracy
The goal of this part enhance the accuracy of the three ML algorithms. We concluded when dropping the missing values, we obtained a few records of the dataset (14.7% of the dataset). So, there is a large amount of information lost. So, we dropped the irrelevant columns and then computed the correlation between the dataset as shown in Fig. 11. We dropped the missing values in the dataset after dropping all irrelevant columns. We obtained 2902 records (28% of the dataset). After this, we doubled the size of the selected dataset. After this, we selected the nine columns (Country, Fulfill Via, Vendor INCO Term, Shipment Mode, Scheduled Delivery Date, Delivered to Client Date, Line-Item Value, Freight Cost (USD), Line Item Insurance (USD)). The nine selected features have various factors such as the type of shipment to deliver the medicines to clients, data of scheduled and delivered date, cost of transportation, and country of delivered medicines.
These selected features have the text categorical data such as country. So, we encode these data via the Label Encoder method. Then there are outliers in the dataset. So, we normalize these data to ensure that all are in the specific range and prevent the outlier. We used the Standard Scaler method to normalize the dataset. Then we split the dataset into a training (2466 records) and a testing set (436 records). We applied the three ML algorithms to the selected feature. Figure 12 shows the performance measurement of the three ML algorithms.
We observed the CB has the highest accuracy (98.8%), recall (87.5%), and f1 score (93.3%), but all algorithms are equal in precision (100%). So, the CB is the best ML algorithm.
Table 7 shows the confusion matrix of the three ML algorithms. We found the CB has 396 records of true positive and five records are predicted false.
ROC-AUC curve
The area under the receiver operating characteristics (ROC) curve, which represents the degree of class separation, is a crucial performance metric for categorization algorithms. The area under the curve (AUC) measures how likely it is that a random positive example would be ranked higher than a random negative instance by the model being depicted. The area under the ROC curve from 0 to 1 is represented by the AUC, which is a metric with a maximum value of 1. Figure 13 shows the ROC curve for the three ML algorithms. We found that CB is the best algorithm with an accuracy of 99.48% followed by GB with an accuracy (of 98.88), and the lowest accuracy is HGB with 98.79%.
Analysis with other ML algorithms
We compare the three ML models (GB, HGB, and CB) with other ML models such as logistic regression, support vector machine, random forest, ada boosting, and k-nearest neighbors [49,50,51,52,53]. We used the ML in the previous studies and applied it to our dataset. We found the three applied ML models are the best algorithms. We found that ada boosting has the least accuracy, precision, recall, and f1 score followed by logistic regression, and decision tree. Table 8 shows the comparative analysis results.
Statistical test
We used the statistical test method to test the applied three models with the comparative models as shown in Table 9. We used McNemar’s test to test the significant statistical or not [54]. McNemar’s test determines whether two sets of conflicts are consistent with one another. In statistical parlance, this is referred to as marginal homogeneity of the contingency table. For this reason, we may classify McNemar’s test as a homogeneity test for contingency tables. The test provides feedback on whether or not two models disagree in the same manner (or not) when comparing binary classification algorithms. It does not evaluate the relative quantity of different models or the likelihood of their making errors. We compute the p-value to compare with the threshold value (0.05). If the p-value is less than the threshold value, there is a significant statistical. If not, we conclude there is no significant statistical. In this study, we conclude the applied three models have significant statistics with the comparative model.
Ablation study
In this part, we discuss the impact of the normalization method and LR when applied in the ML algorithms.
Impact of normalization
In this part, we measure the impact of the normalization method on the ML algorithms. In this study, we applied the standard scaler method to manage outliers and normalize all data by putting it in a specific range. Table 10 shows the accuracy precision, recall, and f1 score after and before applying three ML models. Before applying the normalization method, we obtained the CB has the highest accuracy (97.2%), followed by HGB with accuracy (96.5%), and GB with accuracy (95.6%). After applying the normalization method, we found the accuracy is increased, the CB has the highest accuracy with (98.85%) followed by the HGB with accuracy (98.39), and GB with accuracy (98.17%). We conclude the normalization method has a big impact on the applied three ML algorithms.
Sensitivity analysis
In this part, we applied the sensitivity analysis in this study to show the changes in the performance measures of the applied three ML algorithms. We change the values of the LR and then compute the performance measurement. We put three values of LR such as 1.0, 0.1, and 0.01, and then computed the performance measurements as shown in Table 11. We found the performance decreased from the original value of the LR. In GB, the highest accuracy in the LR is 0.1. The highest accuracy in the LR is 0.1. The highest accuracy in the CB is LR 0.1, 1. So, the CB is not sensitive to the LR between 0.1 and 1, and the other algorithms are sensitive to changing LRs.
We applied the k-fold cross-validation method for five folds for all models. We show a range of accuracy between 90 and 98.8% which means to accuracy is high and stable under different folds, the range of accuracy is small that show the consistent performance of the model and model has not the overfitting problem.
Comparative analysis with previous studies
We compare the CB (which has the highest accuracy) with the previous studies as shown in Table 12. All these algorithms are applied in our dataset. From Table 12, we show that LSTM has 63.9% accuracy and CNN has 63.9% accuracy. The decision tree has two accuracies 88.98% and 80%. The KNN has a 98.09%. The SVM has an 83.3%. We show our applied algorithm is the best with 98.85% accuracy.
Managerial implications
There are various managerial implications in the prediction of the delivery of medicines on time or not in HSCM.
The results of this study can aid managers and organizations in predicting the delivered medicines in the HSCM in inventory management to make considerations about the delay in the delivered medicine or not, so the inventory has an important role in storing more medicines if the delivered medicines are delayed.
The results of this study can aid organizations in demand for drugs. This can help the manager to order the most accurate number of medicines in the future based on the scheduled date of delivery.
Patient care and customer satisfaction can be increased by using this study model to predict whether medicines are delivered on time or not. This allows managers to enhance customer service by communicating with the service provider in the predicted delay.
This study can help managers in cost management in the HSCM by knowing the best shipment mode, inventory cost, and some shortcuts.
This model can aid in risk management by reducing the risk when predicting whether the medicines are delivered on time or not. So, the manager can develop a plan to manage the orders of medicines in the HSCM.
Conclusions
This study applied the data preprocessing methods and ML algorithms in the HSCM dataset to predict the delivered medicines on time or not. We applied the data preprocessing method to the dataset. We handle the missing values by dropping all of them. Then we use the normalization method to manage the outliers and put all data in a specific range. Then we encode the text categorical data. We applied the three ML algorithms in the HSCM dataset to show whether the medicines were delivered on time or not. We obtained the CB is the largest accuracy, followed by the HGB and GB. We used the ROC-AUC curve to show the accuracy of the three applied ML algorithms, we show the CB is the best algorithm. We made a comparative analysis to show the robustness of the applied three ML algorithms. We compare the applied ML with five ML algorithms. We show that the three applied ML algorithms have the highest accuracy, precision, recall, and f1 score. We made a sensitivity analysis to show the changes in the LR values and then computed the performance measurement. We put the leering rate with three values such as 0.1, 0.01, and 1. We conclude the CB is not sensitive between 0.1 and 1 LR and other ML algorithms are sensitive to changes in LR. There is a limitation in this study, a few number of samples in the dataset, in future studies this problem can be solved by collecting datasets from different sources to increase the volume of the dataset. Also, the deep learning models can be used in the future direction to show the accuracy, precision, recall, and f1 score to show the difference between deep learning and machine learning. Transfer learning can be used to train a large number of datasets to obtain higher accuracy and performance.
Data availability
No datasets were generated or analysed during the current study.
References
Kim S-H, Kwon I-WG. The study of healthcare supply chain management in United States: literature review. Manag Rev Int J. 2015;10(2):34.
Polater A, Demirdogen O. An investigation of healthcare supply chain management and patient responsiveness: an application on public hospitals. Int J Pharm Healthc Mark. 2018;12(3):325–47.
Mathur B, Gupta S, Meena ML, Dangayach GS. Healthcare supply chain management: literature review and some issues. J Adv Manag Res. 2018;15(3):265–87.
Kwon I-WG, Kim S-H, Martin DG. Healthcare supply chain management; strategic areas for quality and financial improvement. Technol Forecast Soc Change. 2016;113:422–8.
McKone-Sweet KE, Hamilton P, Willis SB. The ailing healthcare supply chain: a prescription for change. J Supply Chain Manag. 2005;41(1):4–17.
Clauson KA, Breeden EA, Davidson C, Mackey TK. Leveraging Blockchain Technology to Enhance Supply Chain Management in Healthcare:: an exploration of challenges and opportunities in the health supply chain. Blockchain Healthc Today, 2018.
Rakovska MA, Stratieva SV. A taxonomy of healthcare supply chain management practices. in Supply Chain Forum: An International Journal. Taylor & Francis; 2018. pp. 4–24.
AbuKhousa E, Al-Jaroodi J, Lazarova-Molnar S, Mohamed N. Simulation and modeling efforts to support decision making in healthcare supply chain management, Sci. World J., vol. 2014, 2014.
Haszlinna Mustaffa N, Potter A. Healthcare supply chain management in Malaysia: a case study. Supply Chain Manag Int J. 2009;14(3):234–43.
Bamakan SMH, Malekinejad P, Ziaeian M. Towards blockchain-based hospital waste management systems; applications and future trends. J Clean Prod. 2022;349:131440.
Bamakan SMH, Moghaddam SG, Manshadi SD. Blockchain-enabled pharmaceutical cold chain: applications, key challenges, and future trends. J Clean Prod. 2021;302:127021.
Bamakan SMH, Faregh N, ZareRavasan A. Di-ANFIS: an integrated blockchain–IoT–big data-enabled framework for evaluating service supply chain performance. J Comput Des Eng. 2021;8(2):676–90.
Bamakan SMH, Malekinejad P, Ziaeian M, Motavali A. Bullwhip effect reduction map for COVID-19 vaccine supply chain. Sustain Oper Comput. 2021;2:139–48.
Rana A, Dumka A, Singh R, Panda MK, Priyadarshi N, Twala B. Imperative role of machine learning algorithm for detection of Parkinson’s disease: review, challenges and recommendations, Diagnostics, vol. 12, no. 8, p. 2003, 2022.
Fu Y, Downey ARJ, Yuan L, Zhang T, Pratt A, Balogun Y. Machine learning algorithms for defect detection in metal laser-based additive manufacturing: a review. J Manuf Process. 2022;75:693–710.
Herm L-V, Heinrich K, Wanner J, Janiesch C. Stop ordering machine learning algorithms by their explainability! A user-centered investigation of performance and explainability. Int J Inf Manage. 2023;69:102538.
Allugunti VR. Breast cancer detection based on thermographic images using machine learning and deep learning algorithms. Int J Eng Comput Sci. 2022;4(1):49–56.
Taha AMH, Ariffin D, Abu-Naser SS. A systematic literature review of Deep and Machine Learning algorithms in Brain Tumor and Meta-Analysis. J Theor Appl Inf Technol. 2023;101(1):21–36.
Nasir N, et al. Water quality classification using machine learning algorithms. J Water Process Eng. 2022;48:102920.
Azadi M, Yousefi S, Saen RF, Shabanpour H, Jabeen F. Forecasting sustainability of healthcare supply chains using deep learning and network data envelopment analysis. J Bus Res. 2023;154:113357.
Detwal PK, Soni G, Jakhar SK, Shrivastava DK, Madaan J, Kayikci Y. Machine learning-based technique for predicting vendor incoterm (contract) in global omnichannel pharmaceutical supply chain. J Bus Res. 2023;158:113688.
Kumar A, Mani V, Jain V, Gupta H, Venkatesh VG. Managing healthcare supply chain through artificial intelligence (AI): a study of critical success factors. Comput Ind Eng. 2023;175:108815.
Kumar A, et al. Digging DEEP: futuristic building blocks of omni-channel healthcare supply chains resiliency using machine learning approach. J Bus Res. 2023;162:113903.
Tirkolaee EB, Sadeghi S, Mooseloo FM, Vandchali HR, Aeini S. Application of machine learning in supply chain management: a comprehensive overview of the main areas, Math. Probl. Eng., vol. 2021, no. 1, p. 1476043, 2021.
Lin H, Lin J, Wang F. An innovative machine learning model for supply chain management. J Innov Knowl. 2022;7(4):100276.
Carbonneau R, Laframboise K, Vahidov R. Application of machine learning techniques for supply chain demand forecasting. Eur J Oper Res. 2008;184(3):1140–54.
Wisetsri W, Donthu S, Mehbodniya A, Vyas S, Quiñonez-Choquecota J, Neware R. An investigation on the impact of digital revolution and machine learning in supply chain management. Mater Today Proc. 2022;56:3207–10.
Makkar S, Devi GNR, Solanki VK. Applications of machine learning techniques in supply chain optimization. in ICICCT 2019–System reliability, Quality Control, Safety, maintenance and management: applications to Electrical, Electronics and Computer Science and Engineering. Springer; 2020. pp. 861–9.
Kumar V, Pallathadka H, Sharma SK, Thakar CM, Singh M, Pallathadka LK. Role of machine learning in green supply chain management and operations management. Mater Today Proc. 2022;51:2485–9.
Pontrandolfo P, Gosavi A, Okogbaa OG, Das TK. Global supply chain management: a reinforcement learning approach. Int J Prod Res. 2002;40(6):1299–317.
Han C, Zhang Q. Optimization of supply chain efficiency management based on machine learning and neural network. Neural Comput Appl. 2021;33(5):1419–33.
Hu H, Xu J, Liu M, Lim MK. Vaccine supply chain management: an intelligent system utilizing blockchain, IoT and machine learning. J Bus Res. 2023;156:113480.
Pasupuleti V, Thuraka B, Kodete CS, Malisetty S. Enhancing supply chain agility and sustainability through machine learning: optimization techniques for logistics and inventory management. Logistics. 2024;8(3):73.
Jahin MA, Shovon MSH, Shin J, Ridoy IA, Mridha MF. Big Data—Supply Chain Management Framework for forecasting: Data Preprocessing and Machine Learning techniques. Arch Comput Methods Eng, pp. 1–27, 2024.
Camur MC, Ravi SK, Saleh S. Enhancing supply chain resilience: a machine learning approach for predicting product availability dates under disruption. Expert Syst Appl. 2024;247:123226.
Tamym L, Moh ANS, Benyoucef L, El MD, Ouadghiri. Goods and activities tracking through supply chain network using machine learning models, in IFIP International Conference on Advances in Production Management Systems, Springer, 2021, pp. 3–12.
Bentéjac C, Csörgő A, Martínez-Muñoz G. A comparative analysis of gradient boosting algorithms. Artif Intell Rev. 2021;54:1937–67.
Natekin A, Knoll A. Gradient boosting machines, a tutorial. Front Neurorobot. 2013;7:21.
Zhou F, et al. Fire prediction based on catboost algorithm. Math Probl Eng. 2021;2021:1–9.
Dorogush AV, Ershov V, Gulin A. CatBoost: gradient boosting with categorical features support, arXiv Prepr. arXiv1810.11363, 2018.
Guryanov A. Histogram-based algorithm for building gradient boosting ensembles of piecewise linear decision trees, in Analysis of Images, Social Networks and Texts: 8th International Conference, AIST 2019, Kazan, Russia, July 17–19, 2019, Revised Selected Papers 8, Springer, 2019, pp. 39–50.
Ong YJ, Zhou Y, Baracaldo N, Ludwig H. Adaptive histogram-based gradient boosted trees for federated learning, arXiv Prepr. arXiv2012.06670, 2020.
Sugianela Y, Ahmad T. Pearson correlation attribute evaluation-based feature selection for intrusion detection system, in 2020 International Conference on Smart Technology and Applications (ICoSTA), IEEE, 2020, pp. 1–5.
Ni L, Fang F, Wan F. Adjusted Pearson Chi-Square feature screening for multi-classification with ultrahigh dimensional data. Metrika. 2017;80:805–28.
Rachburee N, Punlumjeak W. A comparison of feature selection approach between greedy, IG-ratio, Chi-square, and mRMR in educational mining, in 2015 7th international conference on information technology and electrical engineering (ICITEE), IEEE, 2015, pp. 420–424.
Chitsaz E, Taheri M, Katebi SD, Jahromi MZ. An improved fuzzy feature clustering and selection based on chi-squared-test, in Proceedings of the international multiconference of engineers and computer scientists, 2009, pp. 18–20.
Rapeepongpan J, Padungweang P, Lavangnananda K. Logistic Principle Component Analysis (L-PCA) for Feature Selection in Classification, in 2018 14th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), IEEE, 2018, pp. 745–751.
Morchid M, Dufour R, Bousquet P-M, Linares G, Torres-Moreno J-M. Feature selection using principal component analysis for massive retweet detection. Pattern Recognit Lett. 2014;49:33–9.
Zhu Y, Xie C, Sun B, Wang G-J, Yan X-G. Predicting China’s SME credit risk in supply chain financing by logistic regression, artificial neural network and hybrid models. Sustainability. 2016;8(5):433.
Zhang H, Shi Y, Yang X, Zhou R. A firefly algorithm modified support vector machine for the credit risk assessment of supply chain finance. Res Int Bus Financ. 2021;58:101482.
Islam S, Amin SH. Prediction of probable backorder scenarios in the supply chain using Distributed Random Forest and Gradient Boosting Machine learning techniques. J Big Data. 2020;7:1–22.
Nikolopoulos KI, Babai MZ, Bozos K. Forecasting supply chain sporadic demand with nearest neighbor approaches. Int J Prod Econ. 2016;177:139–48.
Yao G, Hu X, Zhou T, Zhang Y. Enterprise credit risk prediction using supply chain information: a decision tree ensemble model based on the differential sampling rate, synthetic minority oversampling technique and AdaBoost. Expert Syst. 2022;39(6):e12953.
Japkowicz N, Shah M. Performance evaluation in machine learning. Mach Learn Radiat Oncol Theory Appl, pp. 41–56, 2015.
Oyewola DO, Dada EG, Omotehinwa TO, Emebo O, Oluwagbemi OO. Application of deep learning techniques and bayesian optimization with tree parzen estimator in the classification of supply chain pricing datasets of health medications. Appl Sci. 2022;12(19):10166.
Arora K, Abbi P, Gupta PK. Analysis of Supply Chain Management Data using machine learning algorithms. in Innovative supply Chain Management via Digitalization and Artificial Intelligence. Springer; 2022. pp. 119–33.
Acknowledgements
This research is supported by the Researchers Supporting Project number (RSP2024R389), King Saud University, Riyadh, Saudi Arabia, and in part by National Natural Science Foundation of China with Grant No.72104020.
Funding
This research is supported by the Researchers Supporting Project number (RSP2024R389), King Saud University, Riyadh, Saudi Arabia, and in part by National Natural Science Foundation of China with Grant No.72104020.
Author information
Authors and Affiliations
Contributions
Ibrahim M. Hezam, Ahmed M. Ali and Mohamed Abdel-Basset: Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Conceptualization, Writing – original draft, Writing – review & editing. Ahmad M. Alshamrani: Formal analysis, Investigation, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing. Xuchong Gao: Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Conceptualization, Resources, Supervision, Writing – original draft, Writing – review & editing.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
This article does not contain any studies with human participants or animals performed by any of the authors.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Hezam, I.M., Ali, A.M., Alshamrani, A.M. et al. Artificial intelligence’s impact on drug delivery in healthcare supply chain management: data, techniques, analysis, and managerial implications. J Big Data 11, 177 (2024). https://doi.org/10.1186/s40537-024-01049-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s40537-024-01049-7