Introduction

Significant strides have been made in recent years in estimating the finite population mean, proportion, and variance using auxiliary information. However, the median, a more robust measure than the mean in the face of exceptionally low or high values, has received attention. Our research stands out by proposing a practical solution for estimating the median of a study variable for a finite population, even in the presence of neutrosophic forms of research and supporting variables with extreme values or outliers.

Several studies have made significant contributions to estimating a finite population's median. Gross1 used the sample median estimator in various sampling methods. Kuk and Mak2 proposed using the known median of the auxiliary variable to estimate the median of the study variable. Francisco and Fuller3 approximated the distribution function of a finite population using the median. Smarandache4 advocated using Neutrosophic statistics in uncertain systems. Singh et al.5 suggested a generalized family of median estimators in double sampling. These studies have laid the foundation for our research, which aims to build upon these findings and propose a practical solution for estimating the median of a study variable for a finite population.

A few chain ratio-type estimators were introduced by6 using the additional knowledge of the range of the auxiliary variable, whereas7 discussed the transformed ratio-type estimator using 's8 idea. Shokri9, 10, 58 presented new approaches to solving second-order initial value problems, providing effectiveness in addressing computational challenges. A generalized median estimator utilizing the transformed auxiliary variable was addressed by 11. Lamichhane et al.12 suggested a unique estimation for the finite population mean using the auxiliary variable's median.

Smarandache13 suggested that the sample size may not be accurately specified in neutrosophic statistics and hence presented the neutrosophic logic may fall within the interval \([a, \, b]\) (unidentified exactly). Sahin14 and Şahin15 proposed a new similarity measure based on falsity value between single-valued neutrosophic sets based on the centroid points of transformed single-valued neutrosophic numbers in decision-making. Shokri16 and Uluçay17 proposed the similarity measures of bipolar neutrosophic sets and their application to multiple criteria decision-making. Jan18 proposed multi-criteria decision-making for cubic linguistic information. Aslam19,20 explained the Neutrosophic analysis of variance on neutrosophic data. Using two supplementary variables,21 and22 presented difference-type median estimators for obtaining the population median. Chakraborty23 developed and categorized a trapezoidal bipolar neutrosophic number in decision-making.

For instance,24's development of complex neutrosophic fuzzy sets contributed to advancing the field of fuzzy sets. In addition, they offered an extensive flowchart of fuzzy sets with extensions, a description of their properties, and an explanation of how interval-valued neutrosophic sets function. Haque25 proposed a multi-criteria group decision-making strategy for the cylindrical neutrosophic domain.

Data visualization, analysis, and inference have long used classical statistics (CS). The CS investigates the assumption of data certainty. When the observations' measurements are precise, new methods for handling uncertain data are needed. Fuzzy logic works with data where the variable being researched lacks accurate measurements. They are quickly evolving and often used in settings where decisions are made. Fuzzy statistics examine data with ambiguous, opaque, or uncertain observations but neglect indeterminacy measurement. In this situation, showing a range of specific observations might be possible. The data included inside the indeterminacy interval cannot thus be analyzed using the CS. Neutrosophic statistical techniques are used to analyze the ambiguous neutrosophic data. The neutrosophic logic is used to interpret vague or unclear observations and allows for the measurement of indeterminacy and the determinate part of the observations.

Aslam26,27 elaborated on Neutrosophic Interval Statistics (NIS), Neutrosophic Applied Statistics (NAS), and Neutrosophic Statistical Quality control (NSQC), respectively. Uluçay28 suggested the idea of interval-valued refined neutrosophic sets and their applications. A large number of neutrosophic sets are described in the literature.

Tahir29 addressed a sampling gap by conducting a study to estimate a population's characteristics in a neutrosophic environment. They presented estimators of the neutrosophic ratio-type for estimating the finite population mean utilizing the additional information. Uluçay30,28 presented Q-neutrosophic soft graphs in operations management and communication networks. Vishwakarma and Singh31 proposed a neutrosophic ranked set sampling strategy for estimating the population mean under uncertainty using neutrosophic auxiliary information.

Table 1 illustrates the versatility and utility of neutrosophic logic in tackling problems across various disciplines, especially in situations where traditional binary logic falls short due to vagueness, uncertainty, or contradictory information.

Table 1 Some real-life problems and domains under neutrosophic logic.

Despite an extensive review of existing research, a study still needs to address the challenge of estimating the unknown population median in survey sampling when additional variables are introduced under neutrosophic information. The effectiveness of estimators remains to be determined in scenarios where the study variable and supporting variables take on neutrosophic forms, and the dataset includes extreme values or outliers. The lack of a practical solution for median estimation in such cases underscores the novelty and importance of our proposed neutrosophic median estimation method based on reliable measures of the auxiliary variable that are already known.

Neutrosophic statistics are applied in decision-making, risk assessment, uncertainty modelling, image processing, medical diagnosis, finance, and engineering for robust analysis as discussed in Table 1.

The paper is organized in a way that Section "The neutrosophic statistics" elaborates the details of Neutrosophic statistics along with symbols and notations. The adapted and proposed Neutrosophic median estimators along with the efficiency comparison are given in Section "The neutrosophic median estimators under simple random sampling". The numerical and graphical results related to real-life and simulated data sets are presented in Section "Real-life application". Finally, the proposed work is concluded in Section "Conclusion".

The neutrosophic statistics

Neutrosophic statistics, a unique approach, are designed to handle datasets with a degree of ambiguity or partial information. This method allows for conflicting beliefs and accommodates a range of uncertain numbers that may represent some observations, including an exact measurement. In contrast, traditional statistics falter when faced with uncertainty. This is where the intriguing potential of neutrosophic statistics comes into play, offering a fresh perspective on data analysis.

In real-world problems, the population parameters are often unknown. In such cases, statistical inference methods may need to be more practical. Instead, acceptable estimates are used, resolving the issue of an unknown parameter value by estimating its values. This pragmatic approach reassures the statistician that the derived data are vague but still useful. Neutrosophic statistics, with their ability to calculate the best interval value with the minimum mean square error, offer a reliable solution to these problems.

Previous study efforts provided a limited range of neutrosophic observations, including quantifiable neutrosophic data. Furthermore32,33,34,35,36,37,38,39,40 discussed numerous approaches, such as interval-based approaches, Triangular or trapezoidal fuzzy numbers, and single-valued fuzzy numbers, exist to express the range of neutrosophic numbers along with Optimal trajectories in reproduction models of economic dynamics. He41 proposed a fractal model for internal temperature response in porous concrete, advancing understanding in applied mathematics. Iskandarov and Komartsova42,57 investigated integral perturbations' influence on boundedness in fourth-order linear differential equations. Khankishiyev43 employed finite differences to solve loaded differential equations, while 44, 56 explored dark energy solutions without a cosmological constant. Furthermore,45 established conditions for complete monotonicity in the differential functions involving trigamma functions.

Let the neutrosophic range is \(T_{N} = T_{L} + T_{U} \ell_{N}\) with \(\ell_{N} \in [\ell_{L} , \, \ell_{U} ]\), the neutrosophic variable \(T_{N}\) indicates the neutrosophic samples selected from a population having imprecise, ambiguous and unclear measurements. Thus, for the neutrosophic data in the interval form, we use notation \(T_{N} \in [a, \, b],\) where \(a\) and \(b\) are the lower and upper values of the neutrosophic data, respectively.

Figure 1 depicts the approach to applying the proposed estimation methods in neutrosophic statistics. This workflow developed a few neutrosophic estimators to estimate the finite population median in the presence of supplementary data, which are well suited for overcoming the sample indeterminacy problem.

Figure 1
figure 1

Workflow of the parameter estimation.

Symbols and notations

Suppose a neutrosophic random sample of size \(n_{N} \in [n_{L} , \, n_{U} ],\) selected from a finite population \(U = \left\{ {U_{1} ,U_{2} ,...,U_{N} } \right\}\) of \(N\) recognisable units. Assume \(y_{N} (i)\) and \(x_{N} (i),\)\(i = (1,2,...,n)\) represent the \(i^{th}\) sampled unit's values of the neutrosophic data of the study variable \(y_{N(i)} \in [y_{L} , \, y_{U} ]\) and the supplementary variable \(x_{N(i)} \in [x_{L} , \, x_{U} ]\), respectively. Let \(M_{yN}\) be the neutrosophic variable of interest and \(\hat{M}_{yN}\) and \(\hat{M}_{xN}\) represent the sample median that corresponds to the population median \(M_{yN}\) and \(M_{xN}\) respectively with the probability density functions \(f_{yN} (M_{yN} )\) and \(f_{xN} (M_{xN} ),\) respectively, where \(\hat{M}_{yN} \in [\hat{M}_{yL} , \, \hat{M}_{yU} ]\) and \(\hat{M}_{xN} \in [\hat{M}_{xL} , \, \hat{M}_{xU} ].\) Let \(\rho_{yxN} = \rho_{{(\hat{M}_{yN} ,\hat{M}_{xN} )}} = 4\rho_{11} (y_{N} ,x_{N} ) - 1\) represent the neutrosophic population correlation coefficient between the neutrosophic sample medians ranging from \(- 1\) to \(+ 1\) as \(\rho_{11}\) increases from 0 to 0.5, where \(\rho_{yxN} \in [\rho_{yxL} , \, \rho_{yxU} ]\) such that \(P_{11} (y_{N} ,x_{N} ) = P(y_{N} \le M_{yN} \cap x_{N} \le M_{xN} ).\) Similarly, \(Bias(\hat{M}_{yN} ) \in [Bias_{L} , \, Bias_{U} ]\) and \(MSE(\hat{M}_{yN} ) \in [MSE_{L} , \, MSE_{U} ]\) belong to the neutrosophic sets.

To get the characteristics of estimators, the relative error terms are defined as follows: Suppose \(e_{{0N}} = (\hat{M}_{{yN}} - M_{{yN}} )/M_{{yN}}\) \(e_{1N} = (\hat{M}_{xN} - M_{xN} )/M_{xN}\) are the neutrosophic errors where \(e_{0N} \in [e_{0L} , \, e_{0U} ]\) and \(e_{1N} \in [e_{1L} , \, e_{1U} ]\) such that \(E(e_{iN} ) = 0 \, (i = 0,1),\) \(E(e_{0N}^{2} ) = \lambda_{N} C_{{M_{yN} }}^{2} ,\) \(E(e_{1N}^{2} ) = \lambda_{N} C_{{M_{xN} }}^{2} ,\) \(E(e_{0N} e_{1N} ) = \lambda_{N} C_{{M_{yxN} }} ,\) where \(C_{{M_{yxN} }} = \rho_{yxN} C_{{M_{yN} }} C_{{M_{xN} }} ,\) \(C_{{M_{yN} }} = \left\{ {M_{yN} f_{yN} (M_{yN} )} \right\}^{ - 1} ,\) \(C_{{M_{xN} }} = \left\{ {M_{xN} f_{xN} (M_{xN} )} \right\}^{ - 1}\) are the neutrosophic coefficients of variation, respectively. Let \(C_{{M_{yxN} }} \in [C_{{M_{yxL} }} , \, C_{{M_{yxU} }} ]\), \(C_{{M_{yN} }} \in [C_{{M_{yL} }} , \, C_{{M_{yU} }} ]\),\(C_{{M_{xN} }} \in [C_{{M_{xL} }} , \, C_{{M_{xU} }} ]\) and \(\lambda_{N} = \frac{1}{4}\left( {\frac{1}{{n_{N} }} - \frac{1}{{N_{N} }}} \right),\) where \(\lambda_{N} \in [\lambda_{L} , \, \lambda_{U} ]\).

The neutrosophic median estimators under simple random sampling

First, we present a few adapted neutrosophic median estimators using auxiliary information under simple random sampling to address uncertainty and neutrosophic data.

Adapted median estimators with auxiliary variable

  1. (i)

    Motivated by1, we propose a neutrosophic traditional median estimator and its variance, along with the expression of variance is given by

    $$\hat{M}_{0N} = \hat{M}_{yN}$$
    (1)
    $$Var(\hat{M}_{0N} ) = \lambda_{N} M_{yN}^{2} C_{{M_{yN} }}^{2}$$
    (2)
  2. (ii)

    Motivated by2, we developed a novel neutrosophic traditional ratio estimator, along with the expressions of Bias and MSE are

    $$\hat{M}_{RN} = \hat{M}_{yN} \left( {\frac{{M_{xN} }}{{\hat{M}_{xN} }}} \right)$$
    (3)
    $$Bias(\hat{M}_{RN} ) \cong \lambda_{N} M_{yN} \left\{ {C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} } \right\}$$
    (4)

and

$$MSE(\hat{M}_{RN} ) \cong \lambda_{N} M_{yN}^{2} \left\{ {C_{{M_{yN} }}^{2} + C_{{M_{xN} }}^{2} - 2C_{{M_{yxN} }} } \right\}$$
(5)

The ratio estimator (\(\hat{M}_{RN}\)) performs better than \(\hat{M}_{0N}\) if \(\rho_{yxN} > 0.5\frac{{C_{{M_{xN} }} }}{{C_{{M_{yN} }} }}.\)

  1. (iii)

    Motivated by46, the neutrosophic exponential ratio-type estimator, along with the expressions of Bias and MSE are given by

    $$\hat{M}_{EN} = \hat{M}_{yN} \exp \left( {\frac{{M_{xN} - \hat{M}_{xN} }}{{M_{xN} + \hat{M}_{xN} }}} \right)$$
    (6)
    $$Bias(\hat{M}_{EN} ) \cong M_{yN} \lambda_{N} \left( {\frac{3}{8}C_{{M_{xN} }}^{2} - \frac{1}{2}C_{{M_{yxN} }} } \right)$$
    (7)

and

$$MSE(\hat{M}_{EN} ) \cong M_{yN}^{2} \lambda_{N} \left( {C_{{M_{yN} }}^{2} + \frac{1}{4}C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} } \right)$$
(8)

The exponential ratio estimator (\(\hat{M}_{EN}\)) is more efficient than \(\hat{M}_{0N}\) and \(\hat{M}_{RN}\) if \(\rho_{yxN} > 0.25\frac{{C_{{M_{xN} }} }}{{C_{{M_{yN} }} }}\) and \(\rho_{yxN} < 0.75\frac{{C_{{M_{xN} }} }}{{C_{{M_{yN} }} }},\) respectively.

  1. (iv)

    The adapted neutrosophic difference estimator along with the expression of variance is given by

    $$\hat{M}_{{D_{0N} }} = \hat{M}_{yN} + d_{0N} (M_{xN} - \hat{M}_{xN} )$$
    (9)

At the optimal value of \(d_{0N} ,\) which is \(d_{0N(opt)} = \frac{{M_{yN} \rho_{yxN} C_{{M_{yN} }} }}{{M_{xN} C_{{M_{xN} }} }},\) the minimum MSE of \(\hat{M}_{{D_{0N} }} ,\) is given by

$$Var(\hat{M}_{{D_{0N} }} )_{\min } \cong M_{yN}^{2} C_{{M_{yN} }}^{2} \lambda_{N} \left( {1 - \rho_{yxN}^{2} } \right)$$
(10)
  1. (xxii)

    Adapted from22, difference-type estimators, along with the expressions of Bias and minimum mean square errors, are given by

    $$\hat{M}_{{D_{1N} }} = d_{1N} \hat{M}_{yN} + d_{2N} (M_{xN} - \hat{M}_{xN} )$$
    (11)
    $$\hat{M}_{{D_{2N} }} = \left\{ {d_{3N} \hat{M}_{yN} + d_{4N} (M_{xN} - \hat{M}_{xN} )} \right\}\left( {\frac{{M_{xN} }}{{\hat{M}_{xN} }}} \right),$$
    (12)
    $$\hat{M}_{{D_{3N} }} = \left\{ {d_{5N} \hat{M}_{yN} + d_{6N} (M_{xN} - \hat{M}_{xN} )} \right\}\exp \left( {\frac{{M_{xN} - \hat{M}_{xN} }}{{M_{xN} + \hat{M}_{xN} }}} \right),$$
    (13)

and

$$\hat{M}_{{D_{4N} }} = \left\{ {d_{7N} \hat{M}_{yN} + d_{8N} (M_{xN} - \hat{M}_{xN} )} \right\}\exp \left( {\frac{{M_{xN} }}{{\hat{M}_{xN} }} - 1} \right),$$
(14)

where \(d_{iN} (i = 1 - 8)\) are constants determined below by optimality considerations as

\(d_{1N(opt)} = \frac{{B_{0N} }}{{A_{0N} B_{0N} - C_{0N}^{2} + B_{0N} }},\) \(d_{2N(opt)} = \frac{{M_{yN} }}{{M_{xN} }}\frac{{C_{0N} }}{{A_{0N} B_{0N} - C_{0N}^{2} + B_{0N} }},\)

\(d_{3N(opt)} = \frac{{B_{1N} (C_{1N} - D_{1N} + 1)}}{{A_{1N} B_{1N} - D_{1N}^{2} + B_{1N} }},\) \(d_{4N(opt)} = \frac{{M_{yN} }}{{M_{xN} }}\frac{{(A_{1N} B_{1N} - C_{1N} D_{1N} + B_{1N} - D_{1N} )}}{{(A_{1N} B_{1N} - D_{1N}^{2} + B_{1N} )}},\)

\(d_{5N(opt)} = \frac{{(B_{2N} C_{2N} - D_{2N} E_{2N} + B_{2N} )}}{{(A_{2N} B_{2N} - E_{2N}^{2} + B_{2N} )}},\) \(d_{6N(opt)} = \frac{{M_{yN} }}{{M_{xN} }}\frac{{(A_{2N} D_{2N} - C_{2N} E_{2N} + D_{2N} - E_{2N} )}}{{(A_{2N} B_{2N} - E_{2N}^{2} + B_{2N} )}},\)

\(d_{7N(opt)} = \frac{{B_{3N} (C_{3N} - D_{3N} + 1)}}{{A_{3N} B_{3N} - D_{3N}^{2} + B_{3N} }},\) \(d_{8N(opt)} = \frac{{M_{yN} }}{{M_{xN} }}\frac{{(A_{3N} B_{3N} - C_{3N} D_{3N} + B_{3N} - D_{3N} )}}{{(A_{3N} B_{3N} - D_{3N}^{2} + B_{3N} )}},\)where \(A_{0N} = \lambda_{N} C_{{M_{yN} }}^{2} ,\) \(B_{0N} = \lambda_{N} C_{{M_{xN} }}^{2} ,\) \(C_{0N} = \lambda_{N} C_{{M_{yxN} }} ,\) \(A_{1N} = \lambda_{N} (C_{{M_{yN} }}^{2} + 3C_{{M_{xN} }}^{2} - 4C_{{M_{yxN} }} ),\) \(B_{1N} = \lambda_{N} C_{{M_{xN} }}^{2} ,\) \(C_{1N} = \lambda_{N} (C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} ),\) \(D_{1N} = \lambda_{N} (2C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} ),\) \(A_{2N} = \lambda_{N} (C_{{M_{yN} }}^{2} + C_{{M_{xN} }}^{2} - 2C_{{M_{yxN} }} ),\) \(B_{2N} = \lambda_{N} C_{{M_{xN} }}^{2} ,\) \(C_{2N} = \lambda_{N} \left( {\frac{3}{8}C_{{M_{xN} }}^{2} - \frac{1}{2}C_{{M_{yxN} }} } \right),\) \(D_{2N} = \lambda_{N} C_{{M_{xN} }}^{2} /2,\) \(E_{2N} = \lambda_{N} (C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} ),\) \(A_{3N} = \lambda_{N} (C_{{M_{yN} }}^{2} + 4C_{{M_{xN} }}^{2} - 4C_{{M_{yxN} }} ),\)\(B_{3N} = \lambda_{N} C_{{M_{xN} }}^{2} ,\) \(C_{3N} = \lambda_{N} \left( {\frac{3}{2}C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} } \right)\) and \(D_{3N} = \lambda_{N} (2C_{{M_{xN} }}^{2} - C_{{M_{yxN} }} ).\)

$$Bias(\hat{M}_{{D_{1N} }} ) \cong (d_{1N} - 1)M_{yN} ,$$
(15)
$$Bias(\hat{M}_{{D_{2N} }} ) \cong (d_{3N} - 1)M_{yN} + d_{3N} M_{yN} C_{1N} + d_{4N} M_{xN} B_{1N} ,$$
(16)
$$Bias(\hat{M}_{{D_{3N} }} ) \cong (d_{5N} - 1)M_{yN} + d_{5N} M_{yN} C_{2N} + d_{6N} M_{xN} D_{2N} ,$$
(17)
$$Bias(\hat{M}_{{D_{4N} }} ) \cong (d_{7N} - 1)M_{yN} + d_{7N} M_{yN} C_{3N} + d_{8N} M_{xN} B_{3N} ,$$
(18)
$$MSE(\hat{M}_{{D_{1N} }} )_{\min } \cong M_{yN}^{2} \left\{ {1 - \frac{{B_{0N} }}{{A_{0N} B_{0N} - C_{0N}^{2} + B_{0N} }}} \right\},$$
(19)
$$MSE(\hat{M}_{{D_{2N} }} )_{\min } \cong M_{yN}^{2} \left\{ {1 - \frac{{A_{1N} B_{1N}^{2} + B_{1N} C_{1N}^{2} - 2B_{1N} C_{1N} D_{1N} + B_{1N}^{2} + 2B_{1N} C_{1N} - 2B_{1N} D_{1N} + B_{1N} }}{{A_{1N} B_{1N} - D_{1N}^{2} + B_{1N} }}} \right\},$$
(20)
$$MSE(\hat{M}_{{D_{3N} }} )_{\min } \cong M_{yN}^{2} \left\{ {1 - \frac{{A_{2N} D_{2N}^{2} + B_{2N} C_{2N}^{2} - 2C_{2N} D_{2N} E_{2N} + 2B_{2N} C_{2N} + D_{2N}^{2} - 2D_{2N} E_{2N} + B_{2N} }}{{A_{2N} B_{2N} - E_{2N}^{2} + B_{2N} }}} \right\},$$
(21)

and

$$MSE(\hat{M}_{{D_{4N} }} )_{\min } \cong M_{yN}^{2} \left\{ {1 - \frac{{A_{3N} B_{3N}^{2} + B_{3N} C_{3N}^{2} - 2B_{3N} C_{3N} D_{3N} + B_{3N}^{2} + 2B_{3N} C_{3N} - 2B_{3N} D_{3N} + B_{3N} }}{{A_{3N} B_{3N} - D_{3N}^{2} + B_{3N} }}} \right\}.$$
(22)

The proposed generalized neutrosophic median estimator

Traditional estimators, often hindered by their reliance on historical data, struggle with accuracy, particularly with outliers. This section introduces advanced neutrosophic estimators for accurately predicting a finite population's median. These estimators blend unique metrics like quartile deviation and interquartile range, enhancing data distribution analysis and outlier exclusion through robust scaling, employing decile means, the Hodges-Lehmann estimator, and tri-mean for reliable median estimation. The tri-mean proposed by47, the Hodges-Lehmann estimator proposed by48 and the decile means proposed by49 are the three robust metrics we used in this study. For further information about these robust measures, readers can see50 and51 for details.

Motivated by52, we develop a neutrosophic generalized ratio-type estimator of finite population median as

$$T_{i(d)N} = \hat{M}_{yN} \left[ {\left\{ {m_{1N} \left( {\frac{{\psi_{N} \hat{M}_{xN} + \delta_{N} }}{{\psi_{N} M_{xN} + \delta_{N} }}} \right)^{{\alpha_{3} }} \exp \left( {\frac{{M_{xN} - \hat{M}_{xN} }}{{M_{xN} + \hat{M}_{xN} }}} \right)} \right\} + \left\{ {m_{2N} \left( {\frac{{\psi_{N} M_{xN} + \delta_{N} }}{{\psi_{N} \hat{M}_{xN} + \delta_{N} }}} \right)^{{\alpha_{4} }} } \right\}} \right],$$
(23)

where \(T_{i(d)N} \in [T_{i(d)L} , \, T_{i(d)U} ],\) \(m_{1N} \in [m_{1L} , \, m_{1U} ]\) and \(m_{2N} \in [m_{2L} , \, m_{2U} ]\) are suitable neutrosophic constants, where \(\alpha_{3} {\text{ and }}\alpha_{4}\) take the values \(1,\,\, - 1,\,\,2,\,\, - 2\) for developing new estimators.

Note \(\psi_{N} \in [\psi_{L} , \, \psi_{U} ]\) and \(\delta_{N} \in [\delta_{L} , \, \delta_{U} ]\) are neutrosophic functions of the known robust and non-conventional measures related to the variable \(X_{N}\). Robust measures associated with \(X_{N}\) are:

  1. (i)

    Tri-mean:\(T_{MN} = (Q_{1N} + 2Q_{2N} + Q_{3N} )/4\), \(T_{MN} \in [T_{ML} , \, T_{MU} ]\)

  2. (ii)

    Hodges–Lehman: \(H_{LN} = Median((x_{jN} + x_{kN} )/2),\) \(1 \le j \le k \le N,\) \(H_{LN} \in [H_{LL} , \, H_{LU} ]\)

  3. (iii)

    Decile mean: \(D_{MN} = \sum\limits_{i = 1}^{9} {D_{iN} /9} ,\) \(D_{MN} \in [D_{ML} , \, D_{MU} ]\)

The non-conventional measures (i.e., interquartile range, midrange, quartile average and quartile deviation) of the supplementary variable are as follows:

  1. (iv)

    Interquartile range: \(Q_{RN} = Q_{3N} - Q_{1N}\), \(Q_{RN} \in [Q_{RL} , \, Q_{RU} ]\)

  2. (v)

    Midrange:\(M_{RN} = ((x_{(1)N} + x_{(N)N} )/2)\), \(M_{RN} \in [M_{RL} ,M_{RU} ]\)

  3. (vi)

    Quartile average:\(Q_{AN} = (Q_{3N} + Q_{1N} )/2\), \(Q_{AN} \in [Q_{AL} , \, Q_{AU} ]\)

  4. (vii)

    Quartile deviation \(Q_{DN} = (Q_{3N} - Q_{1N} )/2,\)\(Q_{DN} \in [Q_{DL} , \, Q_{DU} ]\)

where \(Q_{1N} \in [Q_{1L} , \, Q_{1U} ]\), \(Q_{2N} \in [Q_{2L} , \, Q_{2U} ]\) and \(Q_{3N} \in [Q_{3L} , \, Q_{3U} ]\) are the neutrosophic first, second and third quartiles, respectively and \(D_{iN} \in [D_{iL} , \, D_{iU} ]\) is the neutrosophic decile.

By putting different values of \(\alpha_{i} \, ({\text{for }}i = 3,4)\) into (23), we get the following families of estimators as.

  • i. At \(\alpha_{3} = 1\) and \(\alpha_{4} = 2\), the proposed family of estimators reduces to

    $$T_{{i(d)N}}^{{ \ominus }} = \hat{M}_{{yN}} \left[ {\left\{ {m_{{1N}} \left( {\frac{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}{{\psi _{N} M_{{xN}} + \delta _{N} }}} \right)\exp \left( {\frac{{M_{{xN}} - \hat{M}_{{xN}} }}{{M_{{xN}} + \hat{M}_{{xN}} }}} \right)} \right\} + \left\{ {m_{{2N}} \left( {\frac{{\psi _{N} M_{{xN}} + \delta _{N} }}{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}} \right)^{2} } \right\}} \right].$$
    (24)
  • ii. At \(\alpha_{3} = - 1\) and \(\alpha_{4} = - 1;\) the proposed family of estimators reduces to

    $$T_{{i(d)N}}^{ \oplus } = \hat{M}_{{yN}} \left[ {\left\{ {m_{{1N}} \left( {\frac{{\psi _{N} M_{{xN}} + \delta _{N} }}{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}} \right)\exp \left( {\frac{{M_{{xN}} - \hat{M}_{{xN}} }}{{M_{{xN}} + \hat{M}_{{xN}} }}} \right)} \right\} + \left\{ {m_{{2N}} \left( {\frac{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}{{\psi _{N} M_{{xN}} + \delta _{N} }}} \right)} \right\}} \right].$$
    (25)
  • iii. At \(\alpha_{3} = - 1\) and \(\alpha_{4} = - 2\), the proposed family of estimators becomes

    $$T_{{i(d)N}}^{ \otimes } = \hat{M}_{{yN}} \left[ {\left\{ {m_{{1N}} \left( {\frac{{\psi _{N} M_{{xN}} + \delta _{N} }}{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}} \right)\exp \left( {\frac{{M_{{xN}} - \hat{M}_{{xN}} }}{{M_{{xN}} + \hat{M}_{{xN}} }}} \right)} \right\} + \left\{ {m_{{2N}} \left( {\frac{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}{{\psi _{N} M_{{xN}} + \delta _{N} }}} \right)^{2} } \right\}} \right].$$
    (26)
  • iv. At \(\alpha_{3} = 2\) and \(\alpha_{4} = 2\), the proposed family of estimators reduces to

    $$T_{{i(d)N}}^{{ \circledast }} = \hat{M}_{{yN}} \left[ {\left\{ {m_{{1N}} \left( {\frac{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}{{\psi _{N} M_{{xN}} + \delta _{N} }}} \right)^{2} \exp \left( {\frac{{M_{{xN}} - \hat{M}_{{xN}} }}{{M_{{xN}} + \hat{M}_{{xN}} }}} \right)} \right\} + \left\{ {m_{{2N}} \left( {\frac{{\psi _{N} M_{{xN}} + \delta _{N} }}{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}} \right)^{2} } \right\}} \right].$$
    (27)
  • v. At \(\alpha_{3} = - 2\) and \(\alpha_{4} = - 1\), the proposed family of estimators reduces to

    $$T_{{i(d)N}}^{{ \circledcirc }} = \hat{M}_{{yN}} \left[ {\left\{ {m_{{1N}} \left( {\frac{{\psi _{N} M_{{xN}} + \delta _{N} }}{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}} \right)^{2} \exp \left( {\frac{{M_{{xN}} - \hat{M}_{{xN}} }}{{M_{{xN}} + \hat{M}_{{xN}} }}} \right)} \right\} + \left\{ {m_{{2N}} \left( {\frac{{\psi _{N} \hat{M}_{{xN}} + \delta _{N} }}{{\psi _{N} M_{{xN}} + \delta _{N} }}} \right)} \right\}} \right].$$
    (28)

When we use robust measures with linear combinations of the median, quartile deviation, midrange, interquartile range, and quartile average of the supplementary variable in (23), we get different series of estimators such as \(T^{{\begin{array}{*{20}l} { \ominus } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} \oplus \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} \otimes \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledast } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\) and \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledcirc } \hfill \\ \end{array} } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N} .\) Few members of the family of estimators \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledcirc } \hfill \\ \end{array} } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\) are given in Table 2. Putting the same values of \(\psi_{N}\) and \(\delta_{N}\) in \(T^{{\begin{array}{*{20}l} { \ominus } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} \oplus \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} \otimes \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\), \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledast } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\) and \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledcirc } \hfill \\ \end{array} } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\), we obtain several estimators.

Table 2 Some members of \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledcirc } \hfill \\ \end{array} } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\).

We can obtain several optimal estimators by placing suitable constants or known conventional parameters of the supplementary variable in place of \(\psi_{N}\) and \(\delta_{N}\) into (23). Conventional parameters related to the supplementary variable \(X_{N}\) are variance, standard deviation, coefficient of variation, coefficient of skewness, coefficient of kurtosis, coefficient of correlation, and so forth.

Bias, MSE, and minimum MSE of the proposed neutrosophic generalized family of estimators \(T_{i(d)N}\) in terms of \(e_{oN}\) and \(e_{1N}\) are expressed as,

$$Bias(T_{i(d)N} ) = M_{yN} \left[ {m_{1N} \left( {1 + \frac{{\lambda_{N} C_{{M_{xN} }}^{2} }}{2}\left( {\frac{3}{4} - \alpha_{3} \theta_{N} + \alpha_{3} (\alpha_{3} - 1)\theta_{N}^{2} } \right) + \lambda_{N} \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} \left( {\alpha_{3} \theta_{N} - \frac{1}{2}} \right)} \right)} \right.\left. { + m_{2N} \left( {1 + \frac{{\alpha_{4} (\alpha_{4} + 1)\theta_{N}^{2} \lambda_{N} C_{{M_{xN} }}^{2} }}{2} - \alpha_{4} \theta_{N} \lambda_{N} \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} } \right) - 1} \right]$$
(29)

where \(\theta_{N} = \psi_{N} M_{xN} /(\psi_{N} M_{xN} + \delta_{N} )\).

The MSE of suggested estimator up to the first order of approximation as

$$MSE(T_{i(d)N} ) \cong M_{yN}^{2} [1 + m_{1N}^{2} A_{1N} + m_{2N}^{2} A_{2N} + 2m_{1N} m_{2N} A_{3N} - 2m_{1N} A_{4N} - 2m_{2N} A_{5N} ],$$
(30)

where \(A_{1N} = \left[ {1 + \lambda_{N} \left\{ {C_{{M_{yN} }}^{2} + C_{{M_{xN} }}^{2} (1 + 2\alpha_{3}^{2} \theta_{N}^{2} - \alpha_{3} \theta_{N}^{2} - 2\alpha_{3} \theta_{N} ) - 2\rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} (1 - 2\alpha_{3} \theta_{N} )} \right\}} \right]\)

$$A_{2N} = [1 + \lambda_{N} \{ C_{{M_{yN} }}^{2} + \theta_{N}^{2} C_{{M_{xN} }}^{2} (2\alpha_{4}^{2} + \alpha_{4} ) - 4\alpha_{4} \theta_{N} \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} \} ],$$
$$A_{3N} = \left[ {1 + \lambda_{N} \left\{ {C_{{M_{yN} }}^{2} - \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} (2\theta_{N} (\alpha_{4} - \alpha_{3} ) + 1) - C_{{M_{xN} }}^{2} \left( {\alpha_{3} \alpha_{4} \theta_{N}^{2} - \frac{3}{8} + \frac{{\alpha_{3} \theta_{N} }}{2}} \right.} \right.} \right.$$
$$\, \left. {\left. {\left. { - \frac{{\alpha_{4} \theta_{N} }}{2} - \frac{{\alpha_{4} (\alpha_{4} + 1)\theta_{N}^{2} }}{2} - \frac{1}{2}\alpha_{3} (\alpha_{3} - 1)\theta_{N}^{2} } \right)} \right\}} \right]$$

\(A_{4N} = \left[ {1 + \lambda_{N} \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} \left( {\alpha_{3} \theta_{N} - \frac{1}{2}} \right) + \lambda_{N} \frac{{C_{{M_{xN} }}^{2} }}{2}\left( {\frac{3}{4} - \alpha_{3} \theta_{N} + \alpha_{3} (\alpha_{3} - 1)\theta_{N}^{2} } \right)} \right]\) and \(A_{5N} = \left[ {1 + \frac{1}{2}\alpha_{4} (\alpha_{4} + 1)\theta_{N}^{2} \lambda_{N} C_{{M_{xN} }}^{2} - \alpha_{4} \theta_{N} \lambda_{N} \rho_{N} C_{{M_{yN} }} C_{{M_{xN} }} } \right].\)

The minimum MSE at the optimum values \(m_{1N(opt)} = \frac{{(A_{2N} A_{4N} - A_{3N} A_{5N} )}}{{(A_{1N} A_{2N} - A_{3N}^{2} )}}\) and \(m_{2N(opt)} = \frac{{(A_{1N} A_{5N} - A_{3N} A_{4N} )}}{{(A_{1N} A_{2N} - A_{3N}^{2} )}},\) is given by

$$MSE(T_{i(d)N} )_{\min } \cong M_{yN}^{2} \left[ {1 - \frac{{(A_{2N} A_{4N}^{2} + A_{1N} A_{5N}^{2} - 2A_{3N} A_{4N} A_{5N} )}}{{(A_{1N} A_{2N} - A_{3N}^{2} )}}} \right]$$
(31)

Efficiency comparison

(i). By comparing (2) and (23), \(Var(\hat{M}_{0N} ) > MSE(T_{i(d)N} )_{\min }\) if \(\left[ {\Theta_{2N} \left( {\lambda_{N} M_{yN}^{2} C_{{M_{yN} }}^{2} - 1} \right) + \Theta_{1N} } \right]\, > 0\), where \(\Theta_{1N} = A_{2N} A_{4N}^{2} + A_{1N} A_{5N}^{2} - 2A_{3N} A_{4N} A_{5N}\) and \(\Theta_{2N} = A_{1N} A_{2N} - A_{3N}^{2}\).

(ii) By comparing (5) and (23), \(MSE(\hat{M}_{RN} ) > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\Theta_{2N} \left( {\lambda_{hN} C_{{M_{yN} }}^{2} + \lambda_{N} C_{{M_{xN} }}^{2} - 2\lambda_{N} C_{{M_{yxN} }} - 1} \right) + \Theta_{1N} } \right]\, > 0$$

(iii) By comparing (8) and (23), \(MSE(\hat{M}_{EN} ) > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\Theta_{2N} \left( {\lambda_{N} C_{{M_{yN} }}^{2} + \frac{1}{4}\lambda_{N} C_{{M_{xN} }}^{2} - \lambda_{N} C_{{M_{yxN} }} - 1} \right) + \Theta_{1N} } \right]\, > 0$$

(iv) By comparing (10) and (23), \(Var(\hat{M}_{{D_{0} N}} )_{\min } > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {C_{{M_{yN} }}^{2} \lambda_{N} \left( {1 - \rho_{yxN}^{2} } \right) - 1 + \frac{{\Theta_{1N} }}{{\Theta_{2N} }}} \right] \, > 0$$

(v) By comparing (19) and (23), \(MSE(\hat{M}_{{D_{1} N}} )_{\min } > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\frac{{\Theta_{1N} }}{{\Theta_{2N} }} - \frac{{B_{0N} }}{{A_{0N} B_{0N} - C_{0N}^{2} + B_{0N} }}} \right]\, > 0$$

(vi) By comparing (20) and (23), \(MSE(\hat{M}_{{D_{2} N}} )_{\min } > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\frac{{\Theta_{1N} }}{{\Theta_{2N} }} - \frac{{\left( {A_{1N} B_{1N}^{2} + B_{1N} C_{1N}^{2} - 2B_{1N} C_{1N} D_{1N} + B_{1N}^{2} + 2B_{1N} C_{1N} - 2B_{1N} D_{1N} + B_{1N} } \right)}}{{A_{1N} B_{1N} - D_{1N}^{2} + B_{1N} }}} \right]\, > 0$$

(vii) By comparing (21) and (23), \(\,MSE(\hat{M}_{{D_{3} N}} )_{\min } > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\frac{{\Theta_{1N} }}{{\Theta_{2N} }} - \frac{{A_{2N} D_{2N}^{2} + B_{2N} C_{2N}^{2} - 2C_{2N} D_{2N} E_{2N} + 2B_{2N} C_{2N} + D_{2N}^{2} - 2D_{2N} E_{2N} + B_{2N} }}{{A_{2N} B_{2N} - E_{2N}^{2} + B_{2N} }}} \right]\, > 0$$

(vii) By comparing (22) and (23), \(MSE(\hat{M}_{{D_{4} N}} )_{\min } > MSE(T_{i(d)N} )_{\min }\) if

$$\left[ {\frac{{\Theta_{1N} }}{{\Theta_{2N} }} - \frac{{A_{3N} B_{3N}^{2} + B_{3N} C_{3N}^{2} - 2B_{3N} C_{3N} D_{3N} + B_{3N}^{2} + 2B_{3N} C_{3N} - 2B_{3N} D_{3N} + B_{3N} }}{{A_{3N} B_{3N} - D_{3N}^{2} + B_{3N} }}} \right]\, > 0$$

Hence, robustness is evaluated in this case to compare the proposed neutrosophic generalized estimators with other neutrosophic estimators in (1), (3), (6), (9), (11), (12), (13), and (14) to find the more effective neutrosophic median estimator. Additionally, we use real-world datasets to determine the relative effectiveness of different estimators.

Real-life application

In terms of relative efficiency, we compare the suggested family of estimators' performance to that of other competitive estimators. We chose two real-world indeterminacy interval datasets for this purpose.

Regarding relative efficiency, we compare the suggested family of estimators' performance to that of other competitive estimators. For this purpose, we chose two real-world indeterminacy interval datasets.

The first one is the Daily stock price, which is used as a neutrosophic variable because, on each day, a stock's price fluctuates between an opening price (the price during which trade begins) and a closing price (the price at which trade stops for the day). The price constantly fluctuates between a high (the highest price of the day) and a low (the lowest price), which may or may not be similar to the opening or closing price. We estimate the high and low price intervals within which the stock price falls by utilizing the daily starting price as a supplementary variable that is not a neutrosophic variable since its value is set and known for each day.

Population I Source: 53 (1st June 2022 to 29th July 2022) from the link: https://finance.yahoo.com/quote/SMSN.IL/history?p=SMSN.IL. \(Y_{N} =\) Low & High prices; \(X_{N} =\) Opening price.

Population II Source: 54 (1st Feb 2022 to 29th July 2022) from the link: https://finance.yahoo.com/quote/SZKMY/history?p=SZKMY. \(Y_{N} =\) Low & High prices; \(X_{N} =\) Opening price, where \(Y_{N} \in (Y_{L} ,Y_{U} )\) corresponds to the independent determinate variable \(X_{N} \in (X_{L} ,X_{U} )\).

Figures 2 and 3 show the trend of real-world data sets using box plots, which aids in displaying the skewness of the data. The minimum score (Lower Fence), the lower quartile, the median, the upper quartile, and the maximum score summarise data using boxplots (Upper Fence). As shown in Fig. 2, our positively skewed data suggests that the median is closer to the lower quartile. The boxplot with points outside the whiskers shows a few outliers in the data.

Figure 2
figure 2

Boxplot of Populations I and II for each variable.

Figure 3
figure 3

Normal Q-Q plot of Populations I and II for variable \(X_{N}\) and \(Y_{N}\).

Figure 3 elaborates on the pivotal role of Q-Q (quantile–quantile) plots in statistics. These plots facilitate the graphical comparison of two probability distributions through their quantiles. They are instrumental in determining the distribution type of a random variable, spotting outliers, and assessing skewness. By plotting theoretical quantiles against sample quantiles, Q-Q plots reveal distribution traits, including skewness. Notably, deviations in the plot’s upper end and a pronounced right tail indicate a right-skewed distribution, as demonstrated in standard Q-Q plot interpretations.

In addition, the following formulae are used to get the percentage relative efficiency (PRE) \(PRE(\hat{M}_{0N} ,\hat{M}_{iN} ) = \frac{{Var(\hat{M}_{0N} )}}{{MSE(\hat{M}_{iN} )}} \times 100,\) \(({\text{for }}i = 0, \, R, \, E, \, D_{0} , \, D_{1} , \, D_{2} , \, D_{3} {\text{ and }}D_{4} )\) and \(PRE(\hat{M}_{0N} ,T_{i(d)N} ) = \frac{{Var(\hat{M}_{0N} )}}{{MSE(T_{i(d)N} )}} \times 100,\) \(({\text{for }}i = 1,...,12)\).

Table 3 presents the complete descriptions of each population mentioned below. Table 4 presents the complete descriptions of each population mentioned below. Tables 5, 6, 7, 8, 9 and 10 elaborate the PREs of all neutrosophic estimators relative to \(\hat{M}_{0N}\). It is observed that the PREs of \(T_{i(d)N}\) estimators change with the choices of \(\alpha_{3}\) and \(\alpha_{4}\). It is further noted that the performance of \(T_{i(d)N}\) is the best among all the estimators proposed here.

Table 3 Descriptive statistics of populations for single auxiliary variable.
Table 4 PRE’s of proposed neutrosophic estimators to \(\hat{M}_{0N}\).
Table 5 PRE’s of proposed neutrosophic estimators \(T^{{\begin{array}{*{20}l} { \ominus } \hfill \\ \end{array} }}_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 6 PRE’s of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 7 PRE’s of proposed neutrosophic estimators \(T^{ \otimes }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 8 PRE’s of proposed neutrosophic estimators \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledast } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 9 PRE’s of proposed neutrosophic estimators \(T^{{\begin{array}{*{20}l} {\begin{array}{*{20}l} {\begin{array}{*{20}l} { \circledcirc } \hfill \\ \end{array} } \hfill \\ \end{array} } \hfill \\ \end{array} }}_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 10 Descriptive statistics for simulation study.

As indicated by the indeterminacy interval findings from (23) for the whole data set, the neutrosophic generalized estimator \(T_{i(d)N} ,\) is more efficient than the other suggested estimators studied. Also, the indeterminacy interval findings show that the estimator \(\hat{M}_{{D_{4N} }}\) is superior to all other estimators except \(T_{i(d)N}\) for the neutrosophic population, with a moderate or low correlation between the research variable and the supplementary variable (regardless of correlation is positive or negative).

Simulation study

We evaluate the suggested estimators' efficiency using simulated neutrosophic data, such as \(Y_{N}\) and \(X_{N}\) are neutrosophic random variates (NRV). We generate two sets of neutrosophic random numbers of \(N = 1000\), which are \(x^{\prime}_{N}\) and \(y^{\prime}_{N}\) from neutrosophic bivariate gamma distribution using the R programming language. Additionally, motivated by the simulated population generation strategies used by 55, we generate the population's \(U\) transformed variables using \(\rho_{yxN} = \left[ {0.7573323, \, 0.772941} \right]\), \(\sigma_{yN}^{2} = \left[ {61.55041, \, 174.2082} \right]\), \(\sigma_{xN}^{2} = \left[ {36.45664, \, 100.2741} \right],\)

\(\mu_{yN} = \left[ {21.39525, \, 43.65861} \right],\) \(\mu_{xN} = \left[ {12.30978, \, 24.24975} \right]\) as \(Y_{N} = \mu_{yN} + \sigma_{yN} \left[ {\rho_{yxN} x^{\prime}_{N} + \left( {\sqrt {1 - \rho_{yxN}^{2} } } \right)y^{\prime}_{N} } \right]\) and \(X_{N} = \mu_{xN} + \sigma_{yN} x^{\prime}_{N}\).

where \(\rho_{yxN} \in [\rho_{yxL} ,\rho_{yxU} ]\), \(y^{\prime}_{N} \in [y^{\prime}_{L} ,y^{\prime}_{U} ]\), \(Y_{N} \in [Y_{L} ,Y_{U} ]\), \(\mu_{yN} \in [\mu_{yL} ,\mu_{yU} ]\), \(\sigma_{yN}^{2} \in [\sigma_{yL}^{2} ,\sigma_{yU}^{2} ]\), and \(x^{\prime}_{N} \in [x^{\prime}_{L} ,x^{\prime}_{U} ]\), \(X_{N} \in [X_{L} ,X_{U} ]\),\(\mu_{xN} \in [\mu_{xL} ,\mu_{xU} ]\) and \(\sigma_{xN}^{2} \in [\sigma_{xL}^{2} ,\sigma_{xU}^{2} ]\).

Table 11 summarises the findings of the simulated data set utilized to evaluate the suggested estimators' efficiency to that of traditional estimators under neutrosophic statistics. Tables 12, 13, 14, 15 and 16 contain the percent relative efficiency of neutrosophic estimators. The analysis by simulated data also verifies that \(T_{i(d)N}\), is the most efficient estimator. The simulation results suggest that the neutrosophic generalized estimator \(T_{i(d)N}\), produces more accurate and precise findings than other estimators. All estimators are unbiased (up to the first order of approximation), efficient, and reliable.

Table 11 PRE’s of proposed neutrosophic estimators to \(\hat{M}_{0N}\).
Table 12 PREs of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 13 PRE’s of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 14 PRE’s of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 15 PRE’s of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).
Table 16 PRE’s of proposed neutrosophic estimators \(T^{ \oplus }_{i(d)N}\) to \(\hat{M}_{0N}\).

In Figs. 4 and 5, we have displayed the performance of simulated data by using boxplot and Q-Q plot, respectively. The boxplots show that data is positively skewed, which implies that the median is closer to the lower or bottom quartile. The boxplot with points beyond the whiskers indicates that the data has a few outliers. Normal Q-Q plot displays that the top end of the Q-Q plot deviates from the straight line, while the lower end follows the straight line, and that the curves have a more prominent tail to the right, indicating that they are right-skewed (or positively skewed).

Figure 4
figure 4

Boxplot of simulated data for each variable.

Figure 5
figure 5

Normal Q-Q plot of simulated data for variable \(X_{N}\) and \(Y_{N}\).

Conclusion

Our study introduced neutrosophic estimators for accurately estimating the population median in datasets with uncertain, unclear values. Through precise additional variable measurements under simple random sampling, we developed improved neutrosophic estimators, evaluated them for bias and MSE, and demonstrated their superiority. Our proposed estimators offer the advantage of modelling uncertainty and vagueness inherent in many real-world scenarios, allowing for more flexible and nuanced decision-making processes. However, their disadvantage lies in the complexity of mathematical models and computational processes required, which can lead to increased computational costs and challenges in interpretation and precise quantification. We recommend these advanced estimators for future applications and highlight the ongoing need for research to enhance estimator effectiveness for various neutrosophic data types and sampling methods. Furthermore, future work will extend to multiple sampling designs, such as systematic, successive, and double sampling.