Introduction

Data science is the set of statistical techniques and methods necessary for the extraction, analysis, and interpretation of data. In the era of “Big Data”, where a huge amount of information is available to companies, data-driven choices are essential for defining a company's medium and long-term strategies and can turn into a huge competitive advantage [14, 22]. The major internet and manufacturing companies such as Google, Facebook, and Apple hire the best data science talents to work in their vast data science departments. Being a successful company today means making data-driven decisions [12, 41]. Companies that have overlooked the potential of data science have observed their competitors seize market share and enlarge their customer base over the past years. Pioneers such as Facebook, Amazon, and Google instead developed dominant market positions. Nowadays, companies of all sizes are investing heavily in big data and AI initiatives to narrow the gap with the tech giants [8]. Although the value that data analytics brings to companies has been recognized [14, 15, 25], there is still confusion on how to properly integrate big data initiatives within the organization for long-term planning [24, 43]. This is one the main reason for the failure of more than half of big data programs worldwide [38, 43]. Being a data mature organization means being able to spot new data-driven opportunities in advance, while they are still invisible to the competitors, using analytic insights to deliver business outcomes. Survey results from data-driven companies have shown that decisions made with the help of data are more accurate, faster, and more cost-effective. According to a survey conducted by YouGov and commissioned by Tableau [21], business leaders reported that data-driven decisions lead to better decision-making, improved customer satisfaction, increased revenue, and greater cost savings. In addition, the survey found that data-driven decision-making leads to faster decision-making, with decisions taking days or weeks instead of months or years. Furthermore, the survey revealed that data-driven decisions are more accurate, with fewer errors and more accurate predictions. Finally, the survey showed that data-driven decision-making leads to greater cost savings, with data-driven companies spending less money on research and development.

In this study, we analyze the data maturity of 261 Italian companies of different sectors and sizes to answer the following business questions:

  • RQ1 What are the data maturity estimators considered most important by the managers interviewed for the success of big data initiatives?

  • RQ2 Does the relevance of data maturity estimators to industry experts differ depending on the size of the company or industry?

To respond our research questions and score what the ripeness level of the enterprises is, we relied on an eight-dimensional assessment system derived from the literature De Mauro et al. [9]—the CBDAS—consensual with the existing big data maturity models. The CBDAS applies the analytic hierarchy process to assign weights to the critical success factors for big data initiatives. We analyze how respondents (Senior Manager and IT decision-makers) agreed or disagreed with questions that underlined the importance of each success factor proposed. As a result, the paper derives insights on the importance that the managers give on data-driven choices and on the validity of the CBDAS to apply to companies of different sizes and industry sectors.

Big Data Maturity Models

The big data maturity models represent robust frameworks that support the evaluation of old and new big data initiatives among specific aspects or domains to rule whether they can generate new knowledge for a company [14, 28, 32,33,34].

These models help organizations to create a structure around their big data initiatives and provide guidance on how to structure processes, technologies, and analytics to achieve success [1, 14].

Big data maturity models are comprised of different stages that guide organizations in understanding their current level of big data initiatives and how to move forward with their big data strategy. These models follow a logical progression through defined levels, starting from a foundational stage of weakness to reaching the ultimate goal of full maturity [31]. Each stage is identified by criteria that must be met to advance to the next level. These criteria build upon each other, with each stage requiring the fulfillment of lower-level criteria and additional requirements. This structure enables business managers to determine the necessary steps to achieve further improvement [6, 16, 27].

As organizations mature in their big data journey, they gradually gain the ability to effectively utilize big data to their advantage. The process typically starts with an awareness stage where organizations first become aware of big data and its potential, then progress to a stage where they start building their big data infrastructure and capabilities. With increasing proficiency, organizations start to see success in their big data initiatives and make data-driven decisions. Finally, when organizations reach a mature level in their big data initiatives, they have the ability to harness advanced analytics and machine learning to drive growth and gain a competitive advantage [5, 13].

Overall, big data maturity models provide organizations with the structure to create strategies and achieve success in their big data initiatives. By leveraging these models, organizations can gain a better understanding of their current state and make changes to improve their big data initiatives.

By leveraging a big data maturity model, data maturity can be evaluated at the sub-domain level when it refers to micro-level factors, such as routines and organizational requirements, at the domain level when it refers to the macro-level factors to assess the needed conditions to reach maturity stages. While macrolevels generally assess strategic factors of big data initiatives’ success, microlevels make clear the actions to be taken to guide maturity within organizations [6, 16, 27].

The aspects investigated through the maturity models can be many, such as IT management, business intelligence ecosystem, and data warehouse adoption, among others. In general, big data maturity models give the company the maximum value when used to analyze how business processes and strategies integrate with big data initiatives, providing management with the needed information to support strategic and operational decisions [2].

Data models help to outline the optimal choices for a path of improvement of the business management system. The absence of specific procedures regarding the assessment and operation of maturity models may represent a limitation for the use of the model as an organizational and diagnostic-prescriptive management system.

So far, only a few of the big data maturity models present in the literature contain details on the development, validation, and evaluation processes of the model itself, constituting a limit to the validity and usefulness of many proposals [31, 33]. We rely on the Consensual Big Data Assessment System (CBDAS) proposal [9], which starts from a holistic and conceptual integration of existing models. It encompasses the key elements of success that are coherent with big data’s essential components and agrees with the most cited models in literature [2, 17]. The CBDAS offers a robust conceptual framework complemented by a practical assessment and recommendation system to grant usefulness and applicability for industries.

The CBDAS framework was leveraged as judged from the authors the most comprehensive big data maturity model after a comparison of various models was conducted. Previous research, including the systematic reviews by Comuzzi and Patel [6] and Olszak and Mach-Król [28], and dos Santos-Neto and Costa [33], revealed that only a limited number of maturity models effectively illustrate the maturity stages, which is crucial for evaluating an organization's performance. These models are varied in scope, with some focusing on specific aspects and others aiming to provide a holistic assessment. Various maturity models have been proposed for specific aspects of big data, such as IT management [4] and business intelligence [18], Mikalef et al. [26], or data warehouse adoption. However, it is widely acknowledged that the scope of big data encompasses much more than these specific areas and extends to the operational processes of organizations and the support of decision-making processes [6, 12, 30, 40]. The multidisciplinary nature of big data makes aspect-specific maturity models insufficient for a comprehensive and holistic evaluation [6, 36].

Furthermore, the lack of alignment between maturity models and assessment models can cause confusion for users and hinder their use as organizational management tools (2019).

To overcome these issues, big data maturity models need to broaden their scope to include industry surveys and gray literature, and an assessment tool is required to complement the maturity model to make it practically applicable in industries [2, 37].

Different authors aimed to make the constitutive elements of their maturity model tangible and measurable to ensure its practical applicability [4, 26, 42].

In general, big data maturity models are mostly used as a descriptive tool, assessing the current situation against a framework, while only a minor part of these models provides normative, prescriptive recommendations for performance improvement [2, 20].

The completeness of the maturity model in terms of scope is a critical factor to consider for the implementation of big data programs in real-world industries as it ensures its practicality and actionability [5, 13]. The CBDAS provides both a descriptive tool for assessing the current state against a framework and a normative, prescriptive component for performance improvement recommendations. It is aligned with the key elements of big data and is consistent with widely recognized big data maturity models. This assessment system has the potential to drive future research by providing clear understanding of the critical success factors for big data programs and the specific requirements for each subfactor to achieve organizational success, fulfilling the call by researchers for consolidation of existing maturity models rather than creating new ones [4, 26, 42].

The Consensual Big Data Assessment System (CBDAS)

The CBDAS is a consensual framework designed on a theoretical framework that outlines the key success factors (CSFs) of data analytics. This framework encompasses the factors and sub-domains, and the criteria of success of big data initiatives. The structure of the CBDAS consists of two sections: the first section assesses the current maturity levels of enterprises at various levels of granularity—from factor to subdomain level—while the second section evaluates the relative importance of the CSFs for industry experts. The assessment system is comprised of 59 questions, with 44 questions focusing on the maturity levels of the CSFs and sub-domains and 15 questions addressing the underlying business needs. The results are obtained through a standardized questionnaire with one closed-ended question per item and analyzed using software, such as Cognito, Zapier, and Google Sheets. The CBDAS is a quantitative method that generates 156 potential outcomes based on the combination of sub-domain maturity levels and their interdependencies. The second section of the questionnaire evaluates the importance of each factor using the analytic hierarchy process (AHP) which evaluates individuals’ and groups’ judgments and converts them into numerical values for objective comparison. The output of the AHP is used as input to evaluate the relevance of each organizational area from the manager's perspective, recognizing that the relevance of each factor may vary depending on context and may not be applicable to all organizations.

Methodology

Our aim was to explore the critical data maturity indicators that managers consider for the success of big data initiatives and determine if their importance changes based on the company size or industry. To achieve our scope, we gathered input from industry experts through the administration of the CBDAS framework.

We leveraged multilinear regression and Principal Component Analysis (PCA) to understand which domains are strong predictors of the final questionnaire score, we utilized. PCA helps uncover the fundamental structure of the data, while multilinear regression reveals the relationships between variables.

Then, we applied the Analytical Hierarchy Process (AHP) to assess the experts’ subjective perception of the relative importance of each domain and answer our RQ1. AHP assigns weights to factors to determine their relative significance.

We conducted the Analysis of Variance (ANOVA) tests to check for statistical differences among groups and address our RQ2. ANOVA tests the null hypothesis that the averages of different groups come from the same distribution, determining if there is a significant difference in the means of two or more groups.

To summarize, we employed PCA, multilinear regression, AHP, and ANOVA tests to analyze the data from industry experts and address our research questions. These methods were selected for their appropriateness to the study and their widespread use in the field of data analysis.

Data Collection

In this work, we analyzed the responses of 261 employees from Italian companies to a Likert-scale questionnaire (ranging from 1—strongly disagree to 5—strongly agree), where the participants were asked to answer questions that measure a company data maturity. The participants were mostly company managers and IT experts (139 managers, 122 IT experts). We used the LinkedIn platform to draw a sample of professionals worldwide to conduct the online assessment. Although the LinkedIn community is not encompassing the population of industry representatives exhaustively, it might be considered suitable for targeting professionals in scope. The process of sample selection leverage publicly available information about the respondents provided by LinkedIn users, which increases the credibility of the sample and permits control over its composition.

The inclusion criteria were related to:

  1. (a)

    the seniority of the respondents: Senior Managers and Directors;

  2. (b)

    the position covered in their organization: IT Director, IT Responsible, IT Specialist, Senior Data Specialist, Senior Data Scientist, Senior Business Analyst, IT Consultant;

  3. (c)

    their confirmed experiences and skills in the areas of Data Analytics, Big Data, and IT Management.

With the inclusion criteria defined, we identified a potential audience of over 320,000 unique respondents who were targeted through LinkedIn campaigns conducted from October to December 2021. To ensure the validity and reliability of the data collected, we applied exclusion criteria to exclude (a) incomplete assessments and (b) companies operating outside of Italy, as the focus of the study was on the Italian market. Our sample was derived from a starting pool of 297 respondents, and ultimately consisted of 261 complete responses.

Assessment Structure

The CBDAS assessment was structured in two parts: the first part is composed of 40 questions divided into 8 domains and allows the evaluation of data maturity on critical success factors for big data initiatives; the second part is made of 15 questions that focus on the pairwise comparison of data maturity characteristics of the company, which represent a multifactorial combination of the 8 critical success factors.

In the following chapter we present the results of the survey.

Questionnaire Results

The questionnaire results are summarized in Figs. 1 and 2, where the respondent proportion for each question is represented as vertical bars of distinct colors. The full questionnaire answers are available at https://github.com/SimoneMal/AHP_dataset for consultation.

Fig. 1
figure 1

Relative percentage of answers for each of the eight domains

Fig. 2
figure 2

Final Likert scores for each domain. This graph compares the sum of the Likert scores given by each of the 261 people interviewed

As shown in Figs. 1 and 2 more than half of the respondents agreed, or strongly agreed to the importance of “Integrated architecture”, “Data Friendly” and “Integrated organization” domains. This can also be seen in Fig. 2, where the sum of the Likert score is shown. This result is not surprising, since data friendliness and an integrated data architecture in a company are of paramount importance for making data driven decisions. Data integration in a company involves connecting data sources across different systems and applications to create a unified view of data. This process is essential to ensure the accuracy and reliability of data, and to enable data-driven decision-making and helps improve data quality and reduce data silos. To achieve data integration, companies can use a variety of methods, such as data federation, data replication, data virtualization, and data warehouse consolidation. Data friendliness in a company refers to the collective beliefs and behaviors of the people working in the organization regarding the use of data. To create a data friendly culture, organizations should ensure that everyone in the company is aware of the importance and value of data and data-driven decision-making. In addition, organizations should provide data literacy training and workshops, establish data governance policies and processes, and create an environment, where data are easily accessible, trusted, and used to drive decisions. An integrated organization in a company refers to a unified approach to the management of data and resources across all departments and business units. This approach ensures that all data are collected, stored, managed, and shared in a consistent and reliable manner.

Notably, most of the respondents have considered the “Data interface” domain the least important among the eight. A data interface is an interface that provides access to data by allowing users to interact with the data in some way. It may consist of web app, e standalone app with (GUI) or a command-line interface (CLI). Data interfaces allow users to view and manipulate data in a structured and organized way. The reason why the respondents have considered this domain less important than the others is maybe due to the omnipresence of such interfaces and the unquestionable high level of user friendliness achieved in the last years.

Correlation Among Parameters

We calculated a correlation coefficient matrix between every domain listed in Table 1. To do that, we used the Spearman correlation coefficient [35], which has the advantage of not being limited to continuous numerical variables but can also be applied to discrete ordinal variables. Moreover, this method can spot strictly non-linear correlations and can assess how much two variables are correlated by a monotonic function [44]. For linear relationships, the two methods give similar answers. The value of Spearman’s R is between − 1 (indicating a perfect negative correlation) and + 1 (indicating a perfect positive correlation). Weak correlations have R values \(0<\left|R\right|<0.2\), moderate correlations \(0.2<\left|R\right|<0.6\) and strong correlations \(0.6<\left|R\right|<\) 1 We have created a correlation Matrix using the software R 4.1.2 and the command rcorr. The results are shown in Fig. 3.

Table 1 Assessment submitted to the participants
Fig. 3
figure 3

Spearman correlation matrix among domains scores

It was to be expected that only positive correlations had to be found, since all questions in the Likert scale go in the same direction. The strongest correlations happen to be between data process integration and data-friendly approach, where R = 0.78 indicates a strong correlation. It also appears a strong correlation between the domain integrated architecture and data interface with a Spearman’s R = 0.72. A strong correlation among parameters is also referred to as multicollinearity.

Multicollinearity Analysis

To deepen the analysis of the correlation among parameters, we have examined the phenomenon of multicollinearity in our data. Multicollinearity refers to the high correlation between predictors in a multiple regression model, which can hinder statistical analysis and affect the robustness and accuracy of coefficient estimates. To mitigate this challenge, it is necessary to detect and address multicollinearity. This can be achieved through a dimensionality-reduction technique called Principal Component Analysis (or PCA) [19, 29].

The first step in this kind of analysis is understanding how many components are needed to explain most of the variance in the sample. To this purpose a useful tool is the scree plot (see Fig. 4), that depicts the explained sample variance as a function of the number of principal components.

Fig. 4
figure 4

Scree plot of the principal components. The elbow method suggests that the first 3 PC are enough to explain > 80% of the sample variance

A rule of thumb criteria to understand how many components are really needed to retain most of the information and reduce complexity, is the so-called elbow method. With the elbow method one looks at the start of the bend in the line (“elbow”), that in our case happens at dimension n. 3. We may, therefore, expect that in our case just three domains will be able to explain > 80% of the sample variance.

The biplots plot shows the domain variables as vectors (arrows). These vectors can be interpreted as follows [10]:

Angle—The angle that each vector makes with an axis is related to how much that variable contributes to that principal component, the smaller the angle the greater the contribution. In our case the process_data variable contributes the most to dimension 2 (and dimension 2 contributes for the 15.1% to the total variance), while the domain process_data contributes negatively to dimension 2.

Length—The longer the vector the more variability represented by that variable. In our case the variable data_strat contributes the most to the variance of dimension 1.

Directions—The angles/direction between vectors of different variables are proportional to their correlation, the closer the direction the greater the correlation of those parameters. As can be also seen in Fig. 5 the domains Data Interface (data_interface) vs Integrated Architecture (integ_arch) and Analytical Skills (analytical_skills) vs Data Friendly (data_friendly) are highly correlated.

Fig. 5
figure 5

Biplots of the three principal components. The scalar projection of the single vector on each axis represents the contribution of that variable to the principal component. For ease of visualization a colour scale proportional to the net contribution has been added

The colours in the above plot represent the total contribution of each variable to the total variance. The domains that contribute the most are in order Data Strategy (data_strat) Integrated Architecture (integ_arch) and Data Interface (data_interface). Together these variables explain > 85% of the total sample variance.

To confirm this result, we performed a multilinear regression of these three variables on the total score (Fig. 6).

Fig. 6
figure 6

Graphical representation of the multilinear regression obtained with the R "lm" command. The plotted points are 20 rather than 297, because since the questionnaire has discrete values, the points are overlapped

figure a

The coefficient of determination R2 = 0.86 confirms that alone these three variables combined can explain 86% of the sample variability.

From the PCA and multilinear regression emerges that the Data Strategy, Data Interface, and Integrated Architecture domains are good predictors of the questionnaire final score. This means that the broad range of questions posed in these domains are similar in nature to those of other domains and correlate well with other questions. It should be noted that this result does not evaluate the respondent subjective perception of the domains relative importance (evaluated through AHP) but is only a reflection of the relative correlation among domains.

Stratification by Company Sector and Size

A series of demographic questions were asked to the participants when the assessment was submitted to them. We collected information about the characteristics of the company to which they belonged. The questions were focused on the company size and the sector in which it operates. This allowed us to stratify for such parameters and search for statistically significant differences, to answer our RQ2.

The results of this stratification are shown in Figs. 7 and 8.

Fig. 7
figure 7

AHP-weighted average score by industry size. The results look very similar among different company sizes

Fig. 8
figure 8

AHP-weighted average score by industry sector. The results look similar among different company sizes

The stratification by company size shows that the only domain, where there could verify differences among the different-sized company is the data-friendly approach. However, such differences must be ascertained by means of appropriate statistical tools. We performed an ANOVA (ANalysis Of VAriance) test [11] to search for statistical differences among groups. ANOVA tests the null hypothesis that the averages of the groups belong to the same distribution by testing the variance between and within groups. We first tested for significant differences among domains, the results of which are shown in Table 2.

Table 2 ANOVA output

Since the p value is p <  < 0.05, we can reject the hypothesis that the different domains have equal means (i.e., there are significant differences). Regarding the scores of companies of different sizes, the results were opposite and are summarized in Table 3.

Table 3 ANOVA output

Since the p value >  > 0.05, we observe no statistically significant differences among different company sizes in this case. This can be interpreted as companies of varied sizes having the same data-maturity aspirations and ambitions. The stratification for the company sector also shows similar results. Figure 8 and Table 4 show no statistically significant differences in data needs and maturity scores among companies operating in different sectors.

Table 4 ANOVA output

The Analytic Hierarchy Process

Introduction to the AHP

The AHP is a structured technique for organizing and analyzing complex decisions. It was developed by Thomas L. Saaty in the 1970s, and is based on the concept of hierarchies, which are structures that can be used to represent a problem in terms of its components and sub-components. AHP uses hierarchical decomposition to break down a complex decision problem into its component parts, and then uses a combination of mathematics and psychology to evaluate them and come up with an optimal solution.

The AHP methodology consists of three main steps, namely, pairwise comparisons and local priority vectors, weighted super matrix, and super matrix formation and calculation. In the pairwise comparison step, the user compares each component to every other component and rates them according to their relative importance. In the weighted super matrix step, the relative importance scores from the pairwise comparison step are combined to generate a relative importance score for each component. Finally, the relative importance scores are used in a mathematical optimization process to generate the optimal solution to the decision problem. This solution can then be used to make an informed decision and select the best option.

Here is an example of how the AHP can be used to decide. Let us say we are trying to decide which car to buy. We may have several criteria that we want to consider when making our decision, such as price, safety, performance, and style.

First, we would break down the problem into its component parts. In this case, the component parts would be the criteria that we want to consider when making our decision.

Next, we would use the pairwise comparison method to rate each component relative to every other component. For example, we would rate the performance criterion relative to the price criterion, the safety criterion relative to the style criterion, and so on.

Finally, we would combine these ratings and generate a relative importance score for each criterion. The relative importance scores can then be used in a mathematical optimization process to generate the optimal solution to the problem.

In this case, the optimal solution would be the car that best meets all our criteria.

Mathematical Framework and Results

The AHP process is a quantitative method for making decisions based on the relative importance that people arbitrarily assign to certain factors. This process requires answering pairwise comparison questions structured in the following way:

$$\left( A \right){\text{ is }}X{\text{ times more important than }}\left( B \right)$$

Or

$$\left( A \right){\text{ is as important as }}\left( B \right)$$

Or

$$\left( B \right){\text{ is }}X{\text{ times more important than }}A$$

Weights for the single criteria are then computed according to the formula:

$$V_{1} = \sqrt[3]{{x_{11} *x_{12} *x_{13} }},$$

which is the geometric mean (criterion 1) of the question’s scores. The weights are then calculated by dividing each geometric average for the sum of all the other criteria:

$$W_{1} = \frac{{V_{1} }}{{V_{1} + V_{2} + V_{3} }}.$$

The 15 questions in the second section of the CBDAS require the respondent to choose between a pairwise comparison of data maturity characteristics and how much it counts on a specific aspect vs one other to improve big data management, according to the organization's reality.

In our specific case, the AHP process was used to evaluate the following company characteristics, derived from the conceptual CBDAS:

  1. 1.

    The proliferation of a data culture across the entire organization.

  2. 2.

    Availability of IT services.

  3. 3.

    Managers' support in data-driven projects.

  4. 4.

    Care of analytical talents within the company.

  5. 5.

    Satisfaction of technological needs.

  6. 6.

    Business sponsorship to facilitate data-driven decision-making.

The respondents were allowed to rank one of the options from equally important to three times more important, according to the respondent’s perspective on its organization reality. It is crucial to figure out that not all the domains could always be relevant to a particular context [40]. For the same reason, certain factors may be more important than others in specific sectors. To consider these possibilities we allowed the respondents to assign different weights to each organizational need in this section of the assessment. The resulting AHP super matrix is shown in Table 5.

Table 5 AHP matrix results based on the interview of 261 Italian managers and entrepreneurs

Application to the Survey

The associated weights are depicted in Fig. 9:

Fig. 9
figure 9

AHP weights for each of the dimensions chosen in this study. According to more than 261 managers interviewed, the key factors for a company’s data maturity are that the business facilitates data-driven decisions and the proliferation of data culture across the entire organization

The results showed in Fig. 10 answer our RQ1 on the aspects which are judged as most important by the respondents. Looking at the figure, one can see that once one considers the weights from the AHP, data friendliness in the organization becomes the most sought-after characteristic, followed by “Integrated Organization” and “Analytical Skills.”

Fig. 10
figure 10

Final scores weighted using the AHP process. The results now look quite different, since the domain “data-friendly” was considered the most important (higher weights) in the AHP process

Discussions and Conclusions

We created an AHP-based evaluation system for estimating companies' data maturity and the importance that their managers and IT professionals assign to data-driven choices. Our results answer RQ1 and suggest that the data-maturity estimator that is considered as the most important by the respondents was “Data friendliness”, followed by “Integrated Organization” and “Analytical Skills”. Moreover, we found that the relevance of the eight critical success factors included in CBDAS is statistically independent of the size of the company and the sector in which it is operating, making the assessment of general applicability for a broad range of business organizations, responding to our RQ2. Our findings suggest that, when companies look for new opportunities to use analytics, the presence of data-driven culture is of primary importance for making data initiatives able to generate business value [23, 39]. Data culture in a company refers to the collective beliefs and behaviors of the people working in the organization regarding the use of data. Organizations should ensure that everyone in the company is aware of the importance and value of data and data-driven decision-making. To build a successful data culture in a company, it is important to provide data literacy training and workshops, establish data governance policies and processes, and create an environment, where data are easily accessible, trusted, and used to drive decisions. In addition, organizations should incentivize data-driven decision-making and recognize individuals who demonstrate a commitment to data-driven practices. We believe that managers’ support rule should be promoting a broad sense of data ownership by all employees and a solid connection between data professionals and business functions [3, 6]. This enables data experts to directly impact business decisions and influence business strategy. By having top managers seeking advice from data analysts, organizations recognize and accept the central role of data in decision-making, business transformation, and innovation. Our research also highlighted how the characteristics identified by managers as relevant (i.e., corporate culture) do not correspond linearly to those with a higher degree of maturity. This mismatch between managers' perceptions and the implementation of concrete actions suggests the usefulness of a system of recommendations for bridging the existing maturity gap in higher priority areas.

The current study is affected by some known limitations that provide opportunities for future research. First, the limited sample size requires the assessment to be tested with a broader audience involving a larger number of enterprises respondents to confirm preliminary insights obtained from the current analysis. Second, the scope of the interviewed audience was limited to Italy, causing its findings to be prone to specific local dynamics. Third, more robust qualitative research is needed to assess the sufficiency of the critical success factors included in the assessment model that was used in this study. A future direction of the study would be to create a specific model for different company contexts capable of thoroughly evaluating how every aspect of data management change according to the complexity of the organizational network [7, 13]. This will allow increasing the practical applicability of the rule-based recommendations, obtaining specific indications to be implemented in the process of improving business choices.