• Journal of Electronic Science and Technology
  • Vol. 22, Issue 1, 100233 (2024)
Ali Sorour* and Anthony S. Atkins
Author Affiliations
  • Faculty of Digital, Technology, Innovation, and Business, Staffordshire University, Stoke-on-Trent, ST4 2DE, United Kingdom
  • show less
    DOI: 10.1016/j.jnlest.2024.100233 Cite this Article
    Ali Sorour, Anthony S. Atkins. Big data challenge for monitoring quality in higher education institutions using business intelligence dashboards[J]. Journal of Electronic Science and Technology, 2024, 22(1): 100233 Copy Citation Text show less

    Abstract

    As big data becomes an apparent challenge to handle when building a business intelligence (BI) system, there is a motivation to handle this challenging issue in higher education institutions (HEIs). Monitoring quality in HEIs encompasses handling huge amounts of data coming from different sources. This paper reviews big data and analyses the cases from the literature regarding quality assurance (QA) in HEIs. It also outlines a framework that can address the big data challenge in HEIs to handle QA monitoring using BI dashboards and a prototype dashboard is presented in this paper. The dashboard was developed using a utilisation tool to monitor QA in HEIs to provide visual representations of big data. The prototype dashboard enables stakeholders to monitor compliance with QA standards while addressing the big data challenge associated with the substantial volume of data managed by HEIs’ QA systems. This paper also outlines how the developed system integrates big data from social media into the monitoring dashboard.

    1 Introduction

    Quality assurance (QA) in higher education (HE) is emerging as the higher education institutions (HEIs) find it difficult to monitor their levels of quality of services. Governments impose national QA standards in order to achieve the minimum level of quality of services provided by these institutions. In the Kingdom of Saudi Arabia (KSA), Education and Training Evaluation Commission (ETEC) imposes, through its National Centre for Academic Accreditation and Assessment (NCAAA), the QA standards that all HEIs in the KSA are required to comply with for accreditation [1]. To measure compliance with these standards, in 2018, NCAAA developed 23 key performance indicators (KPIs). In 2022, NCAAA introduced a revised version of KPIs. This revised version comprises a total of 17 KPIs, and all HEIs are mandated to monitor and report on compliance. This evaluation process of compliance with these KPIs should take place during the forthcoming accreditation cycle for HEIs [2].

    In addition, the increasing amount of data generated from traditional sources and through social media plays negative influence on business intelligence (BI). There is a desire to build low-cost data platforms that will be able to handle this increasing amount of data. This is sometimes known as the big data challenge [3,4].

    In this paper, different BI architectures are presented. These BI architectures encompass different BI technologies that HEIs may adopt, taking into consideration their capabilities and specific requirements of HEIs. A proposed architecture is also presented to show how the big data challenge can be addressed to deal with huge amounts of data that the QA system might require particularly if the data is being linked to social media. The recent trend of measuring students’ satisfaction is through students’ comments posted on social media.

    2 Background

    As discussed earlier, the NCAAA standards in the KSA were developed in 2018, and a revised version was released in 2022. Table 1 compares KPIs released in 2018 and 2022 and shows the related standards that are aimed to be measured. The comparison indicates that the revised KPIs include the same KPIs as that in the 2018 version but 5 KPIs have been omitted, while the revised QA standards released in 2022 remain the same (8 QA standards). The omission of these institutional KPIs was not expected as there are no KPIs specified now for measuring the standard 8 (Community Partnership). This reduction in KPIs indicates that the QA systems in HEIs in the KSA are now more mature and also allows the institutions to include additional specific KPIs if they wish to.

    NCAAA standard KPIs numbers in 2018NCAAA standard KPIs numbers in 2022NCAAA KPI descriptionRelated standards
    KPI-I-01N/APercentage of achieved indicators of the institution strategic plan objectives1
    KPI-I-02KPI-I-01Proportion of accredited programs2
    KPI-I-03KPI-I-02Students’ evaluation of quality of learning experience in the programs3
    KPI-I-04KPI-I-03First-year students’ retention rate3
    KPI-I-05KPI-I-04Graduates’ employability and enrolment in postgraduate programs3
    KPI-I-06KPI-I-05Graduation rate for undergraduate students in the specified period3
    KPI-I-07KPI-I-06Satisfaction of beneficiaries with learning resources3
    KPI-I-08KPI-I-07Employers’ evaluation of the institution graduates’ proficiency3
    KPI-I-09N/AAnnual expenditure rate per student4
    KPI-I-10KPI-I-08Students’ satisfaction with the offered services4
    KPI-I-11KPI-I-09Ratio of students to teaching staff5
    KPI-I-12KPI-I-10Proportion of faculty members with doctoral qualifications5
    KPI-I-13N/AProportion of teaching staff leaving the institution5
    KPI-I-14KPI-I-11Percentage of self-income of the institution6
    KPI-I-15KPI-I-12Satisfaction of beneficiaries with technical services6
    KPI-I-16KPI-I-13Percentage of publications of faculty members7
    KPI-I-17KPI-I-14Rate of publications per faculty member7
    KPI-I-18KPI-I-15Citation rate in refereed journals per faculty member7
    KPI-I-19KPI-I-16Number of patents, innovations, and awards of excellence7
    KPI-I-20N/AProportion of the budget dedicated to research7
    KPI-I-21KPI-I-17Proportion of external funding for research7
    KPI-I-22N/ASatisfaction of beneficiaries with community services8
    KPI-I-23N/ARate of community programs and initiatives8

    Table 1. NCAAA KPIs released in 2018 and 2022 [1].

    The NCAAA standards released in 2018 have been benchmarked against international standards [5]. In this study, the NCAAA standards released in 2022 are benchmarked against several international standards including Academic Quality Improvement Programme (AQIP) in the USA, Southern Association of Colleges and Schools (SACS) accreditation standards in the USA, Scottish Qualifications Authority (SQA) in Scotland, European Network for Quality Assurance (ENQA) in HE accreditation standards, Tertiary Education Quality and Standards Agency (TEQSA) in Australia, New Zealand Qualifications Authority (NZQA) standards, and Quality Assurance Agency (QAA) for HE standards in the UK. The comparison is presented in Table 2, which demonstrates that each of the NCAAA standards released in 2022 can be mapped to a similar international standard.

    No.NCAAA standardAQIP in the USASACS in the USASQA in ScotlandENQA in EuropeTEQSA in AustraliaNZQA in New ZealandQAA in the UK
    1Mission, vision, and strategic planning×
    2Governance, leadership, and management
    3Teaching and learning
    4Students
    5Institutional resources
    6Faculty and staff
    7Research and innovation
    8Community partnership×

    Table 2. Benchmarking NCAAA standards released in 2022 to international standards.

    A holistic framework discussed in this paper was used to develop the prototype dashboard which supports the NCAAA standards released in both 2018 and 2022. It is also noted that some institutions treat the NCAAA KPIs as a minimum requirement and may use additional institution specific KPIs. For example, King Saud University adopted a total of 56 KPIs to measure QA performance [6].

    A holistic framework for monitoring quality in HEIs using BI dashboards in the KSA (HF-HEQ-BI) was introduced previously [7]. The holistic framework and the prototype dashboard developed from the framework must therefore also be able to support additional KPIs if required by the institution. There was a need to identify the requirements to design the BI system, which assist decision-makers to monitor compliance with QA standards in HEIs. It was also necessary to identify the factors that HEIs must monitor and consider while developing the BI system for QA monitoring in the KSA. An extensive literature review identified 52 studies that discussed QA in HE using BI systems [8], indicating that there are 45 factors identified. However, none of these studies indicated the role of social media in QA monitoring using BI dashboards. A hybrid framework has been developed through the systematic literature review conducted using five main pillars (technology-organisation-environment-business-social) based on the information system strategic triangle, technology-organisation-environment (TOE) framework, and human, organisation and technology-fit to provide the theoretical underpinning background [912]. Alharbi et al. [9] and Maroufkhani et al. [10] used TOE, diffusion of innovation, and resource-based view theoretical frameworks for small and medium sized enterprises (SMEs). Alaboudi et al. [11] developed a framework to support decision making for healthcare organisations for telemedicine networks in the KSA, and Tashkandi and Al-Jabri [12] used TOE to investigate the adoption of cloud computing in HE.

    A qualitative analysis was carried out by interviewing a panel of experts to confirm the thematic analysis of factors [13]. A total of ten experts were interviewed through semi-structured convergent interviews. The responses obtained were analysed through the computer assisted qualitative data analysis software (CAQDAS), which identified 42 factors. The quantitative analysis was then conducted to validate the second version of the framework. The sample size was calculated and confirmed from literature [1417]. Snowball sampling was used to administer the online questionnaire [18,19]. IBM SPSS was used to conduct the statistical analysis. Cronbach’s alpha was calculated to determine the internal consistency of factors under each pillar. In addition, the confirmatory factor analysis (CFA) was conducted to calculate the relationship between the factors and determine if non-significant factor loadings existed [17,20,21]. T-test was conducted to determine whether significant differences in responses existed among public and private institutions [20,22]. Therefore, the 42 factors were confirmed to monitor QA in HEIs using BI dashboards with the triangulation approach to provide the HF-HEQ-BI framework which has been explained previously [8,13]. The factors outlined are shown in Table 3.

    HF-HEQ-BI framework
    TechnologyOrganisationEnvironmentBusinessSocial
    Special requirementsSafetyFitnessContinuous improvementMotivation
    Administrative servicesLocation
    Technical infrastructureLibrary servicesCareer prospectsResourcesTeam
    Curriculum structure
    Data managementFacilitiesEconomyFinancial factorsAcademic staff
    Top management support
    Data qualityInnovationPoliticsCompetitive advantageHuman elements
    Data sourcesStrategic alignmentSocio-cultureProcessReputation
    Analysis methodsLeadershipGlobalisationKPIsSocial media
    CultureCompetition
    NotificationsPartnershipAcademic qualityStakeholders’ interaction
    Administrative qualityQA regulations

    Table 3. HF-HEQ-BI factors.

    HEIs are collecting increasing amounts of data related to monitoring quality of services provided [23,24]. Social media has become a crucial part of QA monitoring in HEIs as numerous public opinions are expressed in social media channels [2527]. This huge amount of data is making increasing pressure on the BI systems used for monitoring compliance with QA standards. In the BI architecture proposed by Sorour et al. [28], the BI system was fed by different data sources including social media data. Dealing with this huge amount of data requires a BI architecture that takes into consideration the emerging technologies such as service-oriented business intelligence (SoBI) and self-service business intelligence (SSBI).

    3 Big data and analytics in higher education

    Big data refers to the use of big data for system inputs. The big data technology has been used by Li et al. [29] to perform data collection, cleaning, processing, and storage of teaching information in colleges and universities through building a quality monitoring platform.

    In HEIs, the type and volume of data gathered related to the QA requirements are increasing significantly. HEIs are struggling in managing data that feeds QA systems due to several challenges including:

    1) HEIs are increasing in their sizes (Qassim University in the KSA, for example, has expanded to 38 colleges with approximately 58267 students ranging from diplomas to Ph.D. candidates in which 92% are undergraduates) [30] and are operating in several satellite locations;

    2) As the programmes being provided by HEIs are expanding each academic year, the QA related data is also increasing;

    3) HEIs are concerned with the continuous improvement which requires a cross-year analysis that encompasses huge amounts of data.

    Besides, HEIs are needing to deal with huge amounts of data from social media comments by stakeholders. As the level of quality of services provided by HEIs can be partially determined by the public opinions expressed in social media [25,26]. Dealing with these huge amounts of data incorporates a challenge for HEIs in terms of identifying patterns in the expressed opinions in social media to determine what is useful and what is not for QA monitoring. Huge investments in the infrastructure of HEIs may be required as it might require multiple servers to handle the data [31]. As HEIs differ in their sizes and available resources, it might be challenging to satisfy this requirement to handle the big data challenge. In this paper, different BI architectures that could handle the big data challenge are discussed. Some of these architectures are designed to suit SMEs as well as large enterprises outlined in the following section. Therefore, each HEI might choose the appropriate BI architecture based on its own size and available resource.

    4 Business intelligence architectures in higher education

    This section outlines different BI architectures in HEIs to address the big data challenge. These architectures are aimed to show different configurations of BI systems in HEIs, thus helping HEIs choose the most suitable architecture according to their sizes and requirements for QA monitoring.

    4.1 Data warehouse based business intelligence architecture

    A data warehouse (DW) based BI architecture in HEIs has been proposed and outlined [28]. The proposed architecture suggests a traditional DW based BI architecture in HE, where the data source layer, the data movement, storage, and processing layer, and the data visualisation and reporting layer are included.

    4.2 Service-oriented business intelligence architecture

    Recent studies suggested that the extract, transform, and load (ETL) process in BI systems can be handled differently by using the SoBI approach [32,33]. Fig. 1 outlines the process of SoBI through web services. Web services may be used to retrieve data directly from data sources and process it in the Extensible Markup Language (XML) format. Data is then processed using the dashboard application that will store it in DW. Because the data needs further processing and then being transformed into appropriate information for the marketing analysis using data mining techniques. A web service approach is adopted rather than the traditional ETL process, because the web service approach allows retrieving the latest data from the data sources directly. This process may process data in an easier and faster manner in comparison with the traditional BI infrastructure in the circumstances where the organisation cannot afford the costs and expertise associated with developing the DW or ETL process [33]. Web services can bridge the information from different sources and overcome the difference between technologies used in data sources (e.g., some data sources may use Oracle databases while other may use MySQL). Web services also have the ability to conquer the challenge of different data sources as XML is used to encode the data.

    SoBI architecture.

    Figure 1.SoBI architecture.

    4.3 Self-service business intelligence

    According to Bentley [34], embedded analytics is designed, which makes BI more accessible for all kinds of users. It allows more real-time autonomy and self-service of data visualisation and customisation. Prem considered this kind of approach as SSBI, which allows business users to access corporate information without the involvement of the information technology (IT) team [35].

    Many tools can be used to obtain the benefits of data analytics without building complex BI structures or DW [3638]. Among these tools, Tableau, Microsoft Power BI, QlikView, and Birst [36] are available from specialised vendors. SSBI involves gathering data from data sources and inserting it directly into the tool. The tool enables the user to further process data or visualise it through BI dashboards as well as producing reports [34,36,39]. This approach is presented in Fig. 2, which identifies how organisations may benefit from commercial tools for presenting data analytics. Large firms may adopt the conventional BI architecture proposed by Sorour et al. in 2019 [28] as they have the financial capabilities for implementation, but SMEs may use the SSBI approach because it is more affordable for them [38,40]. Conventional BI implementation also requires high level of knowledge and expertise that may not be available in SMEs [40]. This is of particular importance in the KSA as private universities and some of the public universities in the KSA are considered SMEs. Technologies, such as cloud computing and, in particular, software as a service (SaaS), provide smaller organisations with access to BI systems with no complex infrastructures which are required to be owned or implemented.

    SSBI architecture.

    Figure 2.SSBI architecture.

    5 Proposed business intelligence architecture in higher education for the big data challenge

    As discussed above, different methods can be used to realize BI systems, whose requirements are determined by the needs and capabilities of each organisation. Thus, organisations that can afford the implementation costs of the complex BI system may choose a DW based approach. While for the organisations that cannot afford the implementation costs and abilities of DW or cannot handle the complex ETL process, the web service or self-service tools are beneficial to visualise data into the dashboards. Due to the fact that HEIs in the KSA differ in sizes and technical capabilities, an alternative BI architecture is proposed in this paper, as shown in Fig. 3. Thus, each organisation might select the most suitable BI solution on the basis of the organisation’s needs and requirements.

    Alternative BI architecture.

    Figure 3.Alternative BI architecture.

    As shown in Fig. 3, there are three paths in the proposed alternative BI architecture: Grey path, red path, and purple path. The grey path indicates the implementation that does not require any traditional DW, while the red path shows a traditional ETL and DW implementation for organisations that can afford the associated costs and for which this approach is appropriate. The purple path shows the SSBI approach, which is an alternative for SMEs or users who are not capable to afford complex BI architectures that include DW. The grey-dashed line in Fig. 3 indicates an optional route for HEIs to implement BI systems. The proposed BI architecture takes into consideration different data sources. These data sources can deal with huge amounts of data, especially that coming from social media.

    6 Development of the business intelligence dashboard dealing with the big data challenge

    The proposed HF-HEQ-BI architecture has been validated by Sorour et al. in 2019 [28]. An HF-HEQ-BI framework utilisation tool (FUT) has also been proposed by them in 2021 [7] to demonstrate how the HF-HEQ-BI framework can be used to develop dashboards for monitoring quality in HE in the KSA. A prototype dashboard, based on the utilisation tool, is presented in this paper to illustrate the way in which the HF-HEQ-BI framework can be applied in practice. A screen shot of part of the utilisation tool is presented in Fig. 4. This indicates how the measurement and information determination task can be mapped to the graphical presentation facilities in Microsoft Power BI to provide appropriate visualisation. HF-HEQ-BI FUT assists in identifying necessary visualisation for the dashboard. These visualisation icons enable decision-makers to monitor the progress regarding compliance with NCAAA standards, institutional specific KPIs, and social media analytics. For example, an institution may choose a ‘gauge’ visual to represent the number of publications while another may tend to represent this measurement through the line chart to view the annual trend of publications. The prototype dashboard was developed using Microsoft Power BI, and the visualisation to be presented in the dashboard screens was determined through the utilisation of HF-HEQ-BI FUT.

    HF-HEQ-BI FUT.

    Figure 4.HF-HEQ-BI FUT.

    For the purpose of developing the dashboard, the Microsoft Power BI Desktop is adopted as it is intuitive to design and free to Microsoft users. For example, Noonpakdee et al. [41] used the Microsoft Power BI Desktop to develop a dashboard template for SMEs, and Sluijter and Otten [42] used it to develop personalised student dashboards.

    7 Dashboard development phases

    The development of the BI dashboard includes identifying the requirements, producing the design solution, and evaluating the designed dashboard. In this section, the dashboard development phases are outlined in more detail. Fig. 5 illustrates how users can apply the HF-HEQ-BI framework to develop the dashboard for monitoring quality in HE. It is obvious that the SSBI tool obtains the inputs from users responsible for gathering data related to evaluating the QA performance. The developed prototype dashboards allow the top management to monitor QA activities and determine the areas that should be pay more attention to. The end users of the dashboard should only have restricted access to the data specific to their roles in the organisation to maintain data integrity. For example, if a department head is using the dashboard, only the data related to the specific department will be displayed in the dashboard.

    Dashboard development process.

    Figure 5.Dashboard development process.

    During the development process, an approach based on communities of practice (COP) was used to solicit constructive comments about the dashboard design from experts with experience and management roles in QA in the KSA. The prototype dashboard was refined through several iterations using the COP approach. COP are considered useful for socialisation as they allow researchers to obtain valuable knowledge related to their fields of expertise [43]. The dashboard was presented to a panel of experts and senior academics who served as senior managers in QA in the KSA and UK to obtain their opinions in the dashboard presentation, and the several comments suggested were used to refine the presentation.

    The initial dashboard only presented the design solution produced by Microsoft Power BI, leading to limited details presented in the dashboard. NCAAA KPIs were presented in the dashboard under the five pillars of the HF-HEQ-BI framework. Following the comments received, the dashboard was further improved to add details related to QA KPIs in order to monitor all QA factors. It was decided to represent the dashboard through a web portal with a navigation bar on the top of the page. A top pane was included to provide additional information about the project and how the HF-HEQ-BI framework can be used to develop the QA monitoring dashboard. The navigation bar allows users to navigate through the dashboard, use its functionalities (drilling-down, drilling-through, filtering, etc.), and return back to the home screen to assist in navigation.

    The use of web portal for publishing the dashboard facilitates management in assigning user permissions and roles. The top management of HEI would have full access to view all details on the dashboard. The college’s dean would normally have access to the details related to his/her college only.

    7.1 Identifying dashboard context

    The HF-HEQ-BI framework is used to determine the context of the dashboard. The developed prototype dashboard will be used to monitor quality in HEIs. Additionally, the designed dashboard using HF-HEQ-BI takes into consideration the mandatory requirements of NCAAA and provides social media analytics. The dashboard presented is illustrated with Twitter-based analytics, it can also be extended to other forms of social media analytics. The sentiment analysis of the text-based interactions (comments) and other social media analytics will appear in the dashboard to allow the top management to monitor them. Some HEIs may add further KPIs which can sometimes be many as 50 to monitor strategic plan alignment or financial metrics. For example, King Saud University adopted a total of 56 KPIs to measure QA performance [6]. Qassim University adopted 50 KPIs to monitor strategic management performance [44]. Majmaah University adopted 58 KPIs to monitor strategic performance in addition to the NCAAA mandatory KPIs [45]. These KPIs may be presented in the dashboard in the business pillar to provide improved visualisation of institutional performance, such as the successful application rate and diversity of students.

    7.2 Identifying dashboard requirements

    HF-HEQ-BI FUT has been used to determine the requirements and visualisation desired for designing the dashboard. A dashboard has been developed by using an SSBI tool. The reason for choosing an SSBI architecture is that SSBI allows more flexibility and saves time [38]. The SSBI tool also allows dashboards to be developed with no need to heavily invest in developing DW. Designing DW for this project is not practical due to the time and data constraints. As one of the most popular SSBI tools, Microsoft Power BI was selected to implement the dashboard in this study.

    7.3 Producing the design solution

    For the purpose of developing the prototype dashboard, the SSBI tools of Microsoft Power BI and Tableau and the SoBI tool of Cube.js were reviewed and tested. Tableau offers various features, including compatibility for installation on both the Macintosh Operating System (MacOS) and Windows. It can provide a trial for all of its functionalities for one year to all students studying in academic institutions. However, Tableau does not support the Arabic language in its interface. Node.js is SoBI which requires web hosting services to utilize its functionalities. Also, web programming skills are required to portray the visualisation to fit the web pages. As a comparison, the Microsoft Power BI Desktop provides user-friendly design interface, and it is freely accessible to Microsoft users.

    The visualisation requirements have been identified through HF-HEQ-BI FUT, and a dashboard has been used for the development process. Fig. 6 shows a Hypertext Preprocessor (PHP) webpage which is the homepage of the system. The top pane of the homepage provides users with information about the dashboard and the HF-HEQ-BI framework and also includes a navigation bar to navigate to different pages.

    Homepage.

    Figure 6.Homepage.

    The developed dashboard has been divided into several screens to allow decision-makers to monitor QA performance and compliance with NCAAA and QA standards by HEI. A portal was presented and used for monitoring. Hyper Text Markup Language (HTML) pages were designed to embed the dashboard pages from Power BI. Fig. 7 illustrates the first page of the dashboard, which displays the mandatory KPIs and measures the degree of compliance with the respective KPIs in HEI. Users are able to drill-down and then investigate the performance related to a specific college or department. Decision-makers in HEI will be able to monitor compliance with KPIs and identify the areas where special attention is required to achieve/maintain the national NCAAA accreditation. The dashboard is dynamic and agile, which indicates that the shown KPIs can be changed to show the 23 KPIs or the new 17 KPIs depending on the current QA system used in HEIs.

    Quality assurance dashboard (NCAAA KPIs).

    Figure 7.Quality assurance dashboard (NCAAA KPIs).

    In Fig. 8, the dashboard shows the institutional specific KPIs, which are related to the factors measuring compliance with QA standards that have not been covered by the NCAAA standards. Also, the dashboard is agile and dynamic, allowing HEIs to add their own specific KPIs (for example, King Saud University has up to 56 KPIs). As QA monitoring encompasses the activities related to the aspects such as ‘locations covered by the institution’ and ‘health and safety requirements’, decision-makers will be able to monitor compliance with such factors through this screen. This information identified throughout the development process of the HF-HEQ-BI framework is essential to build the QA system in HEIs.

    Quality assurance dashboard (institutional specific KPIs).

    Figure 8.Quality assurance dashboard (institutional specific KPIs).

    The dashboard of social media analytics is shown in Fig. 9, which includes the sentiment analysis of Tweets coming from Twitter. In certain private HEIs in the KSA, the top management consistently monitors Twitter feeds to stay abreast of any negative publicity, given that some institutions may receive several hundred Tweets daily.

    Quality assurance dashboard (social media analytics).

    Figure 9.Quality assurance dashboard (social media analytics).

    To obtain the data related to Tweets from Twitter, the Twitter application programming interface (API) was integrated into the dashboard. API allows the user to find Tweets for specific periods of time and export the sentiment analysis data into spreadsheets, thus allowing the SSBI tool to represent the data. Some SSBI tools also allow the users to connect to other third-party sentiment analysis providers such as Hootsuite and Brand24.

    HEI may be required to monitor additional KPIs to keep track of their institutional performance. These additional KPIs, such as successful application rates, students per class, and revenue streams generated by the programmes, may be monitored through the dashboard shown in Fig. 10.

    Quality assurance dashboard (additional KPIs).

    Figure 10.Quality assurance dashboard (additional KPIs).

    7.4 Design evaluation

    The prototype dashboard was developed with HF-HEQ-BI FUT. To evaluate the BI systems, both Howson [46] and Dyczkowski et al. [47] adopted a BI scorecard approach. Dashboards capabilities may be evaluated using the BI scorecards by evaluating the design, presentation, alerting, analysis, KPIs, interactivity, delivery, and architecture [41]. Further consultancy work might benefit to determine the requirements in evaluating the BI dashboard and thus improve its performance. This might include incorporating the applications of the BI dashboard in several colleges in the KSA and obtaining their feedback on the use of the dashboard.

    8 Discussion of quality monitoring dashboard features

    The dashboard presented in the paper was developed through the use of HF-HEQ-BI FUT. The dashboard presents the mandatory NCAAA KPIs as well as other KPIs related to the QA factors identified by the HF-HEQ-BI framework. The development of the dashboard went through several iterations of improvement until reaching the presentation style presented in this paper. Fig. 11 outlines the development screen from Microsoft Power BI.

    Microsoft Power BI development screen.

    Figure 11.Microsoft Power BI development screen.

    The dashboard design enables additional performance metrics for KPIs to be included and also supports the monitoring of social media streams. Fig. 12 presents the visualisation of the dashboard outputs from FUT.

    Visualisation of dashboard outputs from FUT.

    Figure 12.Visualisation of dashboard outputs from FUT.

    As illustrated in Fig. 12, the main screen of the dashboard presents the mandatory NCAAA KPIs. The user can navigate to access the supporting documentation for QA activities such as QA manuals or compliance with safety procedures audit logs. The dashboard also allows customisation of any further required visualisation and monitors as many KPIs as the institution’s top management desires. The end user of the dashboard will be allowed to monitor QA performance through the dashboard screen. The users also can drill-down to the data in the dashboard to evaluate the performance of a specific college or university department, as shown in Fig. 13.

    User visualisation of the dashboard infrastructure.

    Figure 13.User visualisation of the dashboard infrastructure.

    Fig. 14 illustrates the filter abilities of the dashboard based on keywords and time. As it can be seen in ‘top-level social media analytics’ in Fig. 14, the sentiment analysis dashboard shows the Twitter analytics for a specified period of time. The user can find the numbers of Tweets, Retweets, and Likes on the dashboard. Then the results can be filtered to show the Tweets related to a specified period of time using the slide bar filter, as shown in ‘period filter applied’ in Fig. 14. As a result, the numbers of Tweets, Retweets, and Likes are changed to reflect the new period specified. In addition, the results also can be filtered to show the Tweets containing specific keywords as shown in ‘keyword search applied’ in Fig. 14. The dashboard reflects the changes showing Tweets, Retweets, and Likes that contain these keywords. The prototype dashboard has been developed to illustrate the way in which the HF-HEQ-BI framework can be used to develop a QA dashboard. Institutions which develop such dashboard based on the framework can customise the presentation of the data to obtain more sophisticated sentiment analyses.

    Filtering results in the dashboard.

    Figure 14.Filtering results in the dashboard.

    9 Conclusion

    This paper showed how the HF-HEQ-BI framework can be applied to develop a BI prototype dashboard for monitoring quality in HEIs and a dashboard has been developed from the framework. The paper has also outlined different BI architectures that HEIs may adopt to handle the big data challenge while monitoring QA. These architectures show different configurations of BI systems that HEIs may adopt based on their requirements and available resources. The development process of the BI dashboard was discussed to show how HF-HEQ-BI FUT may be used to determine the requirements to be presented in the dashboard. HF-HEQ-BI FUT assists in determining the required visualisation for each KPI to be displayed in the dashboard. FUT is customisable and can be adapted by users to reflect the requirements of individual HEI. FUT shows how decision-makers in HEIs may select various visualisation dashboards while monitoring QA compliance regarding NCCAA standards.

    A prototype dashboard was developed to illustrate the application of the framework and also to evaluate the HF-HEQ-BI framework. The developed dashboard shows a monitoring display of the mandatory NCAAA KPIs, institutional specific KPIs, additional KPIs, and social media analytics. The social media analytics presented in the dashboard included the sentiment analysis of data drawn from Twitter through Twitter API. Additionally, the proposed dashboard allows the dashboard users to represent additional KPIs related to performance monitoring in the institutions.

    The developed dashboard based on the HF-HEQ-BI framework shows how data retrieved from social media is presented in summarised diagrams to handle the big data challenge. The prototype dashboard is dynamic and KPIs can be added or modified easily based on the changes in regulations or institutional directions of the strategic plans. Our future work will focus on the evaluation of the proposed dashboard to meet the big data challenge.

    Disclosures

    The authors declare no conflicts of interest.

    References

    [1] NCAAA, Institutional accreditation stards 2018 [Online]. Available, https:etec.gov.saarproductsservicesNCAAAAccreditationPagesAccreditationstardsdeve.aspx, May 2021.

    [2] NCAAA, Key perfmance indicats f higher education institutions, 2022 [Online]. Available, https:spapi.etec.gov.sa:2443arMediaAssetsDI101%20Institutional%20KPIS%20Ar.pdfcsf=1&e=Nmm1C6, December 2023.

    [3] Chaudhuri S., Dayal U., Narasayya V.. An overview of business intelligence technology. Commun. ACM, 54, 88-98(2011).

    [4] K.C. Laudon, J.P. Laudon, Management Infmation Systems: Managing the Digital Firm, 15th ed., Pearson, Boston, 2017.

    [5] A. Sour, A.S. Atkins, C. Stanier, F. Alharbi, R. Campion, The development of business intelligence dashboard f moniting quality in higher education institutions in Saudi Arabia including sentiment analysis from social media, in: Proc. of 16th Intl. Technology, Education Development Conf., Online, 2022, pp. 1391–1399.

    [6] King Saud University, Quality assurance system manual 2 [Online]. Available, https:dqd.ksu.edu.sasitesdqd.ksu.edu.safilesattachdlyl_nzm_dr_ljwd2true_arabic_qms_2.pdf, December 2023.

    [7] A. Sour, A.S. Atkins, C. Stanier, F. Alharbi, R. Campion, Quality moniting with business intelligence dashboards in higher educational institutions using NVIVO approach to suppt qualitative analysis, in: Proc. of 14th Annu. Intl. Conf. of Education, Research Innovation, Online, 2021, pp. 897–904.

    [8] A. Sour, A.S. Atkins, C.F. Stanier, F.D. Alharbi, Comparative framewks f moniting quality assurance in higher education institutions using business intelligence, in: Proc. of Intl. Conf. on Computing Infmation Technology, Tabuk, 2020, pp. 1–5.

    [9] Alharbi F., Atkins A., Stanier C.. Understanding the determinants of cloud computing adoption in Saudi healthcare organisations. Complex Intell. Syst., 2, 155-171(2016).

    [10] Maroufkhani P., Tseng M.L., Iranmanesh M., Ismail W.K.W., Khalid H.. Big data analytics adoption: Determinants and performances among small to medium-sized enterprises. Int. J. Inform. Manage., 54, 102190:1-15(2020).

    [11] A.A. Alaboudi, A.S. Atkins, B. Sharp, A holistic framewk f assisting decision makers of healthcare facilities to assess telemedicine applications in Saudi Arabia, in: Proc. of 7th Intl. Conf. on eHealth, Telemedicine, Social Medicine, Lisbon, 2015, pp. 203–208.

    [12] A. Tashki, I. AlJabri, Cloud computing adoption by higher education institutions in Saudi Arabia: Analysis based on TOE, in: Proc. of Intl. Conf. on Cloud Computing, Riyadh, 2015, pp. 1–8.

    [13] A. Sour, A. Atkins, C. Stanier, F. Alharbi, R. Campion, Integrated dashboards with social media analysis capabilities f moniting quality in higher education institutions, in: Proc. of 12th Intl. Conf. on Education New Learning Technologies, Online, 2020, pp. 2862–2870.

    [14] U. Sekaran, R. Bougie, Research Methods F Business: A Skill Building Approach, 7th ed., Wiley, Chichester, 2016.

    [15] T.A. Ping, C.V. Chinn, L.Y. Yin, R. Muthuveloo, The impact of infmation technology capability, business intelligence use collabation capability on ganizational perfmance among public listed companies in Malaysia, Global Bus. Manage. Res. An Intl. J. 10 (1) (2018) 293–312.

    [16] J.F. Hair, R.E. erson, R.L. Tatham, W.C. Black, Multivariate Data Analysis, 4th ed., Prentice Hall, Englewood Cliffs, 1995.

    [17] Williams B., Onsman A., Brown T.. Exploratory factor analysis: A five-step guide for novices. Australasian Journal of Paramedicine, 8, 1-13(2010).

    [18] A.S. Haris, H. Washizaki, Y. Fukazawa, Utilization of ICTs in quality assurance accreditation of higher education: Systematic literature review, in: Proc. of IEEE 6th Intl. Conf. on Teaching, Assessment, Learning f Engineering, Hong Kong, China, 2017, pp. 354–359.

    [19] Sadler G.R., Lee H.C., Lim R.S.H., Fullerton J.. Recruitment of hard-to-reach population subgroups via adaptations of the snowball sampling strategy. Nurs. Health Sci., 12, 369-374(2010).

    [20] T.A. Brown, Confirmaty Fact Analysis f Applied Research, The Guilfd Press, New Yk, 2006.

    [21] Shek D.T.L., Yu L.. Confirmatory factor analysis using AMOS: A demonstration. Int. J. Disabil. Hum. De., 13, 191-204(2014).

    [22] Kim T.K.. T test as a parametric statistic. Korean J. Anesthesiol., 68, 540-546(2015).

    [23] F. Matsebula, E. Mnkla, A big data architecture f learning analytics in higher education, in: Proc. of IEEE AFRICON, Cape Town, 2017, pp. 951–956.

    [24] S. Jha, M. Jha, L. O’Brien, A step towards big data architecture f higher education analytics, in: Proc. of 5th AsiaPacific Wld Congress on Computer Science Engineering, Nadi, 2018, pp. 178–183.

    [25] R.G. Qiu, R.R. Ravi, L.L. Qiu, Aggregating visualizing public opinions sentiment trends on the US higher education, in: Proc. of 17th Intl. Conf. on Infmation Integration Webbased Applications & Services, Brussels, 2015, p. 33.

    [26] R.G. Qiu, H. Ha, R. Ravi, L. Qiu, Y. Badr, A big data based smart evaluation system using public opinion aggregation, in: Proc. of 18th Intl. Conf. on Enterprise Infmation Systems, Setubal, 2016, pp. 520–527.

    [27] Ozturk Z.K., Cicek Z.İ.E., Ergul Z.. Sentiment analysis: An application to Anadolu University. Acta Phys. Pol. A, 132, 753-755(2017).

    [28] A. Sour, A.S. Atkins, C.F. Stanier, F.D. Alharbi, The role of business intelligence analytics in higher education quality: A proposed architecture, in: Proc. of Intl. Conf. on Advances in the Emerging Computing Technologies, Al Madinah Al Munawwarah, 2019, pp. 1–6.

    [29] Y.Q. Li, P. Li, F. Zhu, R.C. Wang, Design of higher education quality moniting evaluation platfm based on big data, in: Proc. of 12th Intl. Conf. on Computer Science Education, Houston, 2017, pp. 337–342.

    [30] Qassim University, Annual rept f academic year 1442H, 2020–2021, 2021 [Online]. Available, https:pdq.qu.edu.salaravelfilemanagerfilesshares%D8%A7%D9%84%D8%AA%D9%82%D8%B1%D9%8A%D8%B1modified%20%D8%A7%D9%84%D8%B3%D9%86%D9%88%D9%8A%20%D9%84%D9%84%D8%B9%D8%A7%D9%85%20%D8%A7%D9%84%D8%AC%D8%A7%D9%85%D8%B9%D9%8A%201442%20(%D9%84%D9%84%D9%86%D8%B4%D8%B1%20).pdf, December 2023.

    [31] Sigman B.P., Garr W., Pongsajapan R., Selvanadin M., McWilliams M., Bolling K.. Visualization of Twitter data in the classroom: Analysis of Twitter data. Decis. Sci. J. Innov. Ed., 14, 362-381(2016).

    [32] Ishaya T., Folarin M.. A service oriented approach to business intelligence in telecoms industry. Telemat. Inform., 29, 273-285(2012).

    [33] R. Somya, D. Manongga, M.A.I. Pakereng, Serviceiented business intelligence (SoBI) f academic financial data integration in university, in: Proc. of Intl. Seminar on Application f Technology of Infmation Communication, Semarang, 2018, pp. 1–5.

    [34] D. Bentley, Business Intelligence Analytics, Larsen & Keller, New Yk, 2017.

    [35] J. Prem, Your guide f starting online business with vital tips, Dr Prem Digital Marketing, 2020 [Online]. Available, https:drprem.comdigitalmarketingpracticaltipsfstartinganonlinebusiness, December 2023.

    [36] M. Paulussen, Smart SelfService Business Intelligence Framewk, M.S. thesis, Dept. of Business Economics, Maastricht Univ., Maastricht, herls, 2018.

    [37] D.V. Damme, Trends models in international quality assurance accreditation in higher education in relation to trade in education services, High. Educ. Manag. Pol. 2002, doi: 10.1787hempv14art21en.

    [38] M. Maryska, P. Doucek, Selfservice business intelligence, in: Proc. of 20th Intl. Conf. on Infmation Technology f Practice, Ostrava, 2017, pp. 259–270.

    [39] M.F. Tutunea, R.V. Rus, Business intelligence solutions f SME’s, Procedia Econ. Finance 3 (2012) 865–870.

    [40] E. Ayoubi, S. Aljawarneh, Challenges opptunities of adopting business intelligence in SMEs: Collabative model, in: Proc. of 1st Intl. Conf. on Data Science, ELearning Infmation Systems, Madrid, 2018, p. 42.

    [41] W. Noonpakdee, T. Khunknsiri, A. Phothichai, K. Danaisawat, A framewk f analyzing developing dashboard templates f small medium enterprises, in: Proc. of 5th Intl. Conf. on Industrial Engineering Applications, Singape, 2018, pp. 479–483.

    [42] J. Sluijter, M. Otten, Business intelligence (BI) f personalized student dashboards, in: Proc. of 7th Intl. Learning Analytics & Knowledge Conf., Vancouver, 2017, pp. 562–563.

    [43] Almuayqil S., Atkins A.S., Sharp B.. Application of the SECI model using web tools to support diabetes self-management and education in the Kingdom of Saudi Arabia. Intell. Inf. Manag., 9, 156-176(2017).

    [44] Qassim University, University strategy 2020–2025 [Online]. Available, https:spm.qu.edu.salaravelfilemanagerfilessharesuniversity_strategy_v1.pdf, December 2023.

    [45] Majmaah University, Key perfmance indicats manual f the second strategy plan, 2020 [Online]. Available, https:m.mu.edu.sasitesdefaultfiles2020122020%20mu_0_0.pdf, December 2023.

    [46] C. Howson, Successful Business Intelligence: Secrets to Making BI a Killer App, McGraw Hill, New Yk, 2008.

    [47] M. Dyczkowski, J. Kczak, H. Dudycz, Multicriteria evaluation of the intelligent dashboard f SME managers based on scecard framewk, in: Proc. of Federated Conf. on Computer Science Infmation Systems, Warsaw, 2014, pp. 1147–1155.

    Ali Sorour, Anthony S. Atkins. Big data challenge for monitoring quality in higher education institutions using business intelligence dashboards[J]. Journal of Electronic Science and Technology, 2024, 22(1): 100233
    Download Citation