Section D
Development of Evaluation System for Defense Informatization Level
Seungbae Sim1, Sangho Lee2,*
1Center for Military Analysis and Planning, Korea Institute for Defense Analyses, Seoul, Republic of Korea,
sbsim@kida.re.kr
2Department of IT Management, Sun Moon University, Asan-si, Republic of Korea,
slee@sunmoon.ac.kr
*Corresponding Author : Sangho Lee, 70 Sunmoon-ro 221beon-gil, Asan-si, Chungcheongnam-do, 31460, Republic of Korea, +82-41-530-2515,
slee@sunmoon.ac.kr
© Copyright 2019 Korea Multimedia Society. This is an Open-Access article distributed under the terms of the
Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0/) which permits
unrestricted non-commercial use, distribution, and reproduction in any
medium, provided the original work is properly cited.
Received: Nov 03, 2019; Revised: Nov 22, 2019; Accepted: Nov 25, 2019
Published Online: Dec 31, 2019
Abstract
There is a description that you cannot manage what you do not measure. The Korea Ministry of National Defense (MND) is conducting evaluations in various fields to obtain meaningful effects from IT investments, and views that the evaluation of the defense informatization sector is divided into defense informatization policy evaluation and defense informatization project evaluation. The defense informatization level evaluation can measure the informatization level of MND and the armed forces or organizations. Since the evaluation system being studied to measure the level of defense informatization is composed mainly of qualitative metrics, it is necessary to reconstruct it based on quantitative metrics that can guarantee objectivity. In addition, for managing the level of change by evaluation objects, the evaluation system should be designed with a focus on homeostasis of metrics so that it can be measured periodically. Moreover, metrics need to be promoted in terms of performance against targets. To this end, this study proposes to measure the level of defense informatization by dividing it into defense information network, computer systems, interoperability and standardization, information security, information environment, and information system use, and suggests their metrics.
Keywords: Defense Informatization Level; Evaluation Metric; Informatization Performance; IT Effect
I. INTRODUCTION
“You can't manage what you don't measure.” This sentence is alleged to have been claimed by Edward Deming or Peter Drucker [1] and explains the importance of measurement. The Korea Ministry of National Defense (MND) is conducting evaluations in various fields to obtain meaningful effects from IT investment. The evaluation in the defense informatization sector is divided into defense informatization policy evaluation and defense informatization project evaluation [2], [3]. The defense informatization policy evaluation assesses various policies of MND and military forces related to the defense informatization, and the defense informatization project evaluation assesses IT projects such as information system (IS) development projects, IT procurement projects, and IS operation projects that are being carried out by the MND, Army, Navy, and Air Force [4]. On the other hand, the defense informatization level evaluation can measure the informatization level such as the informatization capacity of the MND and the related organizations or agencies.
Defense informatization aims to achieve military missions and create results by utilizing information technology. If the level of defense informatization associated with input or process factors is improved, the outcome factor, which is a performance or net effect, can be improved. The purpose of measuring the level of informatization can be improving the level of informatization (input or process factors) apart from the performance management of informatization (output factors).
This study proposes the evaluation system to measure the level of defense informatization in terms of defense information network, computer systems, interoperability and standardization, information security, informatization environment, and IS use, and also describes quantitative evaluation metrics in each field.
II. RELATED WORKS
2.1. Information Systems Success Model
DeLone and McLean [5] suggested an information systems (IS) success model (Fig. 1) after reviewing various existing researches related to IS performance. The model suggests that both IS quality and information quality affect both IS use and user satisfaction, that both IS use and user satisfaction affect the individual performance of IS, and that the individual performance lastly affects organizational performance. In addition, the IS use affects and is affected by the user satisfaction. In the view of the IS success model, the IS performance depends on IS quality and information quality, which is related to an input factor, IS use and user satisfaction, which is a process factor.
DeLone and McLean [6] reviewed the many empirical studies applied with their model [5], and suggested updated IS success model (Fig. 2). They included service quality and intention to use in their IS success model [5] and added a feedback from net benefits to the intention to IS use and the user satisfaction. In the view of the updated IS success model, the net benefits (performance) depends on IS quality, information quality, service quality, intention to use, use, and user satisfaction. This model focuses quality of systems, information, service, IS use, user satisfaction as an informatization level.
Figure. 2.
DeLone and McLean’s updated IS success model.
Sources: DeLone and McLean [6, p. 24], Fig. 3.
Download Original Figure
2.2. ITU ICT Development Index
International Telecommunication Union (ITU) measures the level of informatization for each country by calculating the Information and Communication Technology Development Index (IDI) [7, 8, 9]. IDI is measured in Information and Communication Technology (ICT) access, ICT use, and ICT skills. ICT access measures the fixed-telephone subscriptions per 100 inhabitants, the mobile-cellular telephone subscriptions per 100 inhabitants, the international Internet bandwidth (bit/s) per Internet user, the percentage of households with a computer, and the percentage of households with Internet access.
ICT use applies metrics such as the percentage of individuals using the Internet, the fixed-broadband subscriptions per 100 inhabitants, and the active mobile-broadband subscriptions per 100 inhabitants. ICT skills uses the metrics of the adult literacy rate, the secondary gross enrolment ratio, and the tertiary gross enrolment ratio.
The weights for ICT access, ICT use, and ICT skills are 40%, 40%, and 20%, respectively. The same weights are assigned to the metrics within each item. In 2017, ITU member countries decided to revise and expand the IDI [10], but the results have not yet been published as of October 2019. Infrastructure, use, and skills (user capability) are evaluated in the IDI.
2.3. Evaluation Method of Korean Government
The Korean government's evaluation of informatization capability (level) is being implemented as part of the evaluation of administrative management capability to improve the system in the informatization field [11]. This evaluation is managed by the Ministry of the Interior and Safety, which is based on the Framework Act on Public Service Evaluation [12]. Organizations to be evaluated are 44 central government departments including the MND.
The Ministry of the Interior and Safety conducts a preliminary evaluation of the informatization field with the National Information Society Agency (NIA), and notifies the preliminary evaluation results to each central government department. The central administrative departments finally evaluate their informatization level themselves with the results.
The evaluation system of the informatization field in 2019 shows that a perfect score is 35 points, and measures the implementation of effective e-government (20 points) and the reinforcement of cyber safety level (15 points). Evaluation metrics and their weights are shown in Table 1. The evaluation system also describes the evaluation formula as well as the evaluation method for each metric [11].
Table 1.
Metrics for national informatization level evaluation in Korea.
Type |
Evaluation metric |
Weight |
Implementation of effective e-government (20 points) |
Propulsion of performance management of e-government |
14 |
Measurement of information resource management level |
3 |
Implementation level of phased performance management |
7 |
Prevention of redundant investment |
4 |
Efficiency of website operation management |
6 |
Web compatibility/access level |
3 |
Website plug-in removal ratio |
3 |
Reinforce-ment of cyber safety level (15 points) |
Privacy protection level |
4 |
E-government civil service information security level |
4 |
Cyber security management level |
6 |
Administrative information security level |
3 |
Cyber crisis management and technical information security level |
3 |
Cyber-attack response training result |
1 |
Download Excel Table
A survey on informatization level for small and medium enterprises (SMEs) is conducted by the Ministry of SMEs and Startups and Korea Technology & Information Promotion Agency for SMEs. This survey is for establishing informatization strategies and establishing policy directions for supporting SMEs [13]. It examines the general status of companies, the willingness and plans for informatization reinforcement, the environment for informatization reinforcement, the current status of IS implementation and use, the level of informatization effectiveness, and smart factories and new ICT technologies. Since 2012, the evaluation of SMEs informatization level has not been carried out using all the surveyed items, but only three items: the willingness and plans for informatization reinforcement, the environment for informatization reinforcement, and the current status of IS implementation and use.
Many researches have been conducted on the evaluation system for defense informatization level and the metrics for each item. Lim et al. [14] suggest to measure the level of defense informatization in terms of informatization infrastructure, informatization environment, informatization use, and informatization performance. In the informatization infrastructure, defense information network, computer systems, interoperability and standardization are measured. In the informatization environment, information security, organization, education, investment, and strategy of informatization are measured. The use of informatization measures the use of battlefield management information system (MIS) and resource MIS. The informatization performance measures the level of informatization combat power and the improvement of the efficiency of defense management. Also, various evaluation metrics are presented in the study [14]. The weights of evaluation items or metrics were calculated using Analytic Hierarchy Process (AHP). It is speculated that a method of calculating the evaluation metrics may have been developed, but the study [14] did not provide it. In order to demonstrate the usability of the developed evaluation system for defense informatization level, the informatization level was measured for military units in the operation command or higher with surveys on current status and questionnaires.
For other methods of informatization level measurement, such as the corporate informatization level evaluation, and a survey on the SMEs informatization level prior to 2012, refer the existing study [14].
Since the evaluation system of defense informatization level is mainly composed of qualitative evaluation metrics, it is necessary to reconstruct it based on quantitative evaluation metrics that can guarantee objectivity [15]. It should be able to manage the increase and decrease of the level by evaluation area. It should be designed with a focus on homeostasis of metrics so that it can be measured periodically. In addition, it is necessary to refrain from evaluating all evaluation metrics using the 5-point scale without a clear comparison, and to promote the evaluation in terms of achieved performance against targets. The method of measuring the informatization level with a questionnaire is simple, but there are limitations. This may be a subjective result by the survey respondent's perception rather than an objective one. The informatization level may be measured differently by the change of the survey respondent even though the organization or institution was the same informatization level. Longitudinal analysis also needs to be available for level improvement or trend analysis.
The level evaluation can be performed in the manner of improvement of metrics compared to the previous year, the level comparison to domestic or overseas organizations (institutes), and absolute maturity level. In the short term, it is advisable to conduct a level assessment in order to compare the level of metric improvement compared to the previous year or to compare the level of metrics with other organizations. To use the evaluation system for the informatization maturity level, at least several years of data must be accumulated and available.
III. EVALUATION FOR DEFENSE INFORMATIZATION LEVEL
This study describes an evaluation system for measuring the level of defense informatization. Defense informatization level assessment can be classified into IT infrastructure, informatization environment, and IS use according to the domain of evaluation (Table 2). In particular, the level assessment for IT infrastructure is evaluated by subdividing it into areas such as Defense information network (A), Computer systems (B), Interoperability and standardization (C), and Information security(D) [15]. In the Defense information network (A), speed, traffic, and availability/recovery time are measured. Server utilization, server availability, server throughput, and work efficiency are measured in Computer systems (B). Interoperability among ISs, standardization of reference information (master data), and interoperability between weapon systems and IS are measured in Interoperability and standardization (C). The detection level (capability) of infringement incident, the response level (capability) of infringement incident, and the recovery of infringement incident are measure in Information security (D).
Table 2.
Evaluation metrics for defense informatization level.
Item |
Metric |
IT infrastructure – Defense information network(A) |
Speed(A.1) |
<A-1-1>Average speed among main nodes |
<A-1-2> Average speed between main node and secondary node (branch line) |
<A-1-3> Average speed between secondary node and user |
Traffic(A.2) |
<A-2-1> Average traffic among main nodes |
<A-2-2> Average traffic between main node and secondary node (branch line) |
<A-2-3> Average traffic between secondary node and user |
Availability/recovery time (A.3) |
<A-3-1> Network availability |
<A-3-2> Recovery time |
IT infrastructure – Computer systems(B) |
Server utilization(B.1) |
<B-1-1> Average (maximum) CPU use ratio |
<B-1-2> Average storage use ratio |
Server availability (B.2) |
<B-2-1> Server availability |
<B-2-2> Server recovery time |
Server throughput (B.3) |
<B-3-1> Average tpmC (Transaction per minute by council) |
<B-3-2> Average response time |
Work efficiency (B.4) |
<B-4-1> Input automation level (Input interface level) |
IT infrastructure – Interoperability and standardization(C) |
Interoperability among ISs(C.1) |
<C-1-1> Average interoperability level among battlefield management information systems (MISs) |
<C-1-2> Average data interconnection ratio among battlefield MISs |
<C-1-3> Average interoperability level between battlefield MIS and resource MIS |
<C-1-4> Average data interconnection level between battlefield MIS and resource MIS |
<C-1-5> Average interoperability level among resource MISs |
<C-1-6> Average data interconnection level among resource MISs |
Standardization of reference information (master data) (C.2) |
<C-2-1> Standardization ratio of reference information/code data (master data) |
Interoperability between weapon systems and IS (C.3) |
<C-3-1> Average interoperability level between weapon systems and battlefield MIS |
<C-3-2> Average data interconnection ratio between weapon systems and battlefield MIS |
IT infrastructure – Information security(D) |
Detection level (capability) of infringement incident(D.1) |
<D-1-1> Intrusion detection ratio in advance |
Response level (capability) of infringement incident(D.2) |
<D-2-1> Number of intrusion incidents per year |
<D-2-2> Intrusion incident response time |
Recovery of infringement incident(D.3) |
<D-3-1> Trace ratio of infringement incidents |
<D-3-2> Recovery time after infringement incident |
Informatization environment(E) |
Efforts to improve informatization capability(E.1) |
<E-1-1> Annual time of informatization education |
Informatization master plan(E.2) |
<E-2-1> Informatization master plan |
Efficient execution of budget(E.3) |
<E-3-1> Efficiency of budget execution |
IS use(F) |
Business informatization (F.1) |
<F-1-1> Ratio of tasks implemented by IS |
Business use(F.2) |
<F-2-1> Business use |
Download Excel Table
The level evaluation for defense information network is performed using metrics for Speed (A.1), Traffic (A.2), and Availability/recovery time (A.3). Table 3 shows the evaluation metrics for Speed (A.1). These are average speed among main nodes (main lines), average speed between the main node and the secondary node (branch line), and average speed between the secondary node and the user (user network). The equation is the ratio of measured speed of network against maximum speed (designed speed) to the target ratio. The data for network speed can be acquired from network management systems (NMS).
Table 3.
Metrics for Speed (A.1).
Item |
Description |
Evaluation item |
Defense information network(A) >> Speed(A.1) |
Metric |
<A-1-1> Average speed among main nodes |
<A-1-2> Average speed between main node and secondary node (branch line) |
<A-1-3> Average speed between secondary node and user |
Explanation |
Average speed among main nodes, the main node and the secondary node, and the secondary node and the user in the defense information network |
Measurement method |
|
(satisfaction ratio of network speed) |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with network management systems (NMS) |
Download Excel Table
Table 4 describes metrics for traffic. The evaluation metrics for Traffic (A.2), like the metrics for Speed (A.1), measure a traffic on trunk, branch, and user network. These measure if an excessive traffic against a designed or estimated traffic occurs. The data for network traffic can be acquired from NMS.
Table 4.
Metrics for Traffic (A.2).
Item |
Description |
Evaluation item |
Defense information network(A) >> Traffic(A.2) |
Metric |
<A-2-1> Average traffic among main nodes |
<A-2-2> Average traffic between main node and secondary node (branch line) |
<A-2-3> Average traffic between secondary node and user |
Explanation |
Average traffic among main nodes, the main node and the secondary node, and the secondary node and the user in the defense information network |
Measurement method |
|
(excess ratio of traffic)* |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure average, minimum, and maximum traffic of network with network management systems (NMS) |
Download Excel Table
The evaluation of Availability/recovery time (A.3) metric calculates availability with a probability that the network can be serviced without interruption, and manages data on network average recovery time to calculate availability. Table 5 presents a metric for network availability. The data for network availability can be acquired from NMS. The mean time to failure (MTTF) is average time of network service between failure of network and failure of network.
Table 5.
<A-3-1> Network availability metric.
Item |
Description |
Evaluation item |
Defense information network(A) >> Traffic(A.3) |
Metric |
<A-3-1> Network availability |
Explanation |
A probability that the defense information network can be serviced without interruption |
Measurement method |
|
(satisfaction ratio of availability) |
※ MTTF (Mean Time To Failure): average time of network service |
※ MTTR (Mean Time To Repair): average time to recovery after network service outage |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with network management systems (NMS) |
Download Excel Table
Table 6 shows the measurement method for the Recovery time metric. This calculates MTTR (Mean Time To Repair), and suggests providing service within target time of recovery. Shorter MTTR is better.
Table 6.
<A-3-2> Recovery time metric.
Item |
Description |
Evaluation item |
Defense information network(A) >> Traffic(A.3) |
Metric |
<A-3-1> Network availability |
Explanation |
A probability that the defense information network can be serviced without interruption |
Measurement method |
X = MTTR |
(excess ratio of recovery time) |
※ MTTR (Mean Time To Repair): average time to recovery after network service outage |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with network management systems (NMS) |
Download Excel Table
Other metrics are shown in appendix.
IV. CONCLUSION
This study describes the defense informatization level evaluation system modified from the defense informatization evaluation methodology [4]. The proposed evaluation system for the defense informatization level measures in terms of IT infrastructure, informatization environment, and IS use [15]. This does not focus on a survey method but the direct measurement of the level if possible. The measurement effort should be put into the evaluation. The evaluation method that was being implemented by the government is accommodated as much as possible so that the efficient evaluation can be carried out while reducing the burden on the defense organization.
As with most researches and methodologies, there are limitations in the proposed evaluation system. It is necessary to set the target value for each metric. One also needs to consider metrics such as power usage effectiveness (PUE), which are used as metrics for US government data centers [16, 17]. In addition, the study should have provided some cases or results applying the proposed evaluation system, but it does not due to a defense security issue.
Rather than waiting for the development of a sufficiently reasonable and theoretically complete evaluation system, it is better to complement the evaluation system by measuring the level of informatization. It is more significant to develop and apply the evaluation system that users can intuitively understand or use. Accumulated experiences with repetitive use of the evaluation system can lead to lessons learned and complementary needs, which can make the evaluation system more robust. Through such a virtuous cycle, the evaluation system of the defense informatization level and measurement metrics that are actively accepted by the stakeholders of the evaluation can be developed.
Acknowledgements
This manuscript is based on Research Report [15]. The authors wish to thank the editors and the anonymous reviewers for their careful reviews and constructive suggestions. Their suggestions helped strengthen the paper. All errors are the sole responsibility of the authors.
REFERENCES
A. McAfee and E. Brynjolfsson, “Big data: The management revolution,”
Harvard Business Review, pp. 60-68, October 2012.
Korea Ministry of National Defense (MND),
Act on Establishment of Infrastructure for Informatization of National Defense and Management of Informational Resources for National Defense (Abbreviation: Act on Defense Informatization), Act No. 12553,
http://elaw.klri.re.kr/kor_service/lawView.do?hseq=32670&lang=ENG, May 9, 2014.
H. J. Kwon, J. S. Choi, S. T. Kim, H. J. Lee, and Y. P. Sung, “A study for improving an evaluation systems of defense informatization,” Korea Institute for Defense Analyses, Seoul, Republic of Korea, Research Report, Feb. 2012. (
In Korean)
W. H. DeLone and E. R. McLean, “Information systems success: The quest for the dependent variable,”
Information System Research, vol. 3, no. 1, pp. 60-95, 1992.
W. H. DeLone and E. R. McLean, “The DeLone and McLean model of information systems success: A ten-year update,”
Journal of Management Information Systems, vol. 19, no. 4, pp. 9-30, Spring 2003.
International Telecommunication Union, “
Measuring the Information Society Report-Executive summary,” ITU Publications, 2010-2015.
International Telecommunication Union, “
Measuring the Information Society Report-Executive Summary,” ITU Publications, 2017.
E. J. Lee and B. M. Ahn, “
Evaluation of Informatization Level of ITU (IDI, ICT Development Index),” Korea Institute of S&T Evaluation and Planning, Seoul, Republic of Korea, KISTEP Statistic Brief, 2013-26, 2013. (
In Korean)
International Telecommunication Union, “
Measuring the Information Society Report,” Volume 1, Statistical Reports, ITU Publications, 2018.
Korea Ministry of the Interior and Safety, “
2019 Central Government Department Self-evaluation (Public Administration Capability Part) Plan (Draft),” May 17, 2019. (
In Korean)
Office for Government Policy Coordination,
Framework Act on Public Service Evaluation, Act No. 14839, July 26, 2017. (
In Korean)
G. G. Lim, D. C. Lee, H. J. Kwon, and S. R. Cho, “A case of developing performance evaluation model for Korean defense informatization,”
Information Systems Review, Vol. 19, No. 3, pp. 23-45, 2017. (
In Korean)
S. Lee, H. S. Jung, and S. J. Yoon, “An application with an evaluation methodology for defense informatization and validating the methodology,” Sun Moon University, Asan, Republic of Korea, Research Report, Nov. 2012. (
In Korean)
Appendices
APPENDIX
Table A1.
<B-1-1> Average (maximum) CPU use ratio metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server utilization(B.1) |
Metric |
<B-1-1> Average (maximum) CPU use ratio |
Explanation |
Average CPU use ratio of IS server |
Measurement method |
X = Daily / Monthly / Yearly average (maximum) CPU use ratio |
(excess ratio of CPU use) |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) |
Download Excel Table
Table A2.
<B-1-2> Average storage use ratio metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server utilization(B.1) |
Metric |
<B-1-2> Average storage use ratio |
Explanation |
Average storage use ratio of IS server |
Measurement method |
X = daily / monthly / yearly average storage use ratio |
(excess ratio of storage use) |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) |
Download Excel Table
Table A3.
<B-2-1> Server availability metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server availability(B.2) |
Metric |
<B-2-1> Server availability |
Explanation |
A probability that the server can be serviced without interruption |
Measurement method |
X = MTTF / (MTTF + MTTR) |
(satisfaction ratio of availability) |
※ MTTF (Mean Time To Failure): average time of server service |
※ MTTR (Mean Time To Repair): average time to recovery after server service outage |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) |
Download Excel Table
Table A4.
<B-2-2> Server recovery time metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server availability(B.2) |
Metric |
<B-2-2> Server recovery time |
Explanation |
Average time to recovery after server service outage |
Measurement method |
X = MTTR |
(excess ratio of recovery time) |
※ MTTR (Mean Time To Repair): average time to recovery after server service outage |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) |
Download Excel Table
Table A5.
<B-3-1> Average tpmC metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server throughput(B.3) |
Metric |
<B-3-1> Average tpmC (transaction per minute by council) |
Explanation |
Average transaction processing speed of IS server |
Measurement method |
X = average tpmC |
(satisfaction ratio of throughput speed) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) and server specification |
Download Excel Table
Table A6.
<B-3-2> Average response time metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Server throughput(B.3) |
Metric |
<B-3-2> Average response time |
Explanation |
Average response time of IS server to client requirement |
Measurement method |
X = Average response time of IS server |
(excess ratio of response time) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure with server management systems (SMS) and server specification |
Download Excel Table
Table A7.
<B-4-1> Input automation level (Input interface level) metric.
Item |
Description |
Evaluation item |
Computer systems(B) >> Work efficiency (B.4) |
Metric |
<B-4-1> Input automation level (Input interface level) |
Explanation |
Automation level of data input in IS |
Measurement method |
|
|
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A8.
<C-1-1> Average interoperability level among battlefield MISs metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-1> Average interoperability level among battlefield MISs |
Explanation |
Average Levels of Information System Interoperability (LISI) among battlefield MISs |
Measurement method |
X = Average LISI among battlefield MISs |
(satisfaction ratio of interoperability) |
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
※ Use the LISI result measured by Defense Interoperability Center |
Download Excel Table
Table A9.
<C-1-2> Average data interconnection ratio among battlefield MISs metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-2> Average data interconnection ratio among battlefield MISs |
Explanation |
Average data interconnection ratio of interface among battlefield MISs |
Measurement method |
|
(satisfaction ratio of data interconnection) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the number of the realized interconnection from ISs and the number of the required interconnection from Requirement of Capability (ROC) / Information Exchange Requirement (IER) / System/Subsystem Specification (SSS) |
Download Excel Table
Table A10.
<C-1-3> Average interoperability level between battlefield MIS and resource MIS metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-3> Average interoperability level between battlefield MIS and resource MIS |
Explanation |
Average LISI between battlefield MIS and resource MIS |
Measurement method |
X = Average LISI between battlefield MIS and resource MIS |
(satisfaction ratio of interoperability) |
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
※ Use the LISI result measured by Defense Interoperability Center |
Download Excel Table
Table A11.
<C-1-4> Average data interconnection ratio between battlefield MIS and resource MIS metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-4> Average data interconnection ratio between battlefield MIS and resource MIS |
Explanation |
Average data interconnection ratio of interface between battlefield MIS and resource MIS |
Measurement method |
|
(satisfaction ratio of data interconnection) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the number of the realized interconnection from ISs and the number of the required interconnection from Requirement of Capability (ROC) / Information Exchange Requirement (IER) / System/Subsystem Specification (SSS) |
Download Excel Table
Table A12.
<C-1-5> Average interoperability level among resource MISs metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-5> Average interoperability level among resource MISs |
Explanation |
Average LISI (Levels of Information System Interoperability) among resource MISs |
Measurement method |
X = Average LISI among resource MISs |
(satisfaction ratio of interoperability) |
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
※ Use the LISI result measured by Defense Interoperability Center |
Download Excel Table
Table A13.
<C-1-6> Average data interconnection ratio among resource MISs metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability among ISs(C.1) |
Metric |
<C-1-6> Average data interconnection ratio among resource MISs |
Explanation |
Average data interconnection ratio of interface among resource MISs |
Measurement method |
|
(satisfaction ratio of data interconnection) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the number of the realized interconnection from ISs and the number of the required interconnection from Requirement of Capability (ROC) / Information Exchange Requirement (IER) / System/Subsystem Specification (SSS) |
Download Excel Table
Table A14.
<C-2-1> Standardization ratio of reference information/code data (master data) metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Standardization of reference information (master data)(C.2) |
Metric |
<C-2-1> Standardization ratio of reference information/code data (master data) |
Explanation |
The ratio of standardized code to total code data/reference information |
Measurement method |
|
(satisfaction ratio of standardization) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A15.
<C-3-1> Average interoperability level between weapon systems and battlefield MIS metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability between weapon systems and IS(C.3) |
Metric |
<C-3-1> Average interoperability level between weapon systems and battlefield MIS |
Explanation |
Average LISI (Levels of Information System Interoperability) between weapon systems and battlefield MIS |
Measurement method |
X = Average LISI between weapon systems and battlefield MIS |
(satisfaction ratio of interoperability) |
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
※ Use the LISI result measured by Defense Interoperability Center |
Download Excel Table
Table A16.
<C-3-2> Average data interconnection ratio between weapon systems and battlefield MIS metric.
Item |
Description |
Evaluation item |
Interoperability and standardization(C) >> Interoperability between weapon systems and IS(C.3) |
Metric |
<C-3-2> Average data interconnection ratio between weapon systems and battlefield MIS |
Explanation |
Average data interconnection ratio of interface between weapon systems and battlefield MIS |
Measurement method |
|
(satisfaction ratio of data interconnection) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the number of the realized interconnection from ISs and the number of the required interconnection from ROC / IER / SSS |
Download Excel Table
Table A17.
<D-1-1> Intrusion detection ratio in advance metric.
Item |
Description |
Evaluation item |
Information security(D) >> Detection level (capability) of infringement incident(D.1) |
Metric |
<D-1-1> Intrusion detection ratio in advance |
Explanation |
The ratio of detecting the infringement incident in advance |
Measurement method |
|
(satisfaction ratio of detection in advance) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the total number of infringement incident from target IS and the number of detection in advance from the data of Computer Emergency Response Team (CERT) |
Download Excel Table
Table A18.
<D-2-1> Number of intrusion incidents per year metric.
Item |
Description |
Evaluation item |
Information security(D) >> Response level (capability) of infringement incident(D.2) |
Metric |
<D-2-1> Number of intrusion incidents per year |
Explanation |
The number of the infringement incident per year in organization |
Measurement method |
X = The number of the infringement incident per year in organization |
(excess ratio of infringement incident occurrence) |
Data gathering method (Data sources) |
■ System □ Data |
□ Questionnaires □ Interview |
※ Measure the total number of infringement incident from target IS |
Download Excel Table
Table A19.
<D-2-2> Intrusion incident response time metric.
Item |
Description |
Evaluation item |
Information security(D) >> Response level (capability) of infringement incident(D.2) |
Metric |
<D-2-2> Intrusion incident response time |
Explanation |
The response completion time following the response procedure after the occurrence of the infringement incident (the situation ending time) |
Measurement method |
X = Average processing time till the situation ending after detecting the infringement incident |
(excess ratio of processing time) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
※ Measure the occurrence of infringement incident from target IS and the processing time from related reports |
Download Excel Table
Table A20.
<D-3-1> Trace ratio of infringement incidents metric.
Item |
Description |
Evaluation item |
Information security(D) >> Recovery of infringement incident(D.3) |
Metric |
<D-3-1> Trace ratio of infringement incidents |
Explanation |
The success ratio of trace about the origin of the infringement incident |
Measurement method |
|
|
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires ■ Interview |
※ Measure the occurrence of infringement incident from target IS and the trace information from related reports |
Download Excel Table
Table A21.
<D-3-2> Recovery time after infringement incident metric.
Item |
Description |
Evaluation item |
Information security(D) >> Recovery of infringement incident(D.3) |
Metric |
<D-3-2> Recovery time after infringement incident |
Explanation |
Average time to recovery after infringement incident |
Measurement method |
X = Average time to recovery after the infringement incident |
(excess ratio of time to recovery) |
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires ■ Interview |
※ Measure the occurrence of infringement incident from target IS and the recovery time from related reports |
Download Excel Table
Table A22.
<E-1-1> Annual time of informatization education metric.
Item |
Description |
Evaluation item |
Informatization environment(E) >> Efforts to improve informatization capability(E.1) |
Metric |
<E-1-1> Annual time of informatization education |
Explanation |
Annual time of informatization education |
Measurement method |
X = Annual time of informatization education per person in organization |
(completion ratio of informatization education) |
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A23.
<E-2-1> Informatization master plan metric.
Item |
Description |
Evaluation item |
Informatization environment(E) >> Informatization master plan(E.2) |
Metric |
<E-2-1> Informatization master plan |
Explanation |
Enterprise Architecture (EA) maturity level measured by National Information Society Agency (NIA) method and the execution level of basic informatization plan |
Measurement method |
X1 = (EA maturity level) / 5 |
|
|
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A24.
<E-3-1> Efficiency of budget execution metric.
Item |
Description |
Evaluation item |
Informatization environment(E) >> Efficient execution of budget(E.3) |
Metric |
<E-3-1> Efficiency of budget execution |
Explanation |
Efficiency level of budget execution: executed budget against planned budget, the ratio of budget for new IT project to informatization budget, and the ratio of informatization budget to total defense budget |
Measurement method |
|
|
|
|
Data gathering method (Data sources) |
□ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A25.
<F-1-1> Ratio of tasks implemented by IS metric.
Item |
Description |
Evaluation item |
IS use(F) >> Business informatization(F.1) |
Metric |
<F-1-1> Ratio of tasks implemented by IS |
Explanation |
Ratio of tasks implemented by IS to total defense tasks (Informatization level of defense tasks) |
Measurement method |
|
|
Data gathering method (Data sources) |
■ System ■ Data |
□ Questionnaires □ Interview |
Download Excel Table
Table A26.
<F-2-1> Business use metric.
Item |
Description |
Evaluation item |
IS use(F) >> Business use(F.2) |
Metric |
<F-2-1> Business use |
Explanation |
The ratio of tasks with IS in the defense tasks |
Measurement method |
|
|
|
Data gathering method (Data sources) |
□ System ■ Data |
■ Questionnaires ■ Interview |
Download Excel Table