<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="/resources/xsl/jats-html.xsl"?>
<article article-type="research-article" dtd-version="1.1" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
	<journal-meta>
		<journal-id journal-id-type="publisher-id">jkits</journal-id>
		<journal-title-group>
		<journal-title xml:lang="ko">한국지식정보기술학회 논문지</journal-title>
		<journal-title>Journal of Knowledge Information Technology and Systems</journal-title>
		</journal-title-group>
		<issn pub-type="ppub">1975-7700</issn>
		<publisher>
		<publisher-name xml:lang="ko">한국지식정보기술학회</publisher-name>
		<publisher-name>Korea Knowledge Information Technology Society</publisher-name>
		</publisher>
	</journal-meta>
	<article-meta>
		<article-id pub-id-type="publisher-id">jkits_2020_15_05_623</article-id>
		<article-id pub-id-type="doi">10.34163/jkits.2020.15.5.005</article-id>
		<article-categories>
			<subj-group>
				<subject>Research Article</subject>
			</subj-group>
		</article-categories>
		<title-group>
			<article-title>A Study on Big Data Platform Architecture-based Conceptual Measurement Model Using Comparative Analysis for Social Commerce</article-title>
			<trans-title-group xml:lang="ko">
				<trans-title>소셜커머스 비교 분석을 이용한 빅데이터 플랫폼 아키텍처 기반 개념 측정 모델에 관한 연구</trans-title>
			</trans-title-group>
		</title-group>
		<contrib-group>
			<contrib contrib-type="author" xlink:type="simple">
				<name-alternatives>
					<name name-style="western">
						<surname>Hwang</surname>
						<given-names>Sung-Tae</given-names>
					</name>
					<name name-style="eastern" xml:lang="ko">
						<surname>황</surname>
						<given-names>성태</given-names>
					</name>
				</name-alternatives>
				<xref ref-type="fn" rid="fn001">*</xref>
			</contrib>
		</contrib-group>
		<aff-alternatives>
			<aff><italic>Department of IT Engineering, Hansung University</italic></aff>
			<aff xml:lang="ko"><italic>한성대학교 공과대학 조교수</italic></aff>
		</aff-alternatives>
		<author-notes>
			<fn id="fn001"><label>*</label><p>Corresponding author is with the Department of IT Engineering, Hansung University, 116 Samseongyuoro-16gil Seongbuk-gu Seoul, 02876, KOREA.</p><p><italic>E-mail address</italic>: <email>sthwang@hansung.ac.kr</email></p></fn>
		</author-notes>
		<pub-date pub-type="ppub">
			<month>10</month>
			<year>2020</year>
		</pub-date>
		<volume>15</volume>
		<issue>5</issue>
		<fpage>623</fpage>
		<lpage>630</lpage>
		<history>
			<date date-type="received">
				<day>10</day>
				<month>09</month>
				<year>2020</year>
			</date>
			<date date-type="rev-recd">
				<day>21</day>
				<month>09</month>
				<year>2020</year>
			</date>
			<date date-type="accepted">
				<day>13</day>
				<month>10</month>
				<year>2020</year>
			</date>
		</history>
		<permissions>
			<copyright-statement>&#x00A9; 2020 KKITS All rights reserved</copyright-statement>
			<copyright-year>2020</copyright-year>
		</permissions>
		<abstract>
		<title>ABSTRACT</title>
		<p>With the Big Data growing in the popularity, more and more architectures are being proposed, but no unified solution is at hand. The main objective of this thesis is to specify and prepare a conceptual measurement model on which selected Big Data Platform Architecture will be mapped and compared. The conceptual measurement model is split into multiple components with the specification of responsibilities, and the their scope of the conceptual measurement model is set. All architectures mapped in a single structure are then evaluated and compared. Our research contains a brief overview of the Big Data concept and a summary of technologies used in evaluated architectures, if available. In addition, social commerce is a new paradigm infrastructure model of commercial market, which use of Web 3.0 technologies and social data to support their exchange activities. While it is popular, being a subset of commercial market, has been increasing tremendously since its introduction in 2012, there exists a general particular of research on their framework and applications’ effectiveness, especially in areas beyond the general social commerce practices. This research develops a comprehensive social commerce framework in big data environment that has four key components. Then our research applies the Big Data Platform (BDP) to analyze the usability and adaptability with the framework by applying it to social commerce successful companies. Accordingly, we provide a set of metrics for social commerce-based Big Data Platform design. The proposed our framework and they are used to lead the design and evaluation in social commerce environment. This research makes a contribution to the social commerce in proposing a new design framework and also, we have to examine in providing inside for effective approach of concerning with their social business.</p>
		</abstract>
		<trans-abstract xml:lang="ko">
		<title>요약</title>
		<p>빅데이터의 인기가 높아지면서 점점 더 많은 아키텍처가 제안되고 있지만 통합 솔루션은 없다. 본 논문의 주요 목적은 빅데이터 플랫폼 아키텍처를 매핑하고 비교할 개념적 측정 모델을 지정하고 준비하는 것이다. 개념 측정 모델은 책임 사양과 함께 여러 구성 요소로 분할되고 개념 측정 모델의 범위가 설정된다. 그런 다음 단일 구조로 매핑 된 모든 아키텍처를 평가하고 비교하고자 한다. 우리의 연구에는 빅데이터 개념에 대한 간략한 개요와 평가된 아키텍처에 사용되는 기술 요약이 포함되고 있다. 또한, 소셜커머스는 웹 3.0 기술과 소셜 데이터를 사용하여 교환 활동을 지원하는 상업 시장의 새로운 패러다임 인프라 모델이다. 상업적 시장에서의 하위 집합은 인기가 있었지만, 2012년 이래로 엄청나게 증가하였다. 특히, 일반적인 소셜커머스 관행을 넘어선 영역에서 프레임 워크와 애플리케이션의 효과에 대한 일반적인 연구가 진행 중이다. 본 연구는 4 가지 핵심 구성 요소가 있는 빅데이터 환경에서 포괄적인 소셜커머스 프레임워크를 개발하고자 한다. 그런 다음 빅데이터 플랫폼(BDP)을 적용하여 소셜커머스 성공 기업에 적용해 봄으로서 프레임 워크의 사용성과 적응성을 분석하고자 한다. 이에 따라 소셜커머스 기반 빅데이터 플랫폼 설계를 위한 일련의 지표를 제공하고자 한다. 제안된 프레임 워크는 소셜커머스 환경에서 설계 및 평가를 주도하는 데 사용된다. 본 연구는 새로운 디자인 프레임워크를 제안하는데 있어 소셜커머스에 기여하고 있으며, 그들의 소셜 비즈니스에 대한 효과적인 접근을 위해 내부 제공에 대해서도 검토하고자 한다.</p>
		</trans-abstract>
		<kwd-group kwd-group-type="author" xml:lang="en">
<title>K E Y W O R D S</title>
			<kwd>Social commerce</kwd>
			<kwd>Commercial market</kwd>
			<kwd>Web 3.0 technology</kwd>
			<kwd>Social data</kwd>
			<kwd>Big data platform</kwd>
			<kwd>Social business</kwd>
		</kwd-group>
	</article-meta>
</front>
<body>
<sec id="sec001" sec-type="intro">
	<title>1. Introduction</title>
	<p>Social Commerce is caused by the evolution of commercial market and the widespread use of social network-based data. In a various of systems and smart platform provide to commercial services with product data and customer historical record by online commerce market such as e-commerce, m-commerce or u-commerce [<xref ref-type="bibr" rid="B001">1</xref>]. Their commerce market is based on Amazon, eBay, Facebook, or Coupang and so on, these system is supported on recommendation obtained using the web community and social data. Nevertheless they were fast growth and increasing with social commerce what they want service, they did not satisfy about their commercial policy how to effectively manage to users for social service [<xref ref-type="bibr" rid="B002">2</xref>]. In other words, despite they made up design with online market place, they have a question what components need in a social commerce framework, what they want communications for personalized service, how they connect social channel with system to system, system to user or system to platform [<xref ref-type="bibr" rid="B003">3</xref>]. These are based on the integration of previous system to newly system and also, the existing knowledge of big data environment. Accordingly, our main idea propose a new framework and a set of fundamental principle for social commerce [<xref ref-type="bibr" rid="B004">4</xref>]. Moreover, we provide Big Data Platform architecture analysis for social commerce including four key elements [<xref ref-type="bibr" rid="B005">5</xref>-<xref ref-type="bibr" rid="B009">9</xref>].</p>
	<p>In order to main contributions of our work, this paper are that we propose a users and service provider using BDP Architecture.</p>
	<p>This paper is organized as follows. In Section 2 details the Big Data. Then Big Data Platform architecture is presented and discussed in Section 3. Finally, Section 4 concludes the paper by briefly summarizing the main points and proposing future work.</p>
</sec>
<sec id="sec002">
	<title>2. Big Data</title>
	<p>There is no single formulation of what Big Data actually is. Generally, it is described as the data that enhances the processing capacity of traditional systems and cannot be processed or analyzed using conventional tools or processes. More and more organizations are receiving and collecting large amounts of data, but they often do not have the means of how to store and analyze the information. This spark of information overflow comes from multiple facts. Sensors are becoming a natural part of every device, and many sensors are capturing information on a constant frequency [<xref ref-type="bibr" rid="B010">10</xref>]. It is a trend nowadays to perform most of the activities on the Internet using different services for different needs, and people tend to generate massive volumes of data on social networks. However, the amount of information is only one of the aspects of Big Data. The data, especially from sensors, usually come in a raw, semistructured, or unstructured format. It is necessary to prepare the received data before it can be stored and analyzed. The analysis became more complex as well, requiring statistical, data mining, and machine learning to manage the Big Data.</p>
	<sec id="sec002-1">
		<title>2.1 5V Model</title>
		<p>Three main characteristics of Big Data are defined and were mainly mentioned at the beginning. These characteristics are variety, velocity, and volume [<xref ref-type="bibr" rid="B011">11</xref>]. Later on, this list was extended with two more characteristics: Value and Veracity. Together, these characteristics are often referred to as 5V. The Volume indicates that the Big Data system can contain petabytes of records, and receive terabytes of data every hour. This number is expected to increase to zettabytes in the future. The Velocity refers to the speed of the information retrieval and the speed of the data flow in the system. The incoming data can be in different formats. Either structured, unstructured, semistructured, or a mix of them, the Big Data system should be able to facilitate them all, and that is the characteristic of Variability.</p>
	</sec>
	<sec id="sec002-2">
		<title>2.2 Hadoop</title>
		<p>Hadoop is one of the most popular implementation frameworks developed by the Apache as an open-source project. It is in the Apache Software Foundation, and the code-base is written in Java. Hadoop is built on top of a distributed file system and is designed to scale up naturally with the number of available hardware. The inspiration came from Google’s work on the GFS(Google File System) and the MapReduce paradigm. The main methodology is based on the function-to-data model rather than data-to-function model [<xref ref-type="bibr" rid="B012">12</xref>]. The biggest difference is in the process, where the analysis programs are sent to their data and not the other way.</p>
	</sec>
	<sec id="sec002-3">
		<title>2.3 HDFS</title>
		<p>Distributed file systems are file systems that deal with the storage on a network of machines. This functionality is necessary when the dataset cannot fit on the single physical machine and must be distributed across multiple machines. The management of data consistency and redundancy is even a bigger problem in Big Data environments [<xref ref-type="bibr" rid="B013">13</xref>]. Data in Hadoop are divided into smaller blocks and shared throughout the cluster. It is then easier to apply MapReduce functionality on smaller datasets, thus providing efficient processing in Big Data systems.</p>
	</sec>
	<sec id="sec002-4">
		<title>2.4 MapReduce</title>
		<p>MapReduce is a programming paradigm concentrating on the clustered parallel data processing part of the Big Data system. It consists of two separate phases, firstly map phase takes a set of data on the input and converts it into multiple tuples, which are key-value pairs. For example, it can extract and transform some information from the raw data on the input, if only some of the values are useful for the second phase, the reduce. The reduce phase always comes after the map phase, and it has the output of the map job as the input. The reduce combines tuples on the input into smaller sets of tuples [<xref ref-type="bibr" rid="B014">14</xref>-<xref ref-type="bibr" rid="B016">16</xref>].</p>
	</sec>
</sec>
<sec id="sec003" sec-type="methods">
	<title>3. Big Data Platform Architecture</title>
	<p>Big Data Platform architecture is a regular software architecture that concentrates on the Big Data domain. Since the Big Data is such a complex environment, it is more than necessary to specify an architecture before actually building the system. As mentioned before, the architecture should be evaluated as soon as possible, to prevent future issues. Currently, there are many different Big Data architectures specified, but no general solution is agreed as a standard. The lack of the standard in Big Data may be caused by the fact that it is a relatively new concept and there was simply no time to establish it, or by the complexity of the domain. Thus, it is complex to compare them in order to design a new architecture or to enhance existing ones. The enhancement can be done by comparing the existing architecture against some unified format and then evaluating the parts that were not found in the scope of the current architecture. During the evaluation, it has to be decided whether the part of the system that is currently not implemented can bring some additional value from the business or technical perspective. If yes, a guideline on how this component should work and cooperate with the rest of the system is needed. Moreover, technology has to be selected for the new part of the system that can be easily integrated with the existing solution.</p>
</sec>
<sec id="sec004">
	<title>4. Conceptual Measurement Model for BDP</title>
	<p>The conceptual measurement model was conceived in several steps. Firstly, the grounds of comparison had to be specified. Only the data manipulation part of the whole Big Data systems was selected as the their scope. The security aspect, quality assurance, and other non-functional requirements do not fall under the their scope. The same applies to the end-user decision making regarding the received final data and type specification of the data that enter the architecture. As to the actual model creation process, the initial requirement of all Big Data architectures was defined in the highest level as possible. Specifically, the general requirement for each architecture, which can be later extended or used only partially, is similar to the basic retrieve, transform, and consume process. The process starts with data retrieval from various sources, then the data are processed and transformed, and finally, the result of the manipulation is returned. This simple principle was then extended in every aspect within selected grounds for comparison, and a process of data modification and preservation was applied to each step. Afterward, this measurement model was validated against available architectures both from the scientific domain and architectures used in industrial applications. A few changes were applied to extend the model to cover more business and theoretical scenarios of Big Data capabilities. &#x003C;<xref ref-type="fig" rid="f001">Figure 1</xref>&#x003E; is a result of the analysis process and should cover most of the scenarios required for big data architectures in the set their scope.</p>
	<fig id="f001" orientation="portrait" position="float">
		<caption>
			<p>Figure 1. Measurement Model for BDP</p>
		</caption>
		<graphic xlink:href="../ingestImageView?artiId=ART002640778&amp;imageName=jkits_2020_15_05_623_f001.jpg" position="float" orientation="portrait" xlink:type="simple"></graphic>
	</fig>
	<fig id="f002" orientation="portrait" position="float">
		<caption>
			<p>Figure 2. Social Commerce for BDP</p>
		</caption>
		<graphic xlink:href="../ingestImageView?artiId=ART002640778&amp;imageName=jkits_2020_15_05_623_f002.jpg" position="float" orientation="portrait" xlink:type="simple"></graphic>
	</fig>
</sec>
<sec id="sec005">
	<title>5. Social Commerce for BDP</title>
	<p>Social commerce architecture was built by analyzing the existing architectures from the industry domain, namely Naver, Instagram, Twitter, and so on, Network traffic measurement. All these existing applications were then mapped to the their architecture to prove the universality of the proposed architecture. The high-level design of the resulting architecture can be seen in  &#x003C;<xref ref-type="fig" rid="f002">Figure 2</xref>&#x003E; Data sources, Data extraction, Data loading and pre-processing, Data processing, Data analysis, Data loading and transformation, Interfacing and visualization, Data Storage, and Job and model specification. Data sources can be categorized into two dimensions: mobility and structure of data. Data sources refer only to the actual data and do not describe the data retrieval, that is the purpose of the Data extraction component. These two components of the architecture correspond with the Data acquisition component of the conceptual measurement model.</p>
	<p>Consecutively, the raw data can be transformed, loaded, or compressed in the Data-loading and preprocessing component. This raw data transformation matches the Pre-processing component of the measurement model. The next component of the architecture can also transform data in terms of combining, cleaning, or replication. Furthermore, the stream processing or information extraction can be defined in the scope of Data processing component, which has the same name in the measurement model and can be mapped directly. Similarly, the Data analysis maps to the Data analysis component, where deep analytics or stream analytics take place. The following step can be an additional transformation of the results of data analysis, which takes place in the Data loading and transformation component. The counterpart of this component in the measurement model is Transformation and modeling. It also includes the model specification of the architecture, which is defined separately in</p>
	<p>the Job and model specification as a standalone process. Visualization of the analyzed and transformed data is part of the Interfacing and visualization component, which is a correlative component to the Interpretation part of the measurement model. Data storage component is spanned across all the components described above and resembles the previous architecture, where the data storage solution was very similar. The same way as in the previous example, this is represented in the conceptual measurement model in two parts as primary and secondary Data storage.</p>
</sec>
<sec id="sec006" sec-type="Conclusion">
	<title>6. Conclusions</title>
	<p>In this paper, we have proposed a description of the theoretical background on the Big Data. The main definition and characteristics of Big Data are described together with Hadoop and its specifications in terms of the main parts which are HDFS, MapReduce, and Common. contains the definition of the measurement model provided for the evaluation of Big Data architectures. The model is structured as multiple interacting components, where almost all the components can be treated as optional. Each component is described with expected functionality and the scope of responsibilities. The our scope for this conceptual measurement model was set with the emphasis on the data flow process and its orchestration, not on the data itself nor the attributes of the system. Then selected BDP architecture from both conceptual and industry origin were evaluated and mapped to the measurement model for a unified point of comparison. Each evaluated architecture contains a description of the process used in the architecture. Our BDP architecture was randomly selected with the emphasis on the relevance to the Big Data architecture domain, and we shown &#x003C;<xref ref-type="fig" rid="f003">Figure 3</xref>&#x003E;.</p>
	<fig id="f003" orientation="portrait" position="float">
		<caption>
			<p>Figure 3. The Result of Social Commerce of Each Big Data Platform</p>
		</caption>
		<graphic xlink:href="../ingestImageView?artiId=ART002640778&amp;imageName=jkits_2020_15_05_623_f003.jpg" position="float" orientation="portrait" xlink:type="simple"></graphic>
	</fig>
	<p>It was possible to map all architectures to the conceptual measurement model allowing further evaluation in terms of technologies. This measurement model is not meant as the architecture proposal, but a unified structure on which was later possible to conduct group analysis for all different architectures with different approaches to the Big Data problematics. The architectures were similar in many ways, but some of them used various and unique approaches. Also, the purpose of each architecture differed based on the domain to which it was proposed.</p>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<!--[1] P. Gaikwad, A. Mandal, P. Ruth, G. Juve, D. Kr´ol, and E. Deelman. Anomaly detection for scientific workflow applications on networked clouds. In: High Performance Computing & Simulation (HPCS), 2016 International Conference on. IEEE, pp. 645-652, 2016.-->
<ref id="B001">
<label>[1]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Gaikwad</surname><given-names>P.</given-names></name>
<name><surname>Mandal</surname><given-names>A.</given-names></name>
<name><surname>Ruth</surname><given-names>P.</given-names></name>
<name><surname>Juve</surname><given-names>G.</given-names></name>
<name><surname>Kr´ol</surname><given-names>D.</given-names></name>
<name><surname>Deelman</surname><given-names>E.</given-names></name>
</person-group>
<year>2016</year>
<article-title>Anomaly detection for scientific workflow applications on networked clouds</article-title>
<source>High Performance Computing &#x26; Simulation (HPCS), 2016 International Conference on. IEEE</source>
<fpage>645</fpage><lpage>652</lpage>
<pub-id pub-id-type="doi">10.1109/HPCSim.2016.7568396</pub-id>
</element-citation>
</ref>
<!--[2] N. Pandeeswari, and G. Kumar. Anomaly Detection System in Cloud Environment Using Fuzzy Clustering Based ANN. In: Mob. Netw. Appl. 21.3, pp. 494-505, 2016.-->
<ref id="B002">
<label>[2]</label>
<element-citation publication-type="journal">
<person-group>
<name><surname>Pandeeswari</surname><given-names>N.</given-names></name>
<name><surname>Kumar</surname><given-names>G.</given-names></name>
</person-group>
<year>2016</year>
<article-title>Anomaly Detection System in Cloud Environment Using Fuzzy Clustering Based ANN</article-title>
<source>Mob. Netw. Appl. 21.3</source>
<fpage>494</fpage><lpage>505</lpage>
<pub-id pub-id-type="doi">10.1007/s11036-015-0644-x</pub-id>
</element-citation>
</ref>
<!--[3] R. J. Hyndman, E. Wang, and N. Laptev. Large-scale unusual time series detection. In: 2015 IEEE International Conference on Data Mining Workshop (ICDMW). IEEE, pp. 1616-1619, 2015.-->
<ref id="B003">
<label>[3]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Hyndman</surname><given-names>R. J.</given-names></name>
<name><surname>Wang</surname><given-names>E.</given-names></name>
<name><surname>Laptev</surname><given-names>N.</given-names></name>
</person-group>
<year>2015</year>
<article-title>Large-scale unusual time series detection</article-title>
<conf-name>2015 IEEE International Conference on Data Mining Workshop (ICDMW). IEEE</conf-name>
<fpage>1616</fpage><lpage>1619</lpage>
<pub-id pub-id-type="doi">10.1109/ICDMW.2015.104</pub-id>
</element-citation>
</ref>
<!--[4] B. Agrawal, A. Chakravorty, C. Rong, and T. W. Wlodarczyk. R2Time: A Framework to Analyse Open TSDB Time-Series Data in HBase. In: Cloud Computing Technology and Science (CloudCom), 2014 IEEE 6th International Conference, pp. 970-975, 2014.-->
<ref id="B004">
<label>[4]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Agrawal</surname><given-names>B.</given-names></name>
<name><surname>Chakravorty</surname><given-names>A.</given-names></name>
<name><surname>Rong</surname><given-names>C.</given-names></name>
<name><surname>Wlodarczyk</surname><given-names>T. W.</given-names></name>
</person-group>
<year>2014</year>
<article-title>R2Time: A Framework to Analyse Open TSDB Time-Series Data in HBase</article-title>
<conf-name>Cloud Computing Technology and Science (CloudCom), 2014 IEEE 6th International Conference</conf-name>
<fpage>970</fpage><lpage>975</lpage>
<pub-id pub-id-type="doi">10.1109/CloudCom.2014.84</pub-id>
</element-citation>
</ref>
<!--[5] M. Solaimani, M. Iftekhar, L. Khan, B. Thur aisingham, and J.B. Ingram. Spark-based anomaly detection over multi-source VMware performance data in real-time. In: Computational Intelligence in Cyber Security (CICS), 2014 IEEE Symposium, pp. 1-8, 2014.-->
<ref id="B005">
<label>[5]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Solaimani</surname><given-names>M.</given-names></name>
<name><surname>Iftekhar</surname><given-names>M.</given-names></name>
<name><surname>Khan</surname><given-names>L.</given-names></name>
<name><surname>Thuraisingham</surname><given-names>B.</given-names></name>
<name><surname>Ingram</surname><given-names>J. B.</given-names></name>
</person-group>
<year>2014</year>
<article-title>Spark-basedanomaly detection over multi-source VMware performance data in real-time</article-title>
<conf-name>Computational Intelligence in Cyber Security (CICS), 2014 IEEE Symposium</conf-name>
<fpage>1</fpage><lpage>8</lpage>
<pub-id pub-id-type="doi">10.1109/CICYBS.2014.7013369</pub-id>
</element-citation>
</ref>
<!--[6] Y. Chen, J. Qian, and V. Saligrama. A new one-class SVM for anomaly detection. In: Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference, pp. 3567-3571, 2013.-->
<ref id="B006">
<label>[6]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Chen</surname><given-names>Y.</given-names></name>
<name><surname>Qian</surname><given-names>J.</given-names></name>
<name><surname>Saligrama</surname><given-names>V.</given-names></name>
</person-group>
<year>2013</year>
<article-title>A new one-class SVM for anomaly detection</article-title>
<conf-name>Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference</conf-name>
<fpage>3567</fpage><lpage>3571</lpage>
<pub-id pub-id-type="doi">10.1109/ICASSP.2013.6638322</pub-id>
</element-citation>
</ref>
<!--[7] L. Yu and Z. Lan. A scalable, non-parametric anomaly detection framework for hadoop. In: Proceedings of the 2013 ACM Cloud and Autonomic Computing Conference, pp. 22, 2013.-->
<ref id="B007">
<label>[7]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Yu</surname><given-names>L.</given-names></name>
<name><surname>Lan</surname><given-names>Z.</given-names></name>
</person-group>
<year>2013</year>
<article-title>A scalable, non-parametric anomaly detection framework for hadoop</article-title>
<conf-name>Proceedings of the 2013 ACM Cloud and Autonomic Computing Conference</conf-name>
<fpage>22</fpage>
<pub-id pub-id-type="doi">10.1145/2494621.2494643</pub-id>
</element-citation>
</ref>
<!--[8] M. Zaharia, M. Chowdhury, T. Das, A. Dave, J. Ma, M. McCauley, M. J. Franklin, S. Shenker, and I. Stoica. Resilient distributed datasets: A fault-tolerant abstraction for in-memory cluster computing. In: Proceedings of the 9th USENIX conference on Networked Systems Design and Implementation. USENIX Association, p. 2, 2012.-->
<ref id="B008">
<label>[8]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Zaharia</surname><given-names>M.</given-names></name>
<name><surname>Chowdhury</surname><given-names>M.</given-names></name>
<name><surname>Das</surname><given-names>T.</given-names></name>
<name><surname>Dave</surname><given-names>A.</given-names></name>
<name><surname>Ma</surname><given-names>J.</given-names></name>
<name><surname>McCauley</surname><given-names>M.</given-names></name>
<name><surname>Franklin</surname><given-names>M. J.</given-names></name>
<name><surname>Shenker</surname><given-names>S.</given-names></name>
<name><surname>Stoica</surname><given-names>I.</given-names></name>
</person-group>
<year>2012</year>
<article-title>Resilient distributed datasets: A fault-tolerant abstraction for in-memory cluster computing</article-title>
<conf-name>Proceedings of the 9th USENIX conference on Networked Systems Design and Implementation</conf-name>
<publisher-name>USENIX Association</publisher-name>
<fpage>2</fpage>
</element-citation>
</ref>
<!--[9] V. Chandrasekaran, S. Sanghavi, P. A. Parrilo, and A. S. Willsky. Rank-sparsity incoherence for matrix decomposition. In: SIAM Journal on Optimization 21.2, pp. 572-596, 2011.-->
<ref id="B009">
<label>[9]</label>
<element-citation publication-type="journal">
<person-group>
<name><surname>Chandrasekaran</surname><given-names>V.</given-names></name>
<name><surname>Sanghavi</surname><given-names>S.</given-names></name>
<name><surname>Parrilo</surname><given-names>P. A.</given-names></name>
<name><surname>Willsky</surname><given-names>A. S.</given-names></name>
</person-group>
<year>2011</year>
<article-title>Rank-sparsity incoherence for matrix decomposition</article-title>
<source>SIAM Journal on Optimization</source>
<volume>21</volume><issue>2</issue>
<fpage>572</fpage><lpage>596</lpage>
<pub-id pub-id-type="doi">10.1137/090761793</pub-id>
</element-citation>
</ref>
<!--[10] C. Wang, K. Viswanathan, L. Choudur, V. Talwar, W. Satterfield, and K. Schwan. Statistical techniques for online anomaly detection in data centers. In: Integrated Network Management (IM), 2011 IFIP/IEEE International Symposium on. IEEE, pp. 385-392, 2011.-->
<ref id="B010">
<label>[10]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Wang</surname><given-names>C.</given-names></name>
<name><surname>Viswanathan</surname><given-names>K.</given-names></name>
<name><surname>Choudur</surname><given-names>L.</given-names></name>
<name><surname>Talwar</surname><given-names>V.</given-names></name>
<name><surname>Satterfield</surname><given-names>W.</given-names></name>
<name><surname>Schwan</surname><given-names>K.</given-names></name>
</person-group>
<year>2011</year>
<article-title>Statistical techniques for online anomaly detection in data centers</article-title>
<conf-name>Integrated Network Management (IM), 2011 IFIP/IEEE International Symposium on. IEEE</conf-name>
<fpage>385</fpage><lpage>392</lpage>
<pub-id pub-id-type="doi">10.1109/INM.2011.5990537</pub-id>
</element-citation>
</ref>
<!--[11] J. S. Kim, J. H. Kim, A Study on Adaptive Smart Platform for Intelligent Software in Big Data Environment, Journal of Knowledge Information Technology and Systems(JKITS), Vol. 15, No. 3, pp. 347-355, Jun. 2020.-->
<ref id="B011">
<label>[11]</label>
<element-citation publication-type="journal">
<person-group>
<name><surname>Kim</surname><given-names>J. S.</given-names></name>
<name><surname>Kim</surname><given-names>J. H.</given-names></name>
</person-group>
<year>2020</year>
<month>Jun.</month>
<article-title>A Study on Adaptive Smart Platform for Intelligent Software in Big Data Environment</article-title>
<source>Journal of Knowledge Information Technology and Systems(JKITS)</source>
<volume>15</volume><issue>3</issue>
<fpage>347</fpage><lpage>355</lpage>
<pub-id pub-id-type="doi">10.34163/jkits.2020.15.3.004</pub-id>
</element-citation>
</ref>
<!--[12] C. Wang, V. Talwar, K. Schwan, and P. Ranganathan. Online detection of utility cloud anomalies using metric distributions. In: Network Operations and Management Symposium (NOMS), 2010 IEEE, pp. 96-103, 2010.-->
<ref id="B012">
<label>[12]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Wang</surname><given-names>C.</given-names></name>
<name><surname>Talwar</surname><given-names>V.</given-names></name>
<name><surname>Schwan</surname><given-names>K.</given-names></name>
<name><surname>Ranganathan</surname><given-names>P.</given-names></name>
</person-group>
<year>2010</year>
<article-title>Online detection of utility cloud anomalies using metric distributions</article-title>
<conf-name>Network Operations and Management Symposium (NOMS), 2010 IEEE</conf-name>
<fpage>96</fpage><lpage>103</lpage>
<pub-id pub-id-type="doi">10.1109/NOMS.2010.5488443</pub-id>
</element-citation>
</ref>
<!--[13] J. Wright, A. Ganesh, S. Rao, Y. Peng, and Y. Ma. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. In: Advances in neural information processing systems, pp. 2080-2088, 2009.-->
<ref id="B013">
<label>[13]</label>
<element-citation publication-type="other">
<person-group>
<name><surname>Wright</surname><given-names>J.</given-names></name>
<name><surname>Ganesh</surname><given-names>A.</given-names></name>
<name><surname>Rao</surname><given-names>S.</given-names></name>
<name><surname>Peng</surname><given-names>Y.</given-names></name>
<name><surname>Ma</surname><given-names>Y.</given-names></name>
</person-group>
<year>2009</year>
<source>Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization</source>
<comment>Advances in neural information processing systems</comment>
<fpage>2080</fpage><lpage>2088</lpage>
</element-citation>
</ref>
<!--[14] K. H. Ramah, H. Ayari, and F. Kamoun. Traffic anomaly detection and characterization in the tunisian national university network. In: NETWORKING 2006. Networking Technologies, Services, and Protocols; Performance of Computer and Communication Networks; Mobile and Wireless Communications Systems. Springer, pp. 136-147, 2006-->
<ref id="B014">
<label>[14]</label>
<element-citation publication-type="book">
<person-group>
<name><surname>Ramah</surname><given-names>K. H.</given-names></name>
<name><surname>Ayari</surname><given-names>H.</given-names></name>
<name><surname>Kamoun</surname><given-names>F.</given-names></name>
</person-group>
<year>2006</year>
<chapter-title>Traffic anomaly detection and characterization in the tunisian national university network</chapter-title>
<source>NETWORKING 2006. Networking Technologies, Services, and Protocols; Performance of Computer and Communication Networks; Mobile and Wireless Communications Systems</source>
<publisher-name>Springer</publisher-name>
<fpage>136</fpage><lpage>147</lpage>
</element-citation>
</ref>
<!--[15] M. Jahrer, A. Toscher, J. Y. Lee, J Deng, H. Zhang, and J. Spoelstra. Ensemble of collaborative filtering and feature engineered models for click through rate prediction. In: KDDCup Workshop. 2012.-->
<ref id="B015">
<label>[15]</label>
<element-citation publication-type="paper">
<person-group>
<name><surname>Jahrer</surname><given-names>M.</given-names></name>
<name><surname>Toscher</surname><given-names>A.</given-names></name>
<name><surname>Lee</surname><given-names>J. Y.</given-names></name>
<name><surname>Deng</surname><given-names>J.</given-names></name>
<name><surname>Zhang</surname><given-names>H.</given-names></name>
<name><surname>Spoelstra</surname><given-names>J.</given-names></name>
</person-group>
<year>2012</year>
<article-title>Ensemble of collaborative filtering and feature engineered models for click through rate prediction</article-title>
<conf-name>KDDCup Workshop</conf-name>
</element-citation>
</ref>
<!--[16] D. E. Bloom, D. Canning, and A. Lubet. Global population aging: Facts, challenges, solutions & perspectives. In: Daedalus, 144.2, pp. 80-92, 2015.-->
<ref id="B016">
<label>[16]</label>
<element-citation publication-type="journal">
<person-group>
<name><surname>Bloom</surname><given-names>D. E.</given-names></name>
<name><surname>Canning</surname><given-names>D.</given-names></name>
<name><surname>Lubet</surname><given-names>A.</given-names></name>
</person-group>
<year>2015</year>
<article-title>Global population aging: Facts, challenges, solutions &#x26; perspectives</article-title>
<source>Daedalus</source>
<volume>144</volume><issue>2</issue>
<fpage>80</fpage><lpage>92</lpage>
<pub-id pub-id-type="doi">10.1162/DAED_a_00332</pub-id>
</element-citation>
</ref>
</ref-list>
<ack>
<title>Acknowledgments</title>
<p>This research was financially supported by Hansung University.</p>
</ack>
<bio>
	<p><graphic xlink:href="../ingestImageView?artiId=ART002640778&amp;imageName=jkits_2020_15_05_623_f004.jpg"></graphic><bold>Sung-Tae Hwang</bold> he is Assistant Professor of Department of IT Engineering at the Hansung University, Seoul, and Korea. He respectively received his Ph.D. degrees in Physics from Sogang University, Korea, in 2010. He is currently a senior researcher at Sogang University's Institute of Basic Science and is studying big data and quantum optics. His research interests include Quantum optics, Quantum information, Smart Platform, Machine Learning, Big Data, Intelligent Software, and Artificial Intelligent. He is a life member of the KKITS.</p>
	<p><italic>E-mail address</italic>: <email>sthwang@hansung.ac.kr</email></p>
</bio>
</back>
</article>
