All publications of JIPS are indexed in SCOPUS, DOI, DBLP, EBSCO, Google and Google scholar.
JIPS is also selected as the Journal for Accreditation by NRF (National Research Foundation of Korea).
The Journal of Information Processing Systems (ISSN: 1976-913X(Print), ISSN: 2092-805X(Online))
is the official international journal of the Korea Information Processing Society. As information processing systems are progressing at a rapid pace,
the Korea Information Processing Society is committed to providing researchers and other professionals with the academic information and resources
they need to keep abreast with ongoing developments. The JIPS aims to be a premier source that enables researchers and professionals all over the world to promote,
share, and discuss all major research issues and developments in the field of information processing systems and other related fields.
The inactive student rate is becoming a major problem in most open universities worldwide. In Indonesia, roughly 36% of students were found to be inactive, in 2005. Data mining had been successfully employed to solve problems in many domains, such as for educational purposes. We are proposing a method for preventing inactive students by mining knowledge from student record systems with several state of the art ensemble methods, such as Bagging, AdaBoost, Random Subspace, Random Forest, and Rotation Forest. The most influential attributes, as well as demographic attributes (marital status and employment), were successfully obtained which were affecting student of being inactive. The complexity and accuracy of classification techniques were also compared and the experimental results show that Rotation Forest, with decision tree as the base-classifier, denotes the best performance compared to other classifiers.
At present, it is simple to the electronic commerce credit scoring model, as a brush credit phenomenon in E- commerce has emerged. This phenomenon affects the judgment of consumers and hinders the rapid development of E-commerce. In this paper, that E-commerce credit evaluation model that uses a Gaussian density function is put forward by density test and the analysis for the anomalies of E-commerce credit rating, it can be fond out the abnormal point in credit scoring, these points were calculated by nonlinear credit scoring algorithm, thus it can effectively improve the current E-commerce credit score, and enhance the accuracy of E-commerce credit score.
Fuzzy Formal Concept Analysis (FCA) is a mathematical tool for the effective representation of imprecise and vague knowledge. However, with a large number of formal concepts from a fuzzy context, the task of knowledge representation becomes complex. Hence, knowledge reduction is an important issue in FCA with a fuzzy setting. The purpose of this current study is to address this issue by proposing a method that computes the corresponding crisp order for the fuzzy relation in a given fuzzy formal context. The obtained formal context using the proposed method provides a fewer number of concepts when compared to original fuzzy context. The resultant lattice structure is a reduced form of its corresponding fuzzy concept lattice and preserves the specialized and generalized concepts, as well as stability. This study also shows a step-by-step demonstration of the proposed method and its application.
In this paper, we present a virtual laboratory platform (VLP) baptized Mercury allowing students to make practical work (PW) on different aspects of mobile wireless sensor networks (WSNs). Our choice of WSNs is motivated mainly by the use of real experiments needed in most courses about WSNs. These experiments require an expensive investment and a lot of nodes in the classroom. To illustrate our study, we propose a course related to energy efficient and safe weighted clustering algorithm. This algorithm which is coupled with suitable routing protocols, aims to maintain stable clustering structure, to prevent most routing attacks on sensor networks, to guaranty energy saving in order to extend the lifespan of the network. It also offers a better performance in terms of the number of re-affiliations. The platform presented here aims at showing the feasibility, the flexibility and the reduced cost of such a realization. We demonstrate the performance of the proposed algorithms that contribute to the familiarization of the learners in the field of WSNs.
Web shells are programs that are written for a specific purpose in Web scripting languages, such as PHP, ASP, ASP.NET, JSP, PERL-CGI, etc. Web shells provide a means to communicate with the server’s operating system via the interpreter of the web scripting languages. Hence, web shells can execute OS specific commands over HTTP. Usually, web attacks by malicious users are made by uploading one of these web shells to compromise the target web servers. Though there have been several approaches to detect such malicious web shells, no standard dataset has been built to compare various web shell detection techniques. In this paper, we present a collection of web shell files, WebSHArk 1.0, as a standard dataset for current and future studies in malicious web shell detection. To provide baseline results for future studies and for the improvement of current tools, we also present some benchmark results by scanning the WebSHArk dataset directory with three web shell scanning tools that are publicly available on the Internet. The WebSHArk 1.0 dataset is only available upon request via email to one of the authors, due to security and legal issues.
Cryptography aims at transmitting secure data over an unsecure network in coded version so that only the intended recipient can analyze it. Communication through messages, emails, or various other modes requires high security so as to maintain the confidentiality of the content. This paper deals with IDEA’s shortcoming of generating weak keys. If these keys are used for encryption and decryption may result in the easy prediction of ciphertext corresponding to the plaintext. For applying genetic approach, which is well-known optimization technique, to the weak keys, we obtained a definite solution to convert the weaker keys to stronger ones. The chances of generating a weak key in IDEA are very rare, but if it is produced, it could lead to a huge risk of attacks being made on the key, as well as on the information. Hence, measures have been taken to safeguard the key and to ensure the privacy of information.
In this paper, we propose a maximum entropy-based model, which can mathematically explain the bio- molecular event extraction problem. The proposed model generates an event table, which can represent the relationship between an event trigger and its arguments. The complex sentences with distinctive event structures can be also represented by the event table. Previous approaches intuitively designed a pipeline system, which sequentially performs trigger detection and arguments recognition, and thus, did not clearly explain the relationship between identified triggers and arguments. On the other hand, the proposed model generates an event table that can represent triggers, their arguments, and their relationships. The desired events can be easily extracted from the event table. Experimental results show that the proposed model can cover 91.36% of events in the training dataset and that it can achieve a 50.44% recall in the test dataset by using the event table.
In this paper, we propose a novel block cipher mode of operation, which is known as the counter chain (CC) mode. The proposed CC mode integrates the cipher block chaining (CBC) block cipher mode of operation with the counter (CTR) mode in a consistent fashion. In the CC mode, the confidentiality and authenticity of data are assured by the CBC mode, while speed is achieved through the CTR mode. The proposed mode of operation overcomes the parallelization deficiency of the CBC mode and the chaining dependency of the counter mode. Experimental results indicate that the proposed CC mode achieves the encryption speed of the CTR mode, which is exceptionally faster than the encryption speed of the CBC mode. Moreover, our proposed CC mode provides better security over the CBC mode. In summary, the proposed CC block cipher mode of operation takes the advantages of both the Counter mode and the CBC mode, while avoiding their shortcomings.
This paper presents the infectious watermarking model (IWM) for the protection of video contents that are based on biological virus modeling by the infectious route and procedure. Our infectious watermarking is designed as a new paradigm protection for video contents, regarding the hidden watermark for video protection as an infectious virus, video content as host, and codec as contagion medium. We used pathogen, mutant, and contagion as the infectious watermark and defined the techniques of infectious watermark generation and authentication, kernel-based infectious watermarking, and content-based infectious watermarking. We experimented with our watermarking model by using existing watermarking methods as kernel-based infectious watermarking and content-based infectious watermarking medium, and verified the practical applications of our model based on these experiments.
Beacon scheduling is considered to be one of the most significant challenges for energy-efficient Low-Rate Wireless Personal Area Network (LR-WPAN) multi-hop networks. The emerging new standard, IEEE802.15.4e, contains a distributed beacon scheduling functionality that utilizes a specific bitmap and multi-superframe structure. However, this new standard does not provide a critical recipe for superframe duration (SD) allocation in beacon scheduling. Therefore, in this paper, we first introduce three different SD allocation approaches, LSB first, MSB first, and random. Via experiments we show that IEEE802.15.4e DSME beacon scheduling performs differently for different SD allocation schemes. Based on our experimental results we propose an adaptive SD allocation (ASDA) algorithm. It utilizes a single indicator, a distributed neighboring slot incrementer (DNSI). The experimental results demonstrate that the ASDA has a superior performance over other methods from the viewpoint of resource efficiency.
The stream cipher Salsa20 and its reduced versions are among the fastest stream ciphers available today. However, Salsa20/7 is broken and Salsa20/12 is not as safe as before. Therefore, Salsa20 must completely perform all of the four rounds of encryption to achieve a good diffusion in order to resist the known attacks. In this paper, a new variant of Salsa20 that uses the chaos theory and that can achieve diffusion faster than the original Salsa20 is presented. The method has been tested and benchmarked with the original Salsa20 with a series of tests. Most of the tests show that the proposed chaotic Salsa of two rounds is faster than the original four rounds of Salsa20/4, but it offers the same diffusion level.
Since 2005, all papers published in the JIPS are subjected to a peer review and upon acceptance are immediately made
permanently available free of charge for everyone worldwide to read and download from the journal’s homepage (http://www.jips-k.org)
without any subscription fee or personal registration. This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
The KIPS waives paper processing charges for submissions from international authors as well as society members. This waiver policy supports
and encourages the publication of quality papers, making the journal an international forum for the exchange of different ideas and experiences.
With regard to ethical standards, the JIPS takes plagiarism very seriously and thoroughly checks all articles.
The JIPS defines research ethics as securing objectivity and accuracy in the execution of research and the conclusion of results without any unintentional errors resulting from negligence or incorrect knowledge, etc.
and without any intentional misconduct such as falsification, plagiarism, etc. When an author submits a paper to the JIPS online submission and peer-review system,
he/she should also upload the separate file "author check list" which contains a statement that all his/her research has been performed in accordance with ethical standards.
Among the JIPS editorial board members, there are four associate manuscript editors who support the JIPS by dealing with any ethical problems associated with the publication process
and give advice on how to handle cases of suspected research and publication misconduct. When the JIPS managing editor looks over submitted papers and checks that they are suitable for further processing,
the managing editor also routes them to the CrossCheck service provided by iTenticate. Based on the results provided by the CrossCheck service, the JIPS associate manuscript editors inform the JIPS editor-in-chief of any plagiarism that is detected in a paper.
Then, the JIPS editor-in-chief communicates such detection to the author(s) while rejecting the paper.
ABOUT THE SOCIETY
Ever since information processing became one of the most important industries in the country, computing professionals have encountered a growing number of challenges.
Along with scholars and colleagues in related fields, they have gathered together at a variety of forums and meetings over the last few decades to share their knowledge and experiences,
and the outcomes of their research. These exchanges led to the founding of the Korea Information Processing Society (KIPS) on January 15, 1993. The KIPS was registered as an incorporated association under the Ministry of Science,
ICT and Future Planning under the government of the Republic of Korea. The main purpose of the KIPS organization is to improve our society by achieving the highest capability possible in the domain of information technology.
As such, it focuses on close collaboration with the nation’s industry, academic, and research communities to foster technological innovation,
to enhance its members' careers, and to promote the advanced information processing industry.