RonPub

Loading...

RonPub Banner

RonPub -- Research Online Publishing

RonPub (Research online Publishing) is an academic publisher of online, open access, peer-reviewed journals.  RonPub aims to provide a platform for researchers, developers, educators, and technical managers to share and exchange their research results worldwide.

RonPub Is Open Access:

RonPub publishes all of its journals under the open access model, defined under BudapestBerlin, and Bethesda open access declarations:

  • All articles published by RonPub is fully open access and online available to readers free of charge.  
  • All open access articles are distributed under  Creative Commons Attribution License,  which permits unrestricted use, distribution and reproduction free of charge in any medium, provided that the original work is properly cited. 
  • Authors retain all copyright to their work.
  • Authors may also publish the publisher's version of their paper on any repository or website. 

RonPub Is Cost-Effective:

To be able to provide open access journals, RonPub defray publishing cost by charging an one-time publication fee for each accepted article. One of RonPub objectives is providing a fast and high-quality but lower-cost publishing service. In order to ensure that the fee is never a barrier to publication, RonPub offers a fee waiver for authors who do not have funds to cover publication fees. We also offer a partial fee waiver for editors and reviewers of RonPub as as reward for their work. See the respective Journal webpage for the concrete publication fee.

RonPub Publication Ethics Statement:

In order to ensure the publishing quality and the reputation of the publisher, it is important that all parties involved in the act of publishing adhere to the standards of the publishing ethical behaviour. To verify the originality of submissions, we use Plagiarism Detection Tools, like Anti-Plagiarism, PaperRater, Viper, to check the content of manuscripts submitted to our journals against previous publications.

RonPub follows the Code of Conduct of the Committee on Publication Ethics (COPE), and deals with the cases of misconduct according to the COPE Flowcharts

Long-Term Preservation in the German National Library

Our publications are archived and permanently-preserved in the German National Library. The publications, which are archived in the German National Library, are not only long-term preserved but also accessible in the future, because the German National Library ensures that digital data saved in the old formats can be viewed and used on current computer systems in the same way they were on the original systems which are long obsolete.

Where is RonPub?

RonPub is a registered corporation in Lübeck, Germany. Lübeck is a beautiful harbour city, 60 kilometer away from Hamburg.

For Authors

Manuscript Preparation

Authors should first read the author guidelines of the corresponding journal. Manuscripts must be prepared using the manuscript template of the respective journal. It is available as word and latex version for download at the Author Guidelines of the corresponding journal page. The template describes the format and structure of manuscripts and other necessary information for preparing manuscripts. Manuscripts should be written in English. There is no restriction on the length of manuscripts.

Submission

Authors submit their manuscripts via the submit page of the corresponding journal. Authors first submit their manuscripts in PDF format. Once a manuscript is accepted, the author then submits the revised manuscript as PDF file and word file or latex folder (with all the material necessary to generate the PDF file). The work described in the submitted manuscript must be previously unpublished; it is not under consideration for publication anywhere else. 

Authors are welcome to suggest qualified reviewers for their papers, but this is not mandatory. If the author wants to do so, please provide the name, affiliations and e-mail addresses for all suggested reviewers.

Manuscript Status

After submission of manuscripts, authors will receive an email to confirm receipt of manuscripts within a few days. Subsequent enquiries concerning paper progress should be made to the corresponding editorial office (see individual journal webpage for concrete contact information).

Review Procedure

RonPub is committed to enforcing a rigorous peer-review process. All manuscripts submitted for publication in RonPub journals are strictly and thoroughly peer-reviewed. Our editorial offices will organize the peer-reviewing and collect three review reports per manuscript. The editorial board will make an accept/revision/reject decision based on the reports of reviewers. Authors will be informed with the decision and reviewing results within 6-8 weeks after the manuscript submission. Authors should perform an adequate revision to address the concerns from reviewers. A second round of peer-review will be performed if necessary.

Accepted manuscripts are published online immediately.

Copyrights

Authors publishing with RonPub open journals retain the copyright to their work. 

All articles published by RonPub is fully open access and online available to readers free of charge.  RonPub publishes all open access articles under the Creative Commons Attribution License,  which permits unrestricted use, distribution and reproduction freely, provided that the original work is properly cited.

Digital Archiving Policy

Our publications have been archived and permanently-preserved in the German National Library. The publications, which are archived in the German National Library, are not only long-term preserved but also accessible in the future, because the German National Library ensures that digital data saved in the old formats can be viewed and used on current computer systems in the same way they were on the original systems which are long obsolete. Further measures will be taken if necessary. Furthermore, we also encourage our authors to self-archive their articles published on the website of RonPub.

For Editors

About RonPub

RonPub is academic publisher of online, open access, peer-reviewed journals. All articles published by RonPub is fully open access and online available to readers free of charge.

RonPub is located in Lübeck, Germany. Lübeck is a beautiful harbour city, 60 kilometer away from Hamburg.

Editor-in-Chief Responsibilities

The Editor-in-Chief of each journal is mainly responsible for the scientific quality of the journal and for assisting in the management of the journal. The Editor-in-Chief suggests topics for the journal, invites distinguished scientists to join the editorial board, oversees the editorial process, and makes the final decision whether a paper can be published after peer-review and revisions.

As a reward for the work of a Editor-in-Chief, the Editor-in-Chief will obtain a 25% discount of the standard publication fee for her/his papers (the Editor-in-Chief is one of authors) published in any of RonPub journals.

Editors’ Responsibilities

Editors assist the Editor-in-Chief in the scientific quality and in decision about topics of the journal. Editors are also encouraged to help to promote the journal among their peers and at conferences. An editor invites at least three reviewers to review a manuscript, but may also review him-/herself the manuscript. After carefully evaluating the review reports and the manuscript itself, the editor makes a commendation about the status of the manuscript. The editor's evaluation as well as the review reports are then sent to EiC, who make the final decision whether a paper can be published after peer-review and revisions. 

The communication with Editorial Board members is done primarily by E-mail, and the Editors are expected to respond within a few working days on any question sent by the Editorial Office so that manuscripts can be processed in a timely fashion. If an editor does not respond or cannot process the work in time, and under some special situations, the editorial office may forward the requests to the Publishers or Editor-in-Chief, who will take the decision directly.

As a reward for the work of editors, an editor will obtain a 25% discount of the standard publication fee for her/his papers (the editor is one of authors) published in any of RonPub journals.

Guest Editors’ Responsibilities

Guest Editors are responsible of the scientific quality of their special issues. Guest Editors will be in charge of inviting papers, of supervising the refereeing process (each paper should be reviewed at least by three reviewers), and of making decisions on the acceptance of manuscripts submitted to their special issue. As regular issues, all accepted papers by (guest) editors will be sent to the EiC of the journal, who will check the quality of the papers, and make the final decsion whether a paper can be published.

Our editorial office will have the right directly asking authors to revise their paper if there are quality issues, e.g. weak quality of writing, and missing information. Authors are required to revise their paper several times if necessary. A paper accepted by it's quest editor may be rejected by the EiC of the journal due to a low quality. However, this occurs only when authors do not really take efforts to revise their paper. A high-quality publication needs the common efforts from the journal, reviewers, editors, editor-in-chief and authors.

The Guest Editors are also expected to write an editorial paper for the special issue. As a reward for work, all guest editors and reviewers working on a special issue will obtain a 25% discount of the standard publication fee for any of their papers published in any of RonPub journals for one year.

Reviewers’ Responsiblity

A reviewer is mainly responsible for reviewing of manuscripts, writing reviewing report and suggesting acception or deny of manuscripts. Reviews are encouraged to provide input about the quality and management of the journal, and help promote the journal among their peers and at conferences.  

Upon the quality of reviewing work, a reviewer will have the potential to be promoted to a full editorial board member. 

As a reward for the reviewing work, a reviewer will obtain a 25% discount of the standard publication fee for her/his papers (the review is one of authors) published in any of RonPub journals.

Launching New Journals

RonPub always welcomes suggestions for new open access journals in any research area. We are also open for publishing collaborations with research societies. Please send your proposals for new journals or for publishing collaboration to This email address is being protected from spambots. You need JavaScript enabled to view it. .

Publication Criteria

This part provides important information for both the scientific committees and authors.

Ethic Requirement:

For scientific committees: Each editor and reviewer should conduct the evaluation of manuscripts objectively and fairly.
For authors: Authors should present their work honestly without fabrication, falsification, plagiarism or inappropriate data manipulation.

Before review:

In order to filter fabricated submissions, the editorial office will check the authenticity of the authors and their affiliations before a peer-review begins. It is important that the authors communicate with us using the email addresses of their affiliations and provide us the URL addresses of their affiliations. The overall quality of paper will be also checked including format, figures, tables, integrity and adequacy. Authors may be required to improve the quality of their paper before sending it out for review. If a paper is obviously of low quality, the paper will be directly rejected.

Acceptance Criteria:

The criteria for acceptance of manuscripts are the quality of work. This will concretely be reflected in the following aspects:

  • Novelty and Practical Impact
  • Technical Soundness
  • Appropriateness and Adequacy of 
    • Literature Review
    • Background Discussion
    • Analysis of Issues
  • Presentation, including 
    • Overall Organization 
    • English 
    • Readability

For a contribution to be acceptable for publication, these points should be at least in middle level.

Guidelines for Rejection:

  • If the work described in the manuscript has been published, or is under consideration for publication anywhere else, it will not be evaluated.
  • If the work is a plagiarism, or contains data falsification or fabrication, it will be rejected.
  • Manuscripts, which have seriously technical flaws, will not be accepted.

Call for Journals

Research Online Publishing (RonPub, www.ronpub.com) is a publisher of online, open access and peer-reviewed scientific journals.  For more information about RonPub please visit this link.

RonPub always welcomes suggestions for new journals in any research area. Please send your proposals for journals along with your Curriculum Vitae to This email address is being protected from spambots. You need JavaScript enabled to view it. .

We are also open for publishing collaborations with research societies. Please send your publishing collaboration also to This email address is being protected from spambots. You need JavaScript enabled to view it. .

Be an Editor / Be a Reviewer

RonPub always welcomes qualified academicians and practitioners to join as editors and reviewers. Being an editor/a reviewer is a matter of prestige and personnel achievement. Upon the quality of reviewing work, a reviewer will have the potential to be promoted to a full editorial board member.

If you would like to participate as a scientific committee member of any of RonPub journals, please send an email to This email address is being protected from spambots. You need JavaScript enabled to view it. with your curriculum vitae. We will revert back as soon as possible. For more information about editors/reviewers, please visit this link.

Contact RonPub

Location

RonPub UG (haftungsbeschränkt)
Hiddenseering 30
23560 Lübeck
Germany

Comments and Questions

For general inquiries, please e-mail to This email address is being protected from spambots. You need JavaScript enabled to view it. .

For specific questions on a certain journal, please visit the corresponding journal page to see the email address.

  1.  Open Access 

    Which NoSQL Database? A Performance Overview

    Veronika Abramova, Jorge Bernardino, Pedro Furtado

    Open Journal of Databases (OJDB), 1(2), Pages 17-24, 2014, Downloads: 12759, Citations: 24

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194607 | GNL-LP: 1132360862 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: NoSQL data stores are widely used to store and retrieve possibly large amounts of data, typically in a key-value format. There are many NoSQL types with different performances, and thus it is important to compare them in terms of performance and verify how the performance is related to the database type. In this paper, we evaluate five most popular NoSQL databases: Cassandra, HBase, MongoDB, OrientDB and Redis. We compare those databases in terms of query performance, based on reads and updates, taking into consideration the typical workloads, as represented by the Yahoo! Cloud Serving Benchmark. This comparison allows users to choose the most appropriate database according to the specific mechanisms and application needs.

    BibTex:

        @Article{OJDB-v1i2n02_Abramova,
            title     = {Which NoSQL Database? A Performance Overview},
            author    = {Veronika Abramova and
                         Jorge Bernardino and
                         Pedro Furtado},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {17--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194607},
            urn       = {urn:nbn:de:101:1-201705194607},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {NoSQL data stores are widely used to store and retrieve possibly large amounts of data, typically in a key-value format. There are many NoSQL types with different performances, and thus it is important to compare them in terms of performance and verify how the performance is related to the database type. In this paper, we evaluate five most popular NoSQL databases: Cassandra, HBase, MongoDB, OrientDB and Redis. We compare those databases in terms of query performance, based on reads and updates, taking into consideration the typical workloads, as represented by the Yahoo! Cloud Serving Benchmark. This comparison allows users to choose the most appropriate database according to the specific mechanisms and application needs.}
        }
    
  2.  Open Access 

    A Comparative Evaluation of Current HTML5 Web Video Implementations

    Martin Hoernig, Andreas Bigontina, Bernd Radig

    Open Journal of Web Technologies (OJWT), 1(2), Pages 1-9, 2014, Downloads: 12687, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291328 | GNL-LP: 1133021514 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: HTML5 video is the upcoming standard for playing videos on the World Wide Web. Although its specification has not been fully adopted yet, all major browsers provide the HTML5 video element and web developers already rely on its functionality. But there are differences between implementations and inaccuracies that trouble the web developer community. To help to improve the current situation we draw a comparison between the most important web browsers. We focus on the event mechanism, since it is essential for interacting with the video element. Furthermore, we compare the seeking accuracy, which is relevant for more specialized applications. Our tests reveal varieties of differences between browser interfaces and show that even simple software solutions may still need third-party plugins in today's browsers.

    BibTex:

        @Article{OJWT-v1i2n01_Hoernig,
            title     = {A Comparative Evaluation of Current HTML5 Web Video Implementations},
            author    = {Martin Hoernig and
                         Andreas Bigontina and
                         Bernd Radig},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--9},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291328},
            urn       = {urn:nbn:de:101:1-201705291328},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {HTML5 video is the upcoming standard for playing videos on the World Wide Web. Although its specification has not been fully adopted yet, all major browsers provide the HTML5 video element and web developers already rely on its functionality. But there are differences between implementations and inaccuracies that trouble the web developer community. To help to improve the current situation we draw a comparison between the most important web browsers. We focus on the event mechanism, since it is essential for interacting with the video element. Furthermore, we compare the seeking accuracy, which is relevant for more specialized applications. Our tests reveal varieties of differences between browser interfaces and show that even simple software solutions may still need third-party plugins in today's browsers.}
        }
    
  3.  Open Access 

    Scalable Distributed Computing Hierarchy: Cloud, Fog and Dew Computing

    Karolj Skala, Davor Davidovic, Enis Afgan, Ivan Sovic, Zorislav Sojat

    Open Journal of Cloud Computing (OJCC), 2(1), Pages 16-24, 2015, Downloads: 7132, Citations: 40

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194519 | GNL-LP: 1132360749 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The paper considers the conceptual approach for organization of the vertical hierarchical links between the scalable distributed computing paradigms: Cloud Computing, Fog Computing and Dew Computing. In this paper, the Dew Computing is described and recognized as a new structural layer in the existing distributed computing hierarchy. In the existing computing hierarchy, the Dew computing is positioned as the ground level for the Cloud and Fog computing paradigms. Vertical, complementary, hierarchical division from Cloud to Dew Computing satisfies the needs of high- and low-end computing demands in everyday life and work. These new computing paradigms lower the cost and improve the performance, particularly for concepts and applications such as the Internet of Things (IoT) and the Internet of Everything (IoE). In addition, the Dew computing paradigm will require new programming models that will efficiently reduce the complexity and improve the productivity and usability of scalable distributed computing, following the principles of High-Productivity computing.

    BibTex:

        @Article{OJCC_2015v2i1n03_Skala,
            title     = {Scalable Distributed Computing Hierarchy: Cloud, Fog and Dew Computing},
            author    = {Karolj Skala and
                         Davor Davidovic and
                         Enis Afgan and
                         Ivan Sovic and
                         Zorislav Sojat},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {16--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194519},
            urn       = {urn:nbn:de:101:1-201705194519},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The paper considers the conceptual approach for organization of the vertical hierarchical links between the scalable distributed computing paradigms: Cloud Computing, Fog Computing and Dew Computing. In this paper, the Dew Computing is described and recognized as a new structural layer in the existing distributed computing hierarchy. In the existing computing hierarchy, the Dew computing is positioned as the ground level for the Cloud and Fog computing paradigms. Vertical, complementary, hierarchical division from Cloud to Dew Computing satisfies the needs of high- and low-end computing demands in everyday life and work. These new computing paradigms lower the cost and improve the performance, particularly for concepts and applications such as the Internet of Things (IoT) and the Internet of Everything (IoE). In addition, the Dew computing paradigm will require new programming models that will efficiently reduce the complexity and improve the productivity and usability of scalable distributed computing, following the principles of High-Productivity computing.}
        }
    
  4.  Open Access 

    Eventual Consistent Databases: State of the Art

    Mawahib Musa Elbushra, Jan Lindström

    Open Journal of Databases (OJDB), 1(1), Pages 26-41, 2014, Downloads: 6701, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194582 | GNL-LP: 1132360846 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: One of the challenges of cloud programming is to achieve the right balance between the availability and consistency in a distributed database. Cloud computing environments, particularly cloud databases, are rapidly increasing in importance, acceptance and usage in major applications, which need the partition-tolerance and availability for scalability purposes, but sacrifice the consistency side (CAP theorem). In these environments, the data accessed by users is stored in a highly available storage system, thus the use of paradigms such as eventual consistency became more widespread. In this paper, we review the state-of-the-art database systems using eventual consistency from both industry and research. Based on this review, we discuss the advantages and disadvantages of eventual consistency, and identify the future research challenges on the databases using eventual consistency.

    BibTex:

        @Article{OJDB-v1i1n03_Elbushra,
            title     = {Eventual Consistent Databases: State of the Art},
            author    = {Mawahib Musa Elbushra and
                         Jan Lindstr{\"o}m},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {26--41},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194582},
            urn       = {urn:nbn:de:101:1-201705194582},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {One of the challenges of cloud programming is to achieve the right balance between the availability and consistency in a distributed database. Cloud computing environments, particularly cloud databases, are rapidly increasing in importance, acceptance and usage in major applications, which need the partition-tolerance and availability for scalability purposes, but sacrifice the consistency side (CAP theorem). In these environments, the data accessed by users is stored in a highly available storage system, thus the use of paradigms such as eventual consistency became more widespread. In this paper, we review the state-of-the-art database systems using eventual consistency from both industry and research. Based on this review, we discuss the advantages and disadvantages of eventual consistency, and identify the future research challenges on the databases using eventual consistency.}
        }
    
  5.  Open Access 

    Accurate Distance Estimation between Things: A Self-correcting Approach

    Ho-sik Cho, Jianxun Ji, Zili Chen, Hyuncheol Park, Wonsuk Lee

    Open Journal of Internet Of Things (OJIOT), 1(2), Pages 19-27, 2015, Downloads: 6454, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244959 | GNL-LP: 1130622525 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper suggests a method to measure the physical distance between an IoT device (a Thing) and a mobile device (also a Thing) using BLE (Bluetooth Low-Energy profile) interfaces with smaller distance errors. BLE is a well-known technology for the low-power connectivity and suitable for IoT devices as well as for the proximity with the range of several meters. Apple has already adopted the technique and enhanced it to provide subdivided proximity range levels. However, as it is also a variation of RSS-based distance estimation, Apple's iBeacon could only provide immediate, near or far status but not a real and accurate distance. To provide more accurate distance using BLE, this paper introduces additional self-correcting beacon to calibrate the reference distance and mitigate errors from environmental factors. By adopting self-correcting beacon for measuring the distance, the average distance error shows less than 10% within the range of 1.5 meters. Some considerations are presented to extend the range to be able to get more accurate distances.

    BibTex:

        @Article{OJIOT_2015v1i2n03_Cho,
            title     = {Accurate Distance Estimation between Things: A Self-correcting Approach},
            author    = {Ho-sik Cho and
                         Jianxun Ji and
                         Zili Chen and
                         Hyuncheol Park and
                         Wonsuk Lee},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {19--27},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244959},
            urn       = {urn:nbn:de:101:1-201704244959},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper suggests a method to measure the physical distance between an IoT device (a Thing) and a mobile device (also a Thing) using BLE (Bluetooth Low-Energy profile) interfaces with smaller distance errors. BLE is a well-known technology for the low-power connectivity and suitable for IoT devices as well as for the proximity with the range of several meters. Apple has already adopted the technique and enhanced it to provide subdivided proximity range levels. However, as it is also a variation of RSS-based distance estimation, Apple's iBeacon could only provide immediate, near or far status but not a real and accurate distance. To provide more accurate distance using BLE, this paper introduces additional self-correcting beacon to calibrate the reference distance and mitigate errors from environmental factors. By adopting self-correcting beacon for measuring the distance, the average distance error shows less than 10\% within the range of 1.5 meters. Some considerations are presented to extend the range to be able to get more accurate distances.}
        }
    
  6.  Open Access 

    Criteria of Successful IT Projects from Management's Perspective

    Mark Harwardt

    Open Journal of Information Systems (OJIS), 3(1), Pages 29-54, 2016, Downloads: 4436, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194797 | GNL-LP: 1132361133 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The aim of this paper is to compile a model of IT project success from management's perspective. Therefore, a qualitative research approach is proposed by interviewing IT managers on how their companies evaluate the success of IT projects. The evaluation of the survey provides fourteen success criteria and four success dimensions. This paper also thoroughly analyzes which of these criteria the management considers especially important and which ones are being missed in daily practice. Additionally, it attempts to identify the relevance of the discovered criteria and dimensions with regard to the determination of IT project success. It becomes evident here that the old-fashioned Iron Triangle still plays a leading role, but some long-term strategical criteria, such as value of the project, customer perspective or impact on the organization, have meanwhile caught up or pulled even.

    BibTex:

        @Article{OJIS_2016v3i1n02_Harwardt,
            title     = {Criteria of Successful IT Projects from Management's Perspective},
            author    = {Mark Harwardt},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {29--54},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194797},
            urn       = {urn:nbn:de:101:1-201705194797},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The aim of this paper is to compile a model of IT project success from management's perspective. Therefore, a qualitative research approach is proposed by interviewing IT managers on how their companies evaluate the success of IT projects. The evaluation of the survey provides fourteen success criteria and four success dimensions. This paper also thoroughly analyzes which of these criteria the management considers especially important and which ones are being missed in daily practice. Additionally, it attempts to identify the relevance of the discovered criteria and dimensions with regard to the determination of IT project success. It becomes evident here that the old-fashioned Iron Triangle still plays a leading role, but some long-term strategical criteria, such as value of the project, customer perspective or impact on the organization, have meanwhile caught up or pulled even.}
        }
    
  7.  Open Access 

    The Potential of Printed Electronics and Personal Fabrication in Driving the Internet of Things

    Paulo Rosa, António Câmara, Cristina Gouveia

    Open Journal of Internet Of Things (OJIOT), 1(1), Pages 16-36, 2015, Downloads: 3811, Citations: 9

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244933 | GNL-LP: 1130621448 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In the early nineties, Mark Weiser, a chief scientist at the Xerox Palo Alto Research Center (PARC), wrote a series of seminal papers that introduced the concept of Ubiquitous Computing. Within this vision, computers and others digital technologies are integrated seamlessly into everyday objects and activities, hidden from our senses whenever not used or needed. An important facet of this vision is the interconnectivity of the various physical devices, which creates an Internet of Things. With the advent of Printed Electronics, new ways to link the physical and digital worlds became available. Common printing technologies, such as screen, flexography, and inkjet printing, are now starting to be used not only to mass-produce extremely thin, flexible and cost effective electronic circuits, but also to introduce electronic functionality into objects where it was previously unavailable. In turn, the growing accessibility to Personal Fabrication tools is leading to the democratization of the creation of technology by enabling end-users to design and produce their own material goods according to their needs. This paper presents a survey of commonly used technologies and foreseen applications in the field of Printed Electronics and Personal Fabrication, with emphasis on the potential to drive the Internet of Things.

    BibTex:

        @Article{OJIOT_2015v1i1n03_Rosa,
            title     = {The Potential of Printed Electronics and Personal Fabrication in Driving the Internet of Things},
            author    = {Paulo Rosa and
                         Ant\'{o}nio C\^{a}mara and
                         Cristina Gouveia},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {16--36},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244933},
            urn       = {urn:nbn:de:101:1-201704244933},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In the early nineties, Mark Weiser, a chief scientist at the Xerox Palo Alto Research Center (PARC), wrote a series of seminal papers that introduced the concept of Ubiquitous Computing. Within this vision, computers and others digital technologies are integrated seamlessly into everyday objects and activities, hidden from our senses whenever not used or needed. An important facet of this vision is the interconnectivity of the various physical devices, which creates an Internet of Things. With the advent of Printed Electronics, new ways to link the physical and digital worlds became available. Common printing technologies, such as screen, flexography, and inkjet printing, are now starting to be used not only to mass-produce extremely thin, flexible and cost effective electronic circuits, but also to introduce electronic functionality into objects where it was previously unavailable. In turn, the growing accessibility to Personal Fabrication tools is leading to the democratization of the creation of technology by enabling end-users to design and produce their own material goods according to their needs. This paper presents a survey of commonly used technologies and foreseen applications in the field of Printed Electronics and Personal Fabrication, with emphasis on the potential to drive the Internet of Things.}
        }
    
  8.  Open Access 

    Causal Consistent Databases

    Mawahib Musa Elbushra, Jan Lindström

    Open Journal of Databases (OJDB), 2(1), Pages 17-35, 2015, Downloads: 3713, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194619 | GNL-LP: 1132360870 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Many consistency criteria have been considered in databases and the causal consistency is one of them. The causal consistency model has gained much attention in recent years because it provides ordering of relative operations. The causal consistency requires that all writes, which are potentially causally related, must be seen in the same order by all processes. The causal consistency is a weaker criteria than the sequential consistency, because there exists an execution, which is causally consistent but not sequentially consistent, however all executions satisfying the sequential consistency are also causally consistent. Furthermore, the causal consistency supports non-blocking operations; i.e. processes may complete read or write operations without waiting for global computation. Therefore, the causal consistency overcomes the primary limit of stronger criteria: communication latency. Additionally, several application semantics are precisely captured by the causal consistency, e.g. collaborative tools. In this paper, we review the state-of-the-art of causal consistent databases, discuss the features, functionalities and applications of the causal consistency model, and systematically compare it with other consistency models. We also discuss the implementation of causal consistency databases and identify limitations of the causal consistency model.

    BibTex:

        @Article{OJDB_2015v2i1n02_Elbushra,
            title     = {Causal Consistent Databases},
            author    = {Mawahib Musa Elbushra and
                         Jan Lindstr{\"o}m},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {17--35},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194619},
            urn       = {urn:nbn:de:101:1-201705194619},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Many consistency criteria have been considered in databases and the causal consistency is one of them. The causal consistency model has gained much attention in recent years because it provides ordering of relative operations. The causal consistency requires that all writes, which are potentially causally related, must be seen in the same order by all processes. The causal consistency is a weaker criteria than the sequential consistency, because there exists an execution, which is causally consistent but not sequentially consistent, however all executions satisfying the sequential consistency are also causally consistent. Furthermore, the causal consistency supports non-blocking operations; i.e. processes may complete read or write operations without waiting for global computation. Therefore, the causal consistency overcomes the primary limit of stronger criteria: communication latency. Additionally, several application semantics are precisely captured by the causal consistency, e.g. collaborative tools. In this paper, we review the state-of-the-art of causal consistent databases, discuss the features, functionalities and applications of the causal consistency model, and systematically compare it with other consistency models. We also discuss the implementation of causal consistency databases and identify limitations of the causal consistency model.}
        }
    
  9.  Open Access 

    Block-level De-duplication with Encrypted Data

    Pasquale Puzio, Refik Molva, Melek Önen, Sergio Loureiro

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 10-18, 2014, Downloads: 3656, Citations: 12

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194448 | GNL-LP: 1132360617 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Deduplication is a storage saving technique which has been adopted by many cloud storage providers such as Dropbox. The simple principle of deduplication is that duplicate data uploaded by different users are stored only once. Unfortunately, deduplication is not compatible with encryption. As a scheme that allows deduplication of encrypted data segments, we propose ClouDedup, a secure and efficient storage service which guarantees blocklevel deduplication and data confidentiality at the same time. ClouDedup strengthens convergent encryption by employing a component that implements an additional encryption operation and an access control mechanism. We also propose to introduce an additional component which is in charge of providing a key management system for data blocks together with the actual deduplication operation. We show that the overhead introduced by these new components is minimal and does not impact the overall storage and computational costs.

    BibTex:

        @Article{OJCC-v1i1n02_Puzio,
            title     = {Block-level De-duplication with Encrypted Data},
            author    = {Pasquale Puzio and
                         Refik Molva and
                         Melek {\"O}nen and
                         Sergio Loureiro},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {10--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194448},
            urn       = {urn:nbn:de:101:1-201705194448},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Deduplication is a storage saving technique which has been adopted by many cloud storage providers such as Dropbox. The simple principle of deduplication is that duplicate data uploaded by different users are stored only once. Unfortunately, deduplication is not compatible with encryption. As a scheme that allows deduplication of encrypted data segments, we propose ClouDedup, a secure and efficient storage service which guarantees blocklevel deduplication and data confidentiality at the same time. ClouDedup strengthens convergent encryption by employing a component that implements an additional encryption operation and an access control mechanism. We also propose to introduce an additional component which is in charge of providing a key management system for data blocks together with the actual deduplication operation. We show that the overhead introduced by these new components is minimal and does not impact the overall storage and computational costs.}
        }
    
  10.  Open Access 

    Big Data in the Cloud: A Survey

    Pedro Caldeira Neves, Jorge Bernardino

    Open Journal of Big Data (OJBD), 1(2), Pages 1-18, 2015, Downloads: 3316, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194365 | GNL-LP: 1132360528 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Big Data has become a hot topic across several business areas requiring the storage and processing of huge volumes of data. Cloud computing leverages Big Data by providing high storage and processing capabilities and enables corporations to consume resources in a pay-as-you-go model making clouds the optimal environment for storing and processing huge quantities of data. By using virtualized resources, Cloud can scale very easily, be highly available and provide massive storage capacity and processing power. This paper surveys existing databases models to store and process Big Data within a Cloud environment. Particularly, we detail the following traditional NoSQL databases: BigTable, Cassandra, DynamoDB, HBase, Hypertable, and MongoDB. The MapReduce framework and its developments Apache Spark, HaLoop, Twister, and other alternatives such as Apache Giraph, GraphLab, Pregel and MapD - a novel platform that uses GPU processing to accelerate Big Data processing - are also analyzed. Finally, we present two case studies that demonstrate the successful use of Big Data within Cloud environments and the challenges that must be addressed in the future.

    BibTex:

        @Article{OJBD_2015v1i2n02_Neves,
            title     = {Big Data in the Cloud: A Survey},
            author    = {Pedro Caldeira Neves and
                         Jorge Bernardino},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194365},
            urn       = {urn:nbn:de:101:1-201705194365},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Big Data has become a hot topic across several business areas requiring the storage and processing of huge volumes of data. Cloud computing leverages Big Data by providing high storage and processing capabilities and enables corporations to consume resources in a pay-as-you-go model making clouds the optimal environment for storing and processing huge quantities of data. By using virtualized resources, Cloud can scale very easily, be highly available and provide massive storage capacity and processing power. This paper surveys existing databases models to store and process Big Data within a Cloud environment. Particularly, we detail the following traditional NoSQL databases: BigTable, Cassandra, DynamoDB, HBase, Hypertable, and MongoDB. The MapReduce framework and its developments Apache Spark, HaLoop, Twister, and other alternatives such as Apache Giraph, GraphLab, Pregel and MapD - a novel platform that uses GPU processing to accelerate Big Data processing - are also analyzed. Finally, we present two case studies that demonstrate the successful use of Big Data within Cloud environments and the challenges that must be addressed in the future.}
        }
    
  11.  Open Access 

    Developing Knowledge Models of Social Media: A Case Study on LinkedIn

    Jinwu Li, Vincent Wade, Melike Sah

    Open Journal of Semantic Web (OJSW), 1(2), Pages 1-24, 2014, Downloads: 3001

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194841 | GNL-LP: 1132361206 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: User Generated Content (UGC) exchanged via large Social Network is considered a very important knowledge source about all aspects of the social engagements (e.g. interests, events, personal information, personal preferences, social experience, skills etc.). However this data is inherently unstructured or semi-structured. In this paper, we describe the results of a case study on LinkedIn Ireland public profiles. The study investigated how the available knowledge could be harvested from LinkedIn in a novel way by developing and applying a reusable knowledge model using linked open data vocabularies and semantic web. In addition, the paper discusses the crawling and data normalisation strategies that we developed, so that high quality metadata could be extracted from the LinkedIn public profiles. Apart from the search engine in LinkedIn.com itself, there are no well known publicly available endpoints that allow users to query knowledge concerning the interests of individuals on LinkedIn. In particular, we present a system that extracts and converts information from raw web pages of LinkedIn public profiles into a machine-readable, interoperable format using data mining and Semantic Web technologies. The outcomes of our research can be summarized as follows: (1) A reusable knowledge model which can represent LinkedIn public users and company profiles using linked data vocabularies and structured data, (2) a public SPARQL endpoint to access structured data about Irish industry and public profiles, (3) a scalable data crawling strategy and mashup based data normalisation approach. The proposed data mining and knowledge representation proposed in this paper are evaluated in four ways: (1) We evaluate metadata quality using automated techniques, such as data completeness and data linkage. (2) Data accuracy is evaluated via user studies. In particular, accuracy is evaluated by comparison of manually entered metadata fields and the metadata which was automatically extracted. (3) User perceived metadata quality is measured by asking users to rate the automatically extracted metadata in user studies. (4) Finally, the paper discusses how the extracted metadata suits for a user interface design. Overall, the evaluations show that the extracted metadata is of high quality and meets the requirements of a data visualisation user interface.

    BibTex:

        @Article{OJSW-v1i2n01_Li,
            title     = {Developing Knowledge Models of Social Media: A Case Study on LinkedIn},
            author    = {Jinwu Li and
                         Vincent Wade and
                         Melike Sah},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194841},
            urn       = {urn:nbn:de:101:1-201705194841},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {User Generated Content (UGC) exchanged via large Social Network is considered a very important knowledge source about all aspects of the social engagements (e.g. interests, events, personal information, personal preferences, social experience, skills etc.). However this data is inherently unstructured or semi-structured. In this paper, we describe the results of a case study on LinkedIn Ireland public profiles. The study investigated how the available knowledge could be harvested from LinkedIn in a novel way by developing and applying a reusable knowledge model using linked open data vocabularies and semantic web. In addition, the paper discusses the crawling and data normalisation strategies that we developed, so that high quality metadata could be extracted from the LinkedIn public profiles. Apart from the search engine in LinkedIn.com itself, there are no well known publicly available endpoints that allow users to query knowledge concerning the interests of individuals on LinkedIn. In particular, we present a system that extracts and converts information from raw web pages of LinkedIn public profiles into a machine-readable, interoperable format using data mining and Semantic Web technologies. The outcomes of our research can be summarized as follows: (1) A reusable knowledge model which can represent LinkedIn public users and company profiles using linked data vocabularies and structured data, (2) a public SPARQL endpoint to access structured data about Irish industry and public profiles, (3) a scalable data crawling strategy and mashup based data normalisation approach. The proposed data mining and knowledge representation proposed in this paper are evaluated in four ways: (1) We evaluate metadata quality using automated techniques, such as data completeness and data linkage. (2) Data accuracy is evaluated via user studies. In particular, accuracy is evaluated by comparison of manually entered metadata fields and the metadata which was automatically extracted. (3) User perceived metadata quality is measured by asking users to rate the automatically extracted metadata in user studies. (4) Finally, the paper discusses how the extracted metadata suits for a user interface design. Overall, the evaluations show that the extracted metadata is of high quality and meets the requirements of a data visualisation user interface.}
        }
    
  12.  Open Access 

    Using Business Intelligence to Improve DBA Productivity

    Eric A. Mortensen, En Cheng

    Open Journal of Databases (OJDB), 1(2), Pages 1-16, 2014, Downloads: 2860

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194595 | GNL-LP: 1132360854 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The amount of data collected and used by companies has grown rapidly in size over the last decade. Business leaders are now using Business Intelligence (BI) systems to make effective business decisions against large amounts of data. The growth in the size of data has been a major challenge for Database Administrators (DBAs). The increase in the number and size of databases at the speed they have grown has made it difficult for DBA teams to provide the same level of service that the business requires they provide. The methods that DBAs have used in the last several decades can no longer be performed with the efficiency needed over all of the databases they administer. This paper presents the first BI system to improve DBA productivity and providing important data metrics for Information Technology (IT) managers. The BI system has been well received by Sherwin Williams Database Administrators. It has i) enabled the DBA team to quickly determine which databases needed work by a DBA without manually logging into the system; ii) helped the DBA team and its management to easily answer other business users' questions without using DBAs' time to research the issue; and iii) helped the DBA team to provide the business data for unanticipated audit request.

    BibTex:

        @Article{OJDB-v1i2n01_Mortensen,
            title     = {Using Business Intelligence to Improve DBA Productivity},
            author    = {Eric A. Mortensen and
                         En Cheng},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--16},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194595},
            urn       = {urn:nbn:de:101:1-201705194595},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The amount of data collected and used by companies has grown rapidly in size over the last decade.  Business leaders are now using Business Intelligence (BI) systems to make effective business decisions against large amounts of data. The growth in the size of data has been a major challenge for Database Administrators (DBAs). The increase in the number and size of databases at the speed they have grown has made it difficult for DBA teams to provide the same level of service that the business requires they provide. The methods that DBAs have used in the last several decades can no longer be performed with the efficiency needed over all of the databases they administer. This paper presents the first BI system to improve DBA productivity and providing important data metrics for Information Technology (IT) managers. The BI system has been well received by Sherwin Williams Database Administrators.  It has i) enabled the DBA team to quickly determine which databases needed work by a DBA without manually logging into the system; ii) helped the DBA team and its management to easily answer other business users' questions without using DBAs' time to research the issue; and iii) helped the DBA team to provide the business data for unanticipated audit request.}
        }
    
  13.  Open Access 

    Getting Indexed by Bibliographic Databases in the Area of Computer Science

    Arne Kusserow, Sven Groppe

    Open Journal of Web Technologies (OJWT), 1(2), Pages 10-27, 2014, Downloads: 2823, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291343 | GNL-LP: 1133021557 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Every author and publisher is interested in adding their publications to the widely used bibliographic databases freely accessible in the world wide web: This ensures the visibility of their publications and hence of the published research. However, the inclusion requirements of publications in the bibliographic databases are heterogeneous even on the technical side. This survey paper aims in shedding light on the various data formats, protocols and technical requirements of getting indexed by widely used bibliographic databases in the area of computer science and provides hints for maximal database inclusion. Furthermore, we point out the possibilities to utilize the data of bibliographic databases, and describes some personal and institutional research repository systems with special regard to the support of inclusion in bibliographic databases.

    BibTex:

        @Article{OJWT_2014v1i2n02_Kusserow,
            title     = {Getting Indexed by Bibliographic Databases in the Area of Computer Science},
            author    = {Arne Kusserow and
                         Sven Groppe},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {10--27},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291343},
            urn       = {urn:nbn:de:101:1-201705291343},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Every author and publisher is interested in adding their publications to the widely used bibliographic databases freely accessible in the world wide web: This ensures the visibility of their publications and hence of the published research. However, the inclusion requirements of publications in the bibliographic databases are heterogeneous even on the technical side. This survey paper aims in shedding light on the various data formats, protocols and technical requirements of getting indexed by widely used bibliographic databases in the area of computer science and provides hints for maximal database inclusion. Furthermore, we point out the possibilities to utilize the data of bibliographic databases, and describes some personal and institutional research repository systems with special regard to the support of inclusion in bibliographic databases.}
        }
    
  14.  Open Access 

    Evidential Sensor Data Fusion in a Smart City Environment

    Aditya Gaur, Bryan W. Scotney, Gerard P. Parr, Sally I. McClean

    Open Journal of Internet Of Things (OJIOT), 1(2), Pages 1-18, 2015, Downloads: 2799, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244969 | GNL-LP: 113062319X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Wireless sensor networks have increasingly become contributors of very large amounts of data. The recent deployment of wireless sensor networks in Smart City infrastructures have led to very large amounts of data being generated each day across a variety of domains, with applications including environmental monitoring, healthcare monitoring and transport monitoring. The information generated through the wireless sensor nodes has made possible the visualization of a Smart City environment for better living. The Smart City offers intelligent infrastructure and cogitative environment for the elderly and other people living in the Smart society. Different types of sensors are present that help in monitoring inhabitants' behaviour and their interaction with real world objects. To take advantage of the increasing amounts of data, there is a need for new methods and techniques for effective data management and analysis, to generate information that can assist in managing the resources intelligently and dynamically. Through this research a Smart City ontology model is proposed, which addresses the fusion process related to uncertain sensor data using semantic web technologies and Dempster-Shafer uncertainty theory. Based on the information handling methods, such as Dempster-Shafer theory (DST), an equally weighted sum operator and maximization operation, a higher level of contextual information is inferred from the low-level sensor data fusion process. In addition, the proposed ontology model helps in learning new rules that can be used in defining new knowledge in the Smart City system.

    BibTex:

        @Article{OJIOT_2015v1i2n02_Gaur,
            title     = {Evidential Sensor Data Fusion in a Smart City Environment},
            author    = {Aditya Gaur and
                         Bryan W. Scotney and
                         Gerard P. Parr and
                         Sally I. McClean},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244969},
            urn       = {urn:nbn:de:101:1-201704244969},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Wireless sensor networks have increasingly become contributors of very large amounts of data. The recent deployment of wireless sensor networks in Smart City infrastructures have led to very large amounts of data being generated each day across a variety of domains, with applications including environmental monitoring, healthcare monitoring and transport monitoring. The information generated through the wireless sensor nodes has made possible the visualization of a Smart City environment for better living. The Smart City offers intelligent infrastructure and cogitative environment for the elderly and other people living in the Smart society. Different types of sensors are present that help in monitoring inhabitants' behaviour and their interaction with real world objects. To take advantage of the increasing amounts of data, there is a need for new methods and techniques for effective data management and analysis, to generate information that can assist in managing the resources intelligently and dynamically. Through this research a Smart City ontology model is proposed, which addresses the fusion process related to uncertain sensor data using semantic web technologies and Dempster-Shafer uncertainty theory. Based on the information handling methods, such as Dempster-Shafer theory (DST), an equally weighted sum operator and maximization operation, a higher level of contextual information is inferred from the low-level sensor data fusion process. In addition, the proposed ontology model helps in learning new rules that can be used in defining new knowledge in the Smart City system.}
        }
    
  15.  Open Access 

    A Self-Optimizing Cloud Computing System for Distributed Storage and Processing of Semantic Web Data

    Sven Groppe, Johannes Blume, Dennis Heinrich, Stefan Werner

    Open Journal of Cloud Computing (OJCC), 1(2), Pages 1-14, 2014, Downloads: 2743, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194478 | GNL-LP: 113236065X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Clouds are dynamic networks of common, off-the-shell computers to build computation farms. The rapid growth of databases in the context of the semantic web requires efficient ways to store and process this data. Using cloud technology for storing and processing Semantic Web data is an obvious way to overcome difficulties in storing and processing the enormously large present and future datasets of the Semantic Web. This paper presents a new approach for storing Semantic Web data, such that operations for the evaluation of Semantic Web queries are more likely to be processed only on local data, instead of using costly distributed operations. An experimental evaluation demonstrates the performance improvements in comparison to a naive distribution of Semantic Web data.

    BibTex:

        @Article{OJCC-v1i2n01_Groppe,
            title     = {A Self-Optimizing Cloud Computing System for Distributed Storage and Processing of Semantic Web Data},
            author    = {Sven Groppe and
                         Johannes Blume and
                         Dennis Heinrich and
                         Stefan Werner},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--14},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194478},
            urn       = {urn:nbn:de:101:1-201705194478},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Clouds are dynamic networks of common, off-the-shell computers to build computation farms. The rapid growth of databases in the context of the semantic web requires efficient ways to store and process this data. Using cloud technology for storing and processing Semantic Web data is an obvious way to overcome difficulties in storing and processing the enormously large present and future datasets of the Semantic Web. This paper presents a new approach for storing Semantic Web data, such that operations for the evaluation of Semantic Web queries are more likely to be processed only on local data, instead of using costly distributed operations. An experimental evaluation demonstrates the performance improvements in comparison to a naive distribution of Semantic Web data.}
        }
    
  16.  Open Access 

    Detecting Data-Flow Errors in BPMN 2.0

    Silvia von Stackelberg, Susanne Putze, Jutta Mülle, Klemens Böhm

    Open Journal of Information Systems (OJIS), 1(2), Pages 1-19, 2014, Downloads: 2734, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017052611934 | GNL-LP: 1132836972 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Data-flow errors in BPMN 2.0 process models, such as missing or unused data, lead to undesired process executions. In particular, since BPMN 2.0 with a standardized execution semantics allows specifying alternatives for data as well as optional data, identifying missing or unused data systematically is difficult. In this paper, we propose an approach for detecting data-flow errors in BPMN 2.0 process models. We formalize BPMN process models by mapping them to Petri Nets and unfolding the execution semantics regarding data. We define a set of anti-patterns representing data-flow errors of BPMN 2.0 process models. By employing the anti-patterns, our tool performs model checking for the unfolded Petri Nets. The evaluation shows that it detects all data-flow errors identified by hand, and so improves process quality.

    BibTex:

        @Article{OJIS-2014v1i2n01_Stackelberg,
            title     = {Detecting Data-Flow Errors in BPMN 2.0},
            author    = {Silvia von Stackelberg and
                         Susanne Putze and
                         Jutta M{\"u}lle and
                         Klemens B{\"o}hm},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--19},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017052611934},
            urn       = {urn:nbn:de:101:1-2017052611934},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Data-flow errors in BPMN 2.0 process models, such as missing or unused data, lead to undesired process executions. In particular, since BPMN 2.0 with a standardized execution semantics allows specifying alternatives for data as well as optional data, identifying missing or unused data systematically is difficult. In this paper, we propose an approach for detecting data-flow errors in BPMN 2.0 process models. We formalize BPMN process models by mapping them to Petri Nets and unfolding the execution semantics regarding data. We define a set of anti-patterns representing data-flow errors of BPMN 2.0 process models. By employing the anti-patterns, our tool performs model checking for the unfolded Petri Nets. The evaluation shows that it detects all data-flow errors identified by hand, and so improves process quality.}
        }
    
  17.  Open Access 

    Pattern-sensitive Time-series Anonymization and its Application to Energy-Consumption Data

    Stephan Kessler, Erik Buchmann, Thorben Burghardt, Klemens Böhm

    Open Journal of Information Systems (OJIS), 1(1), Pages 3-22, 2014, Downloads: 2648, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194696 | GNL-LP: 113236096X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Time series anonymization is an important problem. One prominent example of time series are energy consumption records, which might reveal details of the daily routine of a household. Existing privacy approaches for time series, e.g., from the field of trajectory anonymization, assume that every single value of a time series contains sensitive information and reduce the data quality very much. In contrast, we consider time series where it is combinations of tuples that represent personal information. We propose (n; l; k)-anonymity, geared to anonymization of time-series data with minimal information loss, assuming that an adversary may learn a few data points. We propose several heuristics to obtain (n; l; k)-anonymity, and we evaluate our approach both with synthetic and real data. Our experiments confirm that it is sufficient to modify time series only moderately in order to fulfill meaningful privacy requirements.

    BibTex:

        @Article{OJIS-v1i1n02_Kessler,
            title     = {Pattern-sensitive Time-series Anonymization and its Application to Energy-Consumption Data},
            author    = {Stephan Kessler and
                         Erik Buchmann and
                         Thorben Burghardt and
                         Klemens B{\"o}hm},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {3--22},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194696},
            urn       = {urn:nbn:de:101:1-201705194696},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Time series anonymization is an important problem. One prominent example of time series are energy consumption records, which might reveal details of the daily routine of a household. Existing privacy approaches for time series, e.g., from the field of trajectory anonymization, assume that every single value of a time series contains sensitive information and reduce the data quality very much. In contrast, we consider time series where it is combinations of tuples that represent personal information. We propose (n; l; k)-anonymity, geared to anonymization of time-series data with minimal information loss, assuming that an adversary may learn a few data points. We propose several heuristics to obtain (n; l; k)-anonymity, and we evaluate our approach both with synthetic and real data. Our experiments confirm that it is sufficient to modify time series only moderately in order to fulfill meaningful privacy requirements.}
        }
    
  18.  Open Access 

    P-LUPOSDATE: Using Precomputed Bloom Filters to Speed Up SPARQL Processing in the Cloud

    Sven Groppe, Thomas Kiencke, Stefan Werner, Dennis Heinrich, Marc Stelzner, Le Gruenwald

    Open Journal of Semantic Web (OJSW), 1(2), Pages 25-55, 2014, Downloads: 2632, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194858 | GNL-LP: 1132361214 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: Increasingly data on the Web is stored in the form of Semantic Web data. Because of today's information overload, it becomes very important to store and query these big datasets in a scalable way and hence in a distributed fashion. Cloud Computing offers such a distributed environment with dynamic reallocation of computing and storing resources based on needs. In this work we introduce a scalable distributed Semantic Web database in the Cloud. In order to reduce the number of (unnecessary) intermediate results early, we apply bloom filters. Instead of computing bloom filters, a time-consuming task during query processing as it has been done traditionally, we precompute the bloom filters as much as possible and store them in the indices besides the data. The experimental results with data sets up to 1 billion triples show that our approach speeds up query processing significantly and sometimes even reduces the processing time to less than half.

    BibTex:

        @Article{OJSW-v1i2n02_Groppe,
            title     = {P-LUPOSDATE: Using Precomputed Bloom Filters to Speed Up SPARQL Processing in the Cloud},
            author    = {Sven Groppe and
                         Thomas Kiencke and
                         Stefan Werner and
                         Dennis Heinrich and
                         Marc Stelzner and
                         Le Gruenwald},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {25--55},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194858},
            urn       = {urn:nbn:de:101:1-201705194858},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Increasingly data on the Web is stored in the form of Semantic Web data. Because of today's information overload, it becomes very important to store and query these big datasets in a scalable way and hence in a distributed fashion. Cloud Computing offers such a distributed environment with dynamic reallocation of computing and storing resources based on needs. In this work we introduce a scalable distributed Semantic Web database in the Cloud. In order to reduce the number of (unnecessary) intermediate results early, we apply bloom filters. Instead of computing bloom filters, a time-consuming task during query processing as it has been done traditionally, we precompute the bloom filters as much as possible and store them in the indices besides the data. The experimental results with data sets up to 1 billion triples show that our approach speeds up query processing significantly and sometimes even reduces the processing time to less than half.}
        }
    
  19.  Open Access 

    MapReduce-based Solutions for Scalable SPARQL Querying

    José M. Giménez-Garcia, Javier D. Fernández, Miguel A. Martínez-Prieto

    Open Journal of Semantic Web (OJSW), 1(1), Pages 1-18, 2014, Downloads: 2616, Citations: 6

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194824 | GNL-LP: 1132361168 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The use of RDF to expose semantic data on the Web has seen a dramatic increase over the last few years. Nowadays, RDF datasets are so big and rconnected that, in fact, classical mono-node solutions present significant scalability problems when trying to manage big semantic data. MapReduce, a standard framework for distributed processing of great quantities of data, is earning a place among the distributed solutions facing RDF scalability issues. In this article, we survey the most important works addressing RDF management and querying through diverse MapReduce approaches, with a focus on their main strategies, optimizations and results.

    BibTex:

        @Article{OJSW-v1i1n02_Garcia,
            title     = {MapReduce-based Solutions for Scalable SPARQL Querying},
            author    = {Jos\'{e} M. Gim\'{e}nez-Garcia and
                         Javier D. Fern\'{a}ndez and
                         Miguel A. Mart\'{i}nez-Prieto},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194824},
            urn       = {urn:nbn:de:101:1-201705194824},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The use of RDF to expose semantic data on the Web has seen a dramatic increase over the last few years. Nowadays, RDF datasets are so big and rconnected that, in fact, classical mono-node solutions present significant scalability problems when trying to manage big semantic data. MapReduce, a standard framework for distributed processing of great quantities of data, is earning a place among the distributed solutions facing RDF scalability issues. In this article, we survey the most important works addressing RDF management and querying through diverse MapReduce approaches, with a focus on their main strategies, optimizations and results.}
        }
    
  20.  Open Access 

    Modelling the Integrated QoS for Wireless Sensor Networks with Heterogeneous Data Traffic

    Syarifah Ezdiani, Adnan Al-Anbuky

    Open Journal of Internet Of Things (OJIOT), 1(1), Pages 1-15, 2015, Downloads: 2562, Citations: 10

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244946 | GNL-LP: 1130621979 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The future of Internet of Things (IoT) is envisaged to consist of a high amount of wireless resource-constrained devices connected to the Internet. Moreover, a lot of novel real-world services offered by IoT devices are realized by wireless sensor networks (WSNs). Integrating WSN to the Internet has therefore brought forward the requirements of an end-to-end quality of service (QoS) guarantee. In this paper, the QoS requirements for the WSN-Internet integration are investigated by first distinguishing the Internet QoS from the WSN QoS. Next, this study emphasizes on WSN applications that involve traffic with different levels of importance, thus the way realtime traffic and delay-tolerant traffic are handled to guarantee QoS in the network is studied. Additionally, an overview of the integration strategies is given, and the delay-tolerant network (DTN) gateway, being one of the desirable approaches for integrating WSNs to the Internet, is discussed. Next, the implementation of the service model is presented, by considering both traffic prioritization and service differentiation. Based on the simulation results in OPNET Modeler, it is observed that real-time traffic achieve low bound delay while delay-tolerant traffic experience a lower packet dropped, hence indicating that the needs of real-time and delay-tolerant traffic can be better met by treating both packet types differently. Furthermore, a vehicular network is used as an example case to describe the applicability of the framework in a real IoT application environment, followed by a discussion on the future work of this research.

    BibTex:

        @Article{OJIOT_2015v1i1n02_Syarifah,
            title     = {Modelling the Integrated QoS for Wireless Sensor Networks with Heterogeneous Data Traffic},
            author    = {Syarifah Ezdiani and
                         Adnan Al-Anbuky},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {1--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244946},
            urn       = {urn:nbn:de:101:1-201704244946},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The future of Internet of Things (IoT) is envisaged to consist of a high amount of wireless resource-constrained devices connected to the Internet. Moreover, a lot of novel real-world services offered by IoT devices are realized by wireless sensor networks (WSNs). Integrating WSN to the Internet has therefore brought forward the requirements of an end-to-end quality of service (QoS) guarantee. In this paper, the QoS requirements for the WSN-Internet integration are investigated by first distinguishing the Internet QoS from the WSN QoS. Next, this study emphasizes on WSN applications that involve traffic with different levels of importance, thus the way realtime traffic and delay-tolerant traffic are handled to guarantee QoS in the network is studied. Additionally, an overview of the integration strategies is given, and the delay-tolerant network (DTN) gateway, being one of the desirable approaches for integrating WSNs to the Internet, is discussed. Next, the implementation of the service model is presented, by considering both traffic prioritization and service differentiation. Based on the simulation results in OPNET Modeler, it is observed that real-time traffic achieve low bound delay while delay-tolerant traffic experience a lower packet dropped, hence indicating that the needs of real-time and delay-tolerant traffic can be better met by treating both packet types differently. Furthermore, a vehicular network is used as an example case to describe the applicability of the framework in a real IoT application environment, followed by a discussion on the future work of this research.}
        }
    
  21.  Open Access 

    Perceived Sociability of Use and Individual Use of Social Networking Sites - A Field Study of Facebook Use in the Arctic

    Juhani Iivari

    Open Journal of Information Systems (OJIS), 1(1), Pages 23-53, 2014, Downloads: 2559, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194708 | GNL-LP: 1132360978 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper investigates determinants of individual use of social network sites (SNSs). It introduces a new construct, Perceived Sociability of Use (PSOU), to explain the use of such computer mediated communication applications. Based on a field study of 113 Facebook users it shows that PSOU in the sense of maintaining social contacts is a significant predictor of Perceived Benefits (PB), Perceived Enjoyment (PE), attitude toward use and intention to use. Inspired by Benbasat and Barki, this paper also attempts to answer questions "what makes the system useful", "what makes the system enjoyable to use" and "what makes the system sociable to use". As a consequence it pays special focus on systems characteristics of IT applications as potential predictors of PSOU, PB and PE, introducing seven such designable qualities (user-to-user interactivity, user identifiability, system quality, information quality, usability, user-to-system interactivity, and aesthetics). The results indicate that especially satisfaction with user-to-user interactivity is a significant determinant of PSOU, and that satisfactions with six of these seven designable qualities have significant paths in the proposed nomological network.

    BibTex:

        @Article{OJIS-v1i1n03_Iivari,
            title     = {Perceived Sociability of Use and Individual Use of Social Networking Sites - A Field Study of Facebook Use in the Arctic},
            author    = {Juhani Iivari},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {23--53},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194708},
            urn       = {urn:nbn:de:101:1-201705194708},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper investigates determinants of individual use of social network sites (SNSs). It introduces a new construct, Perceived Sociability of Use (PSOU), to explain the use of such computer mediated communication applications. Based on a field study of 113 Facebook users it shows that PSOU in the sense of maintaining social contacts is a significant predictor of Perceived Benefits (PB), Perceived Enjoyment (PE), attitude toward use and intention to use. Inspired by Benbasat and Barki, this paper also attempts to answer questions "what makes the system useful", "what makes the system enjoyable to use" and "what makes the system sociable to use". As a consequence it pays special focus on systems characteristics of IT applications as potential predictors of PSOU, PB and PE, introducing seven such designable qualities (user-to-user interactivity, user identifiability, system quality, information quality, usability, user-to-system interactivity, and aesthetics). The results indicate that especially satisfaction with user-to-user interactivity is a significant determinant of PSOU, and that satisfactions with six of these seven designable qualities have significant paths in the proposed nomological network.}
        }
    
  22.  Open Access 

    Data Transfers in Hadoop: A Comparative Study

    Ujjal Marjit, Kumar Sharma, Puspendu Mandal

    Open Journal of Big Data (OJBD), 1(2), Pages 34-46, 2015, Downloads: 2528, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194373 | GNL-LP: 1132360536 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Hadoop is an open source framework for processing large amounts of data in distributed computing environment. It plays an important role in processing and analyzing the Big Data. This framework is used for storing data on large clusters of commodity hardware. Data input and output to and from Hadoop is an indispensable action for any data processing job. At present, many tools have been evolved for importing and exporting Data in Hadoop. In this article, some commonly used tools for importing and exporting data have been emphasized. Moreover, a state-of-the-art comparative study among the various tools has been made. With this study, it has been decided that where to use one tool over the other with emphasis on the data transfer to and from Hadoop system. This article also discusses about how Hadoop handles backup and disaster recovery along with some open research questions in terms of Big Data transfer when dealing with cloud-based services.

    BibTex:

        @Article{OJBD_2015v1i2n04_UjjalMarjit,
            title     = {Data Transfers in Hadoop: A Comparative Study},
            author    = {Ujjal Marjit and
                         Kumar Sharma and
                         Puspendu Mandal},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {34--46},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194373},
            urn       = {urn:nbn:de:101:1-201705194373},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Hadoop is an open source framework for processing large amounts of data in distributed computing environment. It plays an important role in processing and analyzing the Big Data. This framework is used for storing data on large clusters of commodity hardware. Data input and output to and from Hadoop is an indispensable action for any data processing job. At present, many tools have been evolved for importing and exporting Data in Hadoop. In this article, some commonly used tools for importing and exporting data have been emphasized. Moreover, a state-of-the-art comparative study among the various tools has been made. With this study, it has been decided that where to use one tool over the other with emphasis on the data transfer to and from Hadoop system. This article also discusses about how Hadoop handles backup and disaster recovery along with some open research questions in terms of Big Data transfer when dealing with cloud-based services.}
        }
    
  23.  Open Access 

    Using Nuisance Telephone Denial of Service to Combat Online Sex Trafficking

    Ross A. Malaga

    Open Journal of Information Systems (OJIS), 2(1), Pages 1-8, 2015, Downloads: 2414, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194736 | GNL-LP: 1132361036 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Over the past few years, sex trafficking has been linked to online classified ads sites such as Craigslist.com and Backpage.com. However, to date technology-based solutions have not been used to attack classified ad sites or the advertisers. This paper proposes and tests a new approach to combating online sex trafficking promulgated via online classified ad sites - nuisance telephone denial of service (TDoS) attacks on the advertisers. The method of attack is described and implications are discussed.

    BibTex:

        @Article{OJIS_2015v2i1n01_Malaga,
            title     = {Using Nuisance Telephone Denial of Service to Combat Online Sex Trafficking},
            author    = {Ross A. Malaga},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {1--8},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194736},
            urn       = {urn:nbn:de:101:1-201705194736},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Over the past few years, sex trafficking has been linked to online classified ads sites such as Craigslist.com and Backpage.com. However, to date technology-based solutions have not been used to attack classified ad sites or the advertisers. This paper proposes and tests a new approach to combating online sex trafficking promulgated via online classified ad sites - nuisance telephone denial of service (TDoS) attacks on the advertisers. The method of attack is described and implications are discussed.}
        }
    
  24.  Open Access 

    An Efficient Approach for Cost Optimization of the Movement of Big Data

    Prasad Teli, Manoj V. Thomas, K. Chandrasekaran

    Open Journal of Big Data (OJBD), 1(1), Pages 4-15, 2015, Downloads: 2390, Citations: 4

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194335 | GNL-LP: 113236048X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: With the emergence of cloud computing, Big Data has caught the attention of many researchers in the area of cloud computing. As the Volume, Velocity and Variety (3 Vs) of big data are growing exponentially, dealing with them is a big challenge, especially in the cloud environment. Looking at the current trend of the IT sector, cloud computing is mainly used by the service providers to host their applications. A lot of research has been done to improve the network utilization of WAN (Wide Area Network) and it has achieved considerable success over the traditional LAN (Local Area Network) techniques. While dealing with this issue, the major questions of data movement such as from where to where this big data will be moved and also how the data will be moved, have been overlooked. As various applications generating the big data are hosted in geographically distributed data centers, they individually collect large volume of data in the form of application data as well as the logs. This paper mainly focuses on the challenge of moving big data from one data center to other. We provide an efficient algorithm for the optimization of cost in the movement of the big data from one data center to another for offline environment. This approach uses the graph model for data centers in the cloud and results show that the adopted mechanism provides a better solution to minimize the cost for data movement.

    BibTex:

        @Article{OJBD_2015v1i1n02_Teli,
            title     = {An Efficient Approach for Cost Optimization of the Movement of Big Data},
            author    = {Prasad Teli and
                         Manoj V. Thomas and
                         K. Chandrasekaran},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194335},
            urn       = {urn:nbn:de:101:1-201705194335},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {With the emergence of cloud computing, Big Data has caught the attention of many researchers in the area of cloud computing. As the Volume, Velocity and Variety (3 Vs) of big data are growing exponentially, dealing with them is a big challenge, especially in the cloud environment. Looking at the current trend of the IT sector, cloud computing is mainly used by the service providers to host their applications. A lot of research has been done to improve the network utilization of WAN (Wide Area Network) and it has achieved considerable success over the traditional LAN (Local Area Network) techniques. While dealing with this issue, the major questions of data movement such as from where to where this big data will be moved and also how the data will be moved, have been overlooked. As various applications generating the big data are hosted in geographically distributed data centers, they individually collect large volume of data in the form of application data as well as the logs. This paper mainly focuses on the challenge of moving big data from one data center to other. We provide an efficient algorithm for the optimization of cost in the movement of the big data from one data center to another for offline environment. This approach uses the graph model for data centers in the cloud and results show that the adopted mechanism provides a better solution to minimize the cost for data movement.}
        }
    
  25.  Open Access 

    BioSStore: A Client Interface for a Repository of Semantically Annotated Bioinformatics Web Services

    Ismael Navas-Delgado, José F. Aldana-Montes

    Open Journal of Semantic Web (OJSW), 1(1), Pages 19-29, 2014, Downloads: 2319

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194836 | GNL-LP: 1132361176 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Bioinformatics has shown itself to be a domain in which Web services are being used extensively. In this domain, simple but real services are being developed. Thus, there are huge repositories of real services available (for example BioMOBY main repository includes more than 1500 services). Besides, bioinformatics repositories usually have active communities using and working on improvements. However, these kinds of repositories do not exploit the full potential of Web services (and SOA, Service Oriented Applications, in general). On the other hand, sophisticated technologies have been proposed to improve SOA, including the annotation on Web services to explicitly describe them. However, these approaches are lacking in repositories with real services. In the work presented here, we address the drawbacks present in bioinformatics services and try to improve the current semantic model by introducing the use of the W3C standard Semantic Annotations for WSDL and XML Schema (SAWSDL) and related proposals (WSMO Lite). This paper focuses on a user interface that takes advantage of a repository of semantically annotated bioinformatics Web services. In this way, we exploit semantics for the discovery of Web services, showing how the use of semantics will improve the user searches. The BioSStore is available at http://biosstore.khaos.uma.es. This portal will contain also future developments of this proposal.

    BibTex:

        @Article{OJSW-v1i1n03_Delgado,
            title     = {BioSStore: A Client Interface for a Repository of Semantically Annotated Bioinformatics Web Services},
            author    = {Ismael Navas-Delgado and
                         Jos\'{e} F. Aldana-Montes},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {19--29},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194836},
            urn       = {urn:nbn:de:101:1-201705194836},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Bioinformatics has shown itself to be a domain in which Web services are being used extensively. In this domain, simple but real services are being developed. Thus, there are huge repositories of real services available (for example BioMOBY main repository includes more than 1500 services). Besides, bioinformatics repositories usually have active communities using and working on improvements. However, these kinds of repositories do not exploit the full potential of Web services (and SOA, Service Oriented Applications, in general). On the other hand, sophisticated technologies have been proposed to improve SOA, including the annotation on Web services to explicitly describe them. However, these approaches are lacking in repositories with real services. In the work presented here, we address the drawbacks present in bioinformatics services and try to improve the current semantic model by introducing the use of the W3C standard Semantic Annotations for WSDL and XML Schema (SAWSDL) and related proposals (WSMO Lite). This paper focuses on a user interface that takes advantage of a repository of semantically annotated bioinformatics Web services. In this way, we exploit semantics for the discovery of Web services, showing how the use of semantics will improve the user searches. The BioSStore is available at http://biosstore.khaos.uma.es. This portal will contain also future developments of this proposal.}
        }
    
  26.  Open Access 

    Deriving Bounds on the Size of Spatial Areas

    Erik Buchmann, Patrick Erik Bradley, Klemens Böhm

    Open Journal of Databases (OJDB), 2(1), Pages 1-16, 2015, Downloads: 2237

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194566 | GNL-LP: 113236082X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Many application domains such as surveillance, environmental monitoring or sensor-data processing need upper and lower bounds on areas that are covered by a certain feature. For example, a smart-city infrastructure might need bounds on the size of an area polluted with fine-dust, to re-route combustion-engine traffic. Obtaining such bounds is challenging, because in almost any real-world application, information about the region of interest is incomplete, e.g., the database of sensor data contains only a limited number of samples. Existing approaches cannot provide upper and lower bounds or depend on restrictive assumptions, e.g., the area must be convex. Our approach in turn is based on the natural assumption that it is possible to specify a minimal diameter for the feature in question. Given this assumption, we formally derive bounds on the area size, and we provide algorithms that compute these bounds from a database of sensor data, based on geometrical considerations. We evaluate our algorithms both with a real-world case study and with synthetic data.

    BibTex:

        @Article{OJDB-2015v2i1n01_Buchmann,
            title     = {Deriving Bounds on the Size of Spatial Areas},
            author    = {Erik Buchmann and
                         Patrick Erik Bradley and
                         Klemens B{\"o}hm},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {1--16},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194566},
            urn       = {urn:nbn:de:101:1-201705194566},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Many application domains such as surveillance, environmental monitoring or sensor-data processing need upper and lower bounds on areas that are covered by a certain feature. For example, a smart-city infrastructure might need bounds on the size of an area polluted with fine-dust, to re-route combustion-engine traffic. Obtaining such bounds is challenging, because in almost any real-world application, information about the region of interest is incomplete, e.g., the database of sensor data contains only a limited number of samples. Existing approaches cannot provide upper and lower bounds or depend on restrictive assumptions, e.g., the area must be convex. Our approach in turn is based on the natural assumption that it is possible to specify a minimal diameter for the feature in question. Given this assumption, we formally derive bounds on the area size, and we provide algorithms that compute these bounds from a database of sensor data, based on geometrical considerations. We evaluate our algorithms both with a real-world case study and with synthetic data.}
        }
    
  27.  Open Access 

    Evaluation of Node Failures in Cloud Computing Using Empirical Data

    Abdulelah Alwabel, Robert John Walters, Gary B. Wills

    Open Journal of Cloud Computing (OJCC), 1(2), Pages 15-24, 2014, Downloads: 2172, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194435 | GNL-LP: 1132360609 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Cloud has emerged as a new computing paradigm that promises to move into computing-as-utility era. Desktop Cloud is a new type of Cloud computing introduced to further achieve this ambition with an aim to reduce costs. It merges two computing models: Cloud computing and volunteer computing. The aim of Desktop Cloud is to provide Cloud services out of infrastructure that is not made for this purpose, like PCs and laptops. Such computing resources lead to a high level of volatility as a result of the fact that they can leave without prior knowledge. This paper studies the impact of node failures using evaluation metrics based on real data collected from public archive to simulate failure events in the infrastructure of a Desktop Cloud. The contribution of this paper is: (i) analysing the failure events, (ii) proposing metrics to evaluate Desktop Clouds, and (iii) evaluating several VM allocation mechanisms in the presence of node failures.

    BibTex:

        @Article{OJCC-2014v1i2n02_Alwabel,
            title     = {Evaluation of Node Failures in Cloud Computing Using Empirical Data},
            author    = {Abdulelah Alwabel and
                         Robert John Walters and
                         Gary B. Wills},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {15--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194435},
            urn       = {urn:nbn:de:101:1-201705194435},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Cloud has emerged as a new computing paradigm that promises to move into computing-as-utility era. Desktop Cloud is a new type of Cloud computing introduced to further achieve this ambition with an aim to reduce costs. It merges two computing models: Cloud computing and volunteer computing. The aim of Desktop Cloud is to provide Cloud services out of infrastructure that is not made for this purpose, like PCs and laptops. Such computing resources lead to a high level of volatility as a result of the fact that they can leave without prior knowledge. This paper studies the impact of node failures using evaluation metrics based on real data collected from public archive to simulate failure events in the infrastructure of a Desktop Cloud. The contribution of this paper is: (i) analysing the failure events, (ii) proposing metrics to evaluate Desktop Clouds, and (iii) evaluating several VM allocation mechanisms in the presence of node failures.}
        }
    
  28.  Open Access 

    IT Governance Practices for Electric Utilities: Insights from Brazil and Europe

    Paulo Rupino da Cunha, Luiz Mauricio Martins, Antão Moura, António Dias de Figueiredo

    Open Journal of Information Systems (OJIS), 2(1), Pages 9-28, 2015, Downloads: 2152, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194743 | GNL-LP: 1132361044 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: We propose a framework of 14 IT governance practices tailored for the electric utilities sector. They were selected and ranked as "essential", "important", or "good" by top executives and IT staff from two multi-billion dollar companies - one in Brazil and another in Europe - from a generic set of 83 collected in the literature and in the field. Our framework addresses a need of electric utilities for which specific guidance was lacking. We have also uncovered a significant impact of social issues in IT governance, whose depth seems to be missing in the current research. As a byproduct of our work, the larger generic framework from which we have departed and the tailoring method that we have proposed can be used to customize the generic framework to different industries.

    BibTex:

        @Article{OJIS_2015v2i1n02_Cunha,
            title     = {IT Governance Practices for Electric Utilities: Insights from Brazil and Europe},
            author    = {Paulo Rupino da Cunha and
                         Luiz Mauricio Martins and
                         Ant\~{a}o Moura and
                         Ant\'{o}nio Dias de Figueiredo},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {9--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194743},
            urn       = {urn:nbn:de:101:1-201705194743},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {We propose a framework of 14 IT governance practices tailored for the electric utilities sector. They were selected and ranked as "essential", "important", or "good" by top executives and IT staff from two multi-billion dollar companies - one in Brazil and another in Europe - from a generic set of 83 collected in the literature and in the field. Our framework addresses a need of electric utilities for which specific guidance was lacking. We have also uncovered a significant impact of social issues in IT governance, whose depth seems to be missing in the current research. As a byproduct of our work, the larger generic framework from which we have departed and the tailoring method that we have proposed can be used to customize the generic framework to different industries.}
        }
    
  29.  Open Access 

    Definition and Categorization of Dew Computing

    Yingwei Wang

    Open Journal of Cloud Computing (OJCC), 3(1), Pages 1-7, 2016, Downloads: 2149, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194546 | GNL-LP: 1132360781 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Dew computing is an emerging new research area and has great potentials in applications. In this paper, we propose a revised definition of dew computing. The new definition is: Dew computing is an on-premises computer software-hardware organization paradigm in the cloud computing environment where the on-premises computer provides functionality that is independent of cloud services and is also collaborative with cloud services. The goal of dew computing is to fully realize the potentials of on-premises computers and cloud services. This definition emphasizes two key features of dew computing: independence and collaboration. Furthermore, we propose a group of dew computing categories. These categories may inspire new applications.

    BibTex:

        @Article{OJCC_2016v3i1n02_YingweiWang,
            title     = {Definition and Categorization of Dew Computing},
            author    = {Yingwei Wang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {1--7},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194546},
            urn       = {urn:nbn:de:101:1-201705194546},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Dew computing is an emerging new research area and has great potentials in applications. In this paper, we propose a revised definition of dew computing. The new definition is: Dew computing is an on-premises computer software-hardware organization paradigm in the cloud computing environment where the on-premises computer provides functionality that is independent of cloud services and is also collaborative with cloud services. The goal of dew computing is to fully realize the potentials of on-premises computers and cloud services. This definition emphasizes two key features of dew computing: independence and collaboration. Furthermore, we propose a group of dew computing categories. These categories may inspire new applications.}
        }
    
  30.  Open Access 

    Runtime Adaptive Hybrid Query Engine based on FPGAs

    Stefan Werner, Dennis Heinrich, Sven Groppe, Christopher Blochwitz, Thilo Pionteck

    Open Journal of Databases (OJDB), 3(1), Pages 21-41, 2016, Downloads: 2137, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194645 | GNL-LP: 1132360900 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper presents the fully integrated hardware-accelerated query engine for large-scale datasets in the context of Semantic Web databases. As queries are typically unknown at design time, a static approach is not feasible and not flexible to cover a wide range of queries at system runtime. Therefore, we introduce a runtime reconfigurable accelerator based on a Field Programmable Gate Array (FPGA), which transparently incorporates with the freely available Semantic Web database LUPOSDATE. At system runtime, the proposed approach dynamically generates an optimized hardware accelerator in terms of an FPGA configuration for each individual query and transparently retrieves the query result to be displayed to the user. During hardware-accelerated execution the host supplies triple data to the FPGA and retrieves the results from the FPGA via PCIe interface. The benefits and limitations are evaluated on large-scale synthetic datasets with up to 260 million triples as well as the widely known Billion Triples Challenge.

    BibTex:

        @Article{OJDB_2016v3i1n02_Werner,
            title     = {Runtime Adaptive Hybrid Query Engine based on FPGAs},
            author    = {Stefan Werner and
                         Dennis Heinrich and
                         Sven Groppe and
                         Christopher Blochwitz and
                         Thilo Pionteck},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {21--41},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194645},
            urn       = {urn:nbn:de:101:1-201705194645},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper presents the fully integrated hardware-accelerated query engine for large-scale datasets in the context of Semantic Web databases. As queries are typically unknown at design time, a static approach is not feasible and not flexible to cover a wide range of queries at system runtime. Therefore, we introduce a runtime reconfigurable accelerator based on a Field Programmable Gate Array (FPGA), which transparently incorporates with the freely available Semantic Web database LUPOSDATE. At system runtime, the proposed approach dynamically generates an optimized hardware accelerator in terms of an FPGA configuration for each individual query and transparently retrieves the query result to be displayed to the user. During hardware-accelerated execution the host supplies triple data to the FPGA and retrieves the results from the FPGA via PCIe interface. The benefits and limitations are evaluated on large-scale synthetic datasets with up to 260 million triples as well as the widely known Billion Triples Challenge.	}
        }
    
  31.  Open Access 

    A 24 GHz FM-CW Radar System for Detecting Closed Multiple Targets and Its Applications in Actual Scenes

    Kazuhiro Yamaguchi, Mitumasa Saito, Takuya Akiyama, Tomohiro Kobayashi, Naoki Ginoza, Hideaki Matsue

    Open Journal of Internet Of Things (OJIOT), 2(1), Pages 1-15, 2016, Downloads: 2121

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704245003 | GNL-LP: 1130623858 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper develops a 24 GHz band FM-CW radar system to detect closed multiple targets in a small displacement environment, and its performance is analyzed by computer simulation. The FM-CW radar system uses a differential detection method for removing any signals from background objects and uses a tunable FIR filtering in signal processing for detecting multiple targets. The differential detection method enables the correct detection of both the distance and small displacement at the same time for each target at the FM-CW radar according to the received signals. The basic performance of the FM-CW radar system is analyzed by computer simulation, and the distance and small displacement of a single target are measured in field experiments. The computer simulations are carried out for evaluating the proposed detection method with tunable FIR filtering for the FM-CW radar and for analyzing the performance according to the parameters in a closed multiple targets environment. The results of simulation show that our 24 GHz band FM-CW radar with the proposed detection method can effectively detect both the distance and the small displacement for each target in multiple moving targets environments. Moreover, we develop an IoT-based application for monitoring several targets at the same time in actual scenes.

    BibTex:

        @Article{OJIOT_2016v2i1n02_Yamaguchi,
            title     = {A 24 GHz FM-CW Radar System for Detecting Closed Multiple Targets and Its Applications in Actual Scenes},
            author    = {Kazuhiro Yamaguchi and
                         Mitumasa Saito and
                         Takuya Akiyama and
                         Tomohiro Kobayashi and
                         Naoki Ginoza and
                         Hideaki Matsue},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2016},
            volume    = {2},
            number    = {1},
            pages     = {1--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704245003},
            urn       = {urn:nbn:de:101:1-201704245003},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper develops a 24 GHz band FM-CW radar system to detect closed multiple targets in a small displacement environment, and its performance is analyzed by computer simulation. The FM-CW radar system uses a differential detection method for removing any signals from background objects and uses a tunable FIR filtering in signal processing for detecting multiple targets. The differential detection method enables the correct detection of both the distance and small displacement at the same time for each target at the FM-CW radar according to the received signals. The basic performance of the FM-CW radar system is analyzed by computer simulation, and the distance and small displacement of a single target are measured in field experiments. The computer simulations are carried out for evaluating the proposed detection method with tunable FIR filtering for the FM-CW radar and for analyzing the performance according to the parameters in a closed multiple targets environment. The results of simulation show that our 24 GHz band FM-CW radar with the proposed detection method can effectively detect both the distance and the small displacement for each target in multiple moving targets environments. Moreover, we develop an IoT-based application for monitoring several targets at the same time in actual scenes.}
        }
    
  32.  Open Access 

    Doing More with the Dew: A New Approach to Cloud-Dew Architecture

    David Edward Fisher, Shuhui Yang

    Open Journal of Cloud Computing (OJCC), 3(1), Pages 8-19, 2016, Downloads: 2114

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194535 | GNL-LP: 1132360773 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: While the popularity of cloud computing is exploding, a new network computing paradigm is just beginning. In this paper, we examine this exciting area of research known as dew computing and propose a new design of cloud-dew architecture. Instead of hosting only one dew server on a user's PC - as adopted in the current dewsite application - our design promotes the hosting of multiple dew servers instead, one for each installed domain. Our design intends to improve upon existing cloud-dew architecture by providing significantly increased freedom in dewsite development, while also automating the chore of managing dewsite content based on the user's interests and browsing habits. Other noteworthy benefits, all at no added cost to dewsite users, are briefly explored as well.

    BibTex:

        @Article{OJCC_2016v3i1n02_Fisher,
            title     = {Doing More with the Dew: A New Approach to Cloud-Dew Architecture},
            author    = {David Edward Fisher and
                         Shuhui Yang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {8--19},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194535},
            urn       = {urn:nbn:de:101:1-201705194535},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {While the popularity of cloud computing is exploding, a new network computing paradigm is just beginning. In this paper, we examine this exciting area of research known as dew computing and propose a new design of cloud-dew architecture. Instead of hosting only one dew server on a user's PC - as adopted in the current dewsite application - our design promotes the hosting of multiple dew servers instead, one for each installed domain. Our design intends to improve upon existing cloud-dew architecture by providing significantly increased freedom in dewsite development, while also automating the chore of managing dewsite content based on the user's interests and browsing habits. Other noteworthy benefits, all at no added cost to dewsite users, are briefly explored as well.}
        }
    
  33.  Open Access 

    A Semantic Question Answering Framework for Large Data Sets

    Marta Tatu, Mithun Balakrishna, Steven Werner, Tatiana Erekhinskaya, Dan Moldovan

    Open Journal of Semantic Web (OJSW), 3(1), Pages 16-31, 2016, Downloads: 2099, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194921 | GNL-LP: 1132361338 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Traditionally, the task of answering natural language questions has involved a keyword-based document retrieval step, followed by in-depth processing of candidate answer documents and paragraphs. This post-processing uses semantics to various degrees. In this article, we describe a purely semantic question answering (QA) framework for large document collections. Our high-precision approach transforms the semantic knowledge extracted from natural language texts into a language-agnostic RDF representation and indexes it into a scalable triplestore. In order to facilitate easy access to the information stored in the RDF semantic index, a user's natural language questions are translated into SPARQL queries that return precise answers back to the user. The robustness of this framework is ensured by the natural language reasoning performed on the RDF store, by the query relaxation procedures, and the answer ranking techniques. The improvements in performance over a regular free text search index-based question answering engine prove that QA systems can benefit greatly from the addition and consumption of deep semantic information.

    BibTex:

        @Article{OJSW_2016v3i1n02_Tatu,
            title     = {A Semantic Question Answering Framework for Large Data Sets},
            author    = {Marta Tatu and
                         Mithun Balakrishna and
                         Steven Werner and
                         Tatiana Erekhinskaya and
                         Dan Moldovan},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {16--31},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194921},
            urn       = {urn:nbn:de:101:1-201705194921},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Traditionally, the task of answering natural language questions has involved a keyword-based document retrieval step, followed by in-depth processing of candidate answer documents and paragraphs. This post-processing uses semantics to various degrees. In this article, we describe a purely semantic question answering (QA) framework for large document collections. Our high-precision approach transforms the semantic knowledge extracted from natural language texts into a language-agnostic RDF representation and indexes it into a scalable triplestore. In order to facilitate easy access to the information stored in the RDF semantic index, a user's natural language questions are translated into SPARQL queries that return precise answers back to the user. The robustness of this framework is ensured by the natural language reasoning performed on the RDF store, by the query relaxation procedures, and the answer ranking techniques. The improvements in performance over a regular free text search index-based question answering engine prove that QA systems can benefit greatly from the addition and consumption of deep semantic information.}
        }
    
  34.  Open Access 

    Designing a Benchmark for the Assessment of Schema Matching Tools

    Fabien Duchateau, Zohra Bellahsene

    Open Journal of Databases (OJDB), 1(1), Pages 3-25, 2014, Downloads: 2082, Citations: 7

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194573 | GNL-LP: 1132360838 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Over the years, many schema matching approaches have been developed to discover correspondences between schemas. Although this task is crucial in data integration, its evaluation, both in terms of matching quality and time performance, is still manually performed. Indeed, there is no common platform which gathers a collection of schema matching datasets to fulfil this goal. Another problem deals with the measuring of the post-match effort, a human cost that schema matching approaches aim at reducing. Consequently, we propose XBenchMatch, a schema matching benchmark with available datasets and new measures to evaluate this manual post-match effort and the quality of integrated schemas. We finally report the results obtained by different approaches, namely COMA++, Similarity Flooding and YAM. We show that such a benchmark is required to understand the advantages and failures of schema matching approaches. Therefore, it could help an end-user to select a schema matching tool which covers his/her needs.

    BibTex:

        @Article{OJDB-v1i1n02_Duchateau,
            title     = {Designing a Benchmark for the Assessment of Schema Matching Tools},
            author    = {Fabien Duchateau and
                         Zohra Bellahsene},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {3--25},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194573},
            urn       = {urn:nbn:de:101:1-201705194573},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Over the years, many schema matching approaches have been developed to discover correspondences between schemas. Although this task is crucial in data integration, its evaluation, both in terms of matching quality and time performance, is still manually performed. Indeed, there is no common platform which gathers a collection of schema matching datasets to fulfil this goal. Another problem deals with the measuring of the post-match effort, a human cost that schema matching approaches aim at reducing. Consequently, we propose XBenchMatch, a schema matching benchmark with available datasets and new measures to evaluate this manual post-match effort and the quality of integrated schemas. We finally report the results obtained by different approaches, namely COMA++, Similarity Flooding and YAM. We show that such a benchmark is required to understand the advantages and failures of schema matching approaches. Therefore, it could help an end-user to select a schema matching tool which covers his/her needs.}
        }
    
  35.  Open Access 

    Fuzzy Color Space for Apparel Coordination

    Pakizar Shamoi, Atsushi Inoue, Hiroharu Kawanaka

    Open Journal of Information Systems (OJIS), 1(2), Pages 20-28, 2014, Downloads: 2079, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194710 | GNL-LP: 1132360994 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Human perception of colors constitutes an important part in color theory. The applications of color science are truly omnipresent, and what impression colors make on human plays a vital role in them. In this paper, we offer the novel approach for color information representation and processing using fuzzy sets and logic theory, which is extremely useful in modeling human impressions. Specifically, we use fuzzy mathematics to partition the gamut of feasible colors in HSI color space based on standard linguistic tags. The proposed method can be useful in various image processing applications involving query processing. We demonstrate its effectivity in the implementation of a framework for the apparel online shopping coordination based on a color scheme. It deserves attention, since there is always some uncertainty inherent in the description of apparels.

    BibTex:

        @Article{OJIS_2014v1i2n02_Shamoi,
            title     = {Fuzzy Color Space for Apparel Coordination},
            author    = {Pakizar Shamoi and
                         Atsushi Inoue and
                         Hiroharu Kawanaka},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {20--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194710},
            urn       = {urn:nbn:de:101:1-201705194710},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Human perception of colors constitutes an important part in color theory. The applications of color science are truly omnipresent, and what impression colors make on human plays a vital role in them. In this paper, we offer the novel approach for color information representation and processing using fuzzy sets and logic theory, which is extremely useful in modeling human impressions. Specifically, we use fuzzy mathematics to partition the gamut of feasible colors in HSI color space based on standard linguistic tags. The proposed method can be useful in various image processing applications involving query processing. We demonstrate its effectivity in the implementation of a framework for the apparel online shopping coordination based on a color scheme. It deserves attention, since there is always some uncertainty inherent in the description of apparels.}
        }
    
  36.  Open Access 

    SIWeb: understanding the Interests of the Society through Web data Analysis

    Marco Furini, Simone Montangero

    Open Journal of Web Technologies (OJWT), 1(1), Pages 1-14, 2014, Downloads: 2059, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291334 | GNL-LP: 1133021522 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The high availability of user-generated contents in the Web scenario represents a tremendous asset for understanding various social phenomena. Methods and commercial products that exploit the widespread use of the Web as a way of conveying personal opinions have been proposed, but a critical thinking is that these approaches may produce a partial, or distorted, understanding of the society, because most of them focus on definite scenarios, use specific platforms, base their analysis on the sole magnitude of data, or treat the different Web resources with the same importance. In this paper, we present SIWeb (Social Interests through Web Analysis), a novel mechanism designed to measure the interest the society has on a topic (e.g., a real world phenomenon, an event, a person, a thing). SIWeb is general purpose (it can be applied to any decision making process), cross platforms (it uses the entire Webspace, from social media to websites, from tags to reviews), and time effective (it measures the time correlatio between the Web resources). It uses fractal analysis to detect the temporal relations behind all the Web resources (e.g., Web pages, RSS, newsgroups, etc.) that talk about a topic and combines this number with the temporal relations to give an insight of the the interest the society has about a topic. The evaluation of the proposal shows that SIWeb might be helpful in decision making processes as it reflects the interests the society has on a specific topic.

    BibTex:

        @Article{OJWT-v1i1n01_Furini,
            title     = {SIWeb: understanding the Interests of the Society through Web data Analysis},
            author    = {Marco Furini and
                         Simone Montangero},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--14},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291334},
            urn       = {urn:nbn:de:101:1-201705291334},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The high availability of user-generated contents in the Web scenario represents a tremendous asset for understanding various social phenomena. Methods and commercial products that exploit the widespread use of the Web as a way of conveying personal opinions have been proposed, but a critical thinking is that these approaches may produce a partial, or distorted, understanding of the society, because most of them focus on definite scenarios, use specific platforms, base their analysis on the sole magnitude of data, or treat the different Web resources with the same importance. In this paper, we present SIWeb (Social Interests through Web Analysis), a novel mechanism designed to measure the interest the society has on a topic (e.g., a real world phenomenon, an event, a person, a thing). SIWeb is general purpose (it can be applied to any decision making process), cross platforms (it uses the entire Webspace, from social media to websites, from tags to reviews), and time effective (it measures the time correlatio between the Web resources). It uses fractal analysis to detect the temporal relations behind all the Web resources (e.g., Web pages, RSS, newsgroups, etc.) that talk about a topic and combines this number with the temporal relations to give an insight of the the interest the society has about a topic. The evaluation of the proposal shows that SIWeb might be helpful in decision making processes as it reflects the interests the society has on a specific topic.}
        }
    
  37.  Open Access 

    An Introductory Approach to Risk Visualization as a Service

    Victor Chang

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 1-9, 2014, Downloads: 2048, Citations: 13

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194429 | GNL-LP: 1132360595 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper introduces the Risk Visualization as a Service (RVaaS) and presents the motivation, rationale, methodology, Cloud APIs used, operations and examples of using RVaaS. Risks can be calculated within seconds and presented in the form of Visualization to ensure that unexploited areas are ex-posed. RVaaS operates in two phases. The first phase includes the risk modeling in Black Scholes Model (BSM), creating 3D Visualization and Analysis. The second phase consists of calculating key derivatives such as Delta and Theta for financial modeling. Risks presented in visualization allow the potential investors and stakeholders to keep track of the status of risk with regard to time, prices and volatility. Our approach can improve accuracy and performance. Results in experiments show that RVaaS can perform up to 500,000 simulations and complete all simulations within 24 seconds for time steps of up to 50. We also introduce financial stock market analysis (FSMA) that can fully blend with RVaaS and demonstrate two examples that can help investors make better decision based on the pricing and market volatility information. RVaaS provides a structured way to deploy low cost, high quality risk assessment and support real-time calculations.

    BibTex:

        @Article{OJCC-v1i1n01_Chang,
            title     = {An Introductory Approach to Risk Visualization as a Service},
            author    = {Victor Chang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--9},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194429},
            urn       = {urn:nbn:de:101:1-201705194429},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper introduces the Risk Visualization as a Service (RVaaS) and presents the motivation, rationale, methodology, Cloud APIs used, operations and examples of using RVaaS. Risks can be calculated within seconds and presented in the form of Visualization to ensure that unexploited areas are ex-posed. RVaaS operates in two phases. The first phase includes the risk modeling in Black Scholes Model (BSM), creating 3D Visualization and Analysis. The second phase consists of calculating key derivatives such as Delta and Theta for financial modeling. Risks presented in visualization allow the potential investors and stakeholders to keep track of the status of risk with regard to time, prices and volatility. Our approach can improve accuracy and performance. Results in experiments show that RVaaS can perform up to 500,000 simulations and complete all simulations within 24 seconds for time steps of up to 50. We also introduce financial stock market analysis (FSMA) that can fully blend with RVaaS and demonstrate two examples that can help investors make better decision based on the pricing and market volatility information. RVaaS provides a structured way to deploy low cost, high quality risk assessment and support real-time calculations.}
        }
    
  38.  Open Access 

    Measuring and analyzing German and Spanish customer satisfaction of using the iPhone 4S Mobile Cloud service

    Victor Chang

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 19-26, 2014, Downloads: 2014, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194450 | GNL-LP: 1132360633 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper presents the customer satisfaction analysis for measuring popularity in the Mobile Cloud, which is an emerging area in the Cloud and Big Data Computing. Organizational Sustainability Modeling (OSM) is the proposed method used in this research. The twelve-month of German and Spanish consumer data are used for the analysis to investigate the return and risk status associated with the ratings of customer satisfaction in the iPhone 4S Mobile Cloud services. Results show that there is a decline in the satisfaction ratings in Germany and Spain due to economic downturn and competitions in the market, which support our hypothesis. Key outputs have been explained and they confirm that all analysis and interpretations fulfill the criteria for OSM. The use of statistical and visualization method proposed by OSM can expose unexploited data and allows the stakeholders to understand the status of return and risk of their Cloud strategies easier than the use of other data analysis.

    BibTex:

        @Article{OJCC-v1i1n03_Chang,
            title     = {Measuring and analyzing German and Spanish customer satisfaction of using the iPhone 4S Mobile Cloud service},
            author    = {Victor Chang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {19--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194450},
            urn       = {urn:nbn:de:101:1-201705194450},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper presents the customer satisfaction analysis for measuring popularity in the Mobile Cloud, which is an emerging area in the Cloud and Big Data Computing. Organizational Sustainability Modeling (OSM) is the proposed method used in this research. The twelve-month of German and Spanish consumer data are used for the analysis to investigate the return and risk status associated with the ratings of customer satisfaction in the iPhone 4S Mobile Cloud services. Results show that there is a decline in the satisfaction ratings in Germany and Spain due to economic downturn and competitions in the market, which support our hypothesis. Key outputs have been explained and they confirm that all analysis and interpretations fulfill the criteria for OSM. The use of statistical and visualization method proposed by OSM can expose unexploited data and allows the stakeholders to understand the status of return and risk of their Cloud strategies easier than the use of other data analysis.}
        }
    
  39.  Open Access 

    An Analytical Model of Multi-Core Multi-Cluster Architecture (MCMCA)

    Norhazlina Hamid, Robert John Walters, Gary B. Wills

    Open Journal of Cloud Computing (OJCC), 2(1), Pages 4-15, 2015, Downloads: 1928, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194487 | GNL-LP: 1132360692 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Multi-core clusters have emerged as an important contribution in computing technology for provisioning additional processing power in high performance computing and communications. Multi-core architectures are proposed for their capability to provide higher performance without increasing heat and power usage, which is the main concern in a single-core processor. This paper introduces analytical models of a new architecture for large-scale multi-core clusters to improve the communication performance within the interconnection network. The new architecture will be based on a multi - cluster architecture containing clusters of multi-core processors.

    BibTex:

        @Article{OJCC_2015v2i1n02_Hamid,
            title     = {An Analytical Model of Multi-Core Multi-Cluster Architecture (MCMCA)},
            author    = {Norhazlina Hamid and
                         Robert John Walters and
                         Gary B. Wills},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194487},
            urn       = {urn:nbn:de:101:1-201705194487},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Multi-core clusters have emerged as an important contribution in computing technology for provisioning additional processing power in high performance computing and communications. Multi-core architectures are proposed for their capability to provide higher performance without increasing heat and power usage, which is the main concern in a single-core processor. This paper introduces analytical models of a new architecture for large-scale multi-core clusters to improve the communication performance within the interconnection network. The new architecture will be based on a multi - cluster architecture containing clusters of multi-core processors.}
        }
    
  40.  Open Access 

    Integrating Human Factors and Semantic Mark-ups in Adaptive Interactive Systems

    Marios Belk, Panagiotis Germanakos, Efi Papatheocharous, Panayiotis Andreou, George Samaras

    Open Journal of Web Technologies (OJWT), 1(1), Pages 15-26, 2014, Downloads: 1919, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017052611313 | GNL-LP: 113283600X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper focuses on incorporating individual differences in cognitive processing and semantic mark-ups in the context of adaptive interactive systems. In particular, a semantic Web-based adaptation framework is proposed that enables Web content providers to enrich content and functionality of Web environments with semantic mark-ups. The Web content is created using a Web authoring tool and is further processed and reconstructed by an adaptation mechanism based on cognitive factors of users. Main aim of this work is to investigate the added value of personalising content and functionality of Web environments based on the unique cognitive characteristics of users. Accordingly, a user study has been conducted that entailed a psychometric-based survey for extracting the users' cognitive characteristics, combined with a real usage scenario of an existing commercial Web environment that was enriched with semantic mark-ups and personalised based on different adaptation effects. The paper provides interesting insights in the design and development of adaptive interactive systems based on cognitive factors and semantic mark-ups.

    BibTex:

        @Article{OJWT-v1i1n02_Belk,
            title     = {Integrating Human Factors and Semantic Mark-ups in Adaptive Interactive Systems},
            author    = {Marios Belk and
                         Panagiotis Germanakos and
                         Efi Papatheocharous and
                         Panayiotis Andreou and
                         George Samaras},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {15--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017052611313},
            urn       = {urn:nbn:de:101:1-2017052611313},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper focuses on incorporating individual differences in cognitive processing and semantic mark-ups in the context of adaptive interactive systems. In particular, a semantic Web-based adaptation framework is proposed that enables Web content providers to enrich content and functionality of Web environments with semantic mark-ups. The Web content is created using a Web authoring tool and is further processed and reconstructed by an adaptation mechanism based on cognitive factors of users. Main aim of this work is to investigate the added value of personalising content and functionality of Web environments based on the unique cognitive characteristics of users. Accordingly, a user study has been conducted that entailed a psychometric-based survey for extracting the users' cognitive characteristics, combined with a real usage scenario of an existing commercial Web environment that was enriched with semantic mark-ups and personalised based on different adaptation effects. The paper provides interesting insights in the design and development of adaptive interactive systems based on cognitive factors and semantic mark-ups.}
        }
    
  41.  Open Access 

    Cognitive Spam Recognition Using Hadoop and Multicast-Update

    Mukund YR, Sunil Sandeep Nayak, K. Chandrasekaran

    Open Journal of Big Data (OJBD), 1(1), Pages 16-28, 2015, Downloads: 1918, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194340 | GNL-LP: 1132360498 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In today's world of exponentially growing technology, spam is a very common issue faced by users on the internet. Spam not only hinders the performance of a network, but it also wastes space and time, and causes general irritation and presents a multitude of dangers - of viruses, malware, spyware and consequent system failure, identity theft, and other cyber criminal activity. In this context, cognition provides us with a method to help improve the performance of the distributed system. It enables the system to learn what it is supposed to do for different input types as different classifications are made over time and this learning helps it increase its accuracy as time passes. Each system on its own can only do so much learning, because of the limited sample set of inputs that it gets to process. However, in a network, we can make sure that every system knows the different kinds of inputs available and learns what it is supposed to do with a better success rate. Thus, distribution and combination of this cognition across different components of the network leads to an overall improvement in the performance of the system. In this paper, we describe a method to make machines cognitively label spam using Machine Learning and the Naive Bayesian approach. We also present two possible methods of implementation - using a MapReduce Framework (hadoop), and also using messages coupled with a multicast-send based network - with their own subtypes, and the pros and cons of each. We finally present a comparative analysis of the two main methods and provide a basic idea about the usefulness of the two in various different scenarios.

    BibTex:

        @Article{OJBD_2015v1i1n03_YR,
            title     = {Cognitive Spam Recognition Using Hadoop and Multicast-Update},
            author    = {Mukund YR and
                         Sunil Sandeep Nayak and
                         K. Chandrasekaran},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {16--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194340},
            urn       = {urn:nbn:de:101:1-201705194340},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In today's world of exponentially growing technology, spam is a very common issue faced by users on the internet. Spam not only hinders the performance of a network, but it also wastes space and time, and causes general irritation and presents a multitude of dangers - of viruses, malware, spyware and consequent system failure, identity theft, and other cyber criminal activity. In this context, cognition provides us with a method to help improve the performance of the distributed system. It enables the system to learn what it is supposed to do for different input types as different classifications are made over time and this learning helps it increase its accuracy as time passes. Each system on its own can only do so much learning, because of the limited sample set of inputs that it gets to process. However, in a network, we can make sure that every system knows the different kinds of inputs available and learns what it is supposed to do with a better success rate. Thus, distribution and combination of this cognition across different components of the network leads to an overall improvement in the performance of the system. In this paper, we describe a method to make machines cognitively label spam using Machine Learning and the Naive Bayesian approach. We also present two possible methods of implementation - using a MapReduce Framework (hadoop), and also using messages coupled with a multicast-send based network - with their own subtypes, and the pros and cons of each. We finally present a comparative analysis of the two main methods and provide a basic idea about the usefulness of the two in various different scenarios.}
        }
    
  42.  Open Access 

    A NoSQL-Based Framework for Managing Home Services

    Marinette Bouet, Michel Schneider

    Open Journal of Information Systems (OJIS), 3(1), Pages 1-28, 2016, Downloads: 1901

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194810 | GNL-LP: 113236115X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Individuals and companies have an increasing need for services by specialized suppliers in their homes or premises. These services can be quite different and can require different amounts of resources. Service suppliers have to specify the activities to be performed, plan those activities, allocate resources, follow up after their completion and must be able to react to any unexpected situation. Various proposals were formulated to model and implement these functions; however, there is no unified approach that can improve the efficiency of software solutions to enable economy of scale. In this paper, we propose a framework that a service supplier can use to manage geo-localized activities. The proposed framework is based on a NoSQL data model and implemented using the MongoDB system. We also discuss the advantages and drawbacks of a NoSQL approach.

    BibTex:

        @Article{OJIS_2016v3i1n02_Marinette,
            title     = {A NoSQL-Based Framework for Managing Home Services},
            author    = {Marinette Bouet and
                         Michel Schneider},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {1--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194810},
            urn       = {urn:nbn:de:101:1-201705194810},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Individuals and companies have an increasing need for services by specialized suppliers in their homes or premises. These services can be quite different and can require different amounts of resources. Service suppliers have to specify the activities to be performed, plan those activities, allocate resources, follow up after their completion and must be able to react to any unexpected situation. Various proposals were formulated to model and implement these functions; however, there is no unified approach that can improve the efficiency of software solutions to enable economy of scale. In this paper, we propose a framework that a service supplier can use to manage geo-localized activities. The proposed framework is based on a NoSQL data model and implemented using the MongoDB system. We also discuss the advantages and drawbacks of a NoSQL approach.}
        }
    
  43.  Open Access 

    Distributed Join Approaches for W3C-Conform SPARQL Endpoints

    Sven Groppe, Dennis Heinrich, Stefan Werner

    Open Journal of Semantic Web (OJSW), 2(1), Pages 30-52, 2015, Downloads: 1835, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194910 | GNL-LP: 1132361303 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: Currently many SPARQL endpoints are freely available and accessible without any costs to users: Everyone can submit SPARQL queries to SPARQL endpoints via a standardized protocol, where the queries are processed on the datasets of the SPARQL endpoints and the query results are sent back to the user in a standardized format. As these distributed execution environments for semantic big data (as intersection of semantic data and big data) are freely accessible, the Semantic Web is an ideal playground for big data research. However, when utilizing these distributed execution environments, questions about the performance arise. Especially when several datasets (locally and those residing in SPARQL endpoints) need to be combined, distributed joins need to be computed. In this work we give an overview of the various possibilities of distributed join processing in SPARQL endpoints, which follow the SPARQL specification and hence are "W3C conform". We also introduce new distributed join approaches as variants of the Bitvector-Join and combination of the Semi- and Bitvector-Join. Finally we compare all the existing and newly proposed distributed join approaches for W3C conform SPARQL endpoints in an extensive experimental evaluation.

    BibTex:

        @Article{OJSW_2015v2i1n04_Groppe,
            title     = {Distributed Join Approaches for W3C-Conform SPARQL Endpoints},
            author    = {Sven Groppe and
                         Dennis Heinrich and
                         Stefan Werner},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {30--52},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194910},
            urn       = {urn:nbn:de:101:1-201705194910},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Currently many SPARQL endpoints are freely available and accessible without any costs to users: Everyone can submit SPARQL queries to SPARQL endpoints via a standardized protocol, where the queries are processed on the datasets of the SPARQL endpoints and the query results are sent back to the user in a standardized format. As these distributed execution environments for semantic big data (as intersection of semantic data and big data) are freely accessible, the Semantic Web is an ideal playground for big data research. However, when utilizing these distributed execution environments, questions about the performance arise. Especially when several datasets (locally and those residing in SPARQL endpoints) need to be combined, distributed joins need to be computed. In this work we give an overview of the various possibilities of distributed join processing in SPARQL endpoints, which follow the SPARQL specification and hence are "W3C conform". We also introduce new distributed join approaches as variants of the Bitvector-Join and combination of the Semi- and Bitvector-Join. Finally we compare all the existing and newly proposed distributed join approaches for W3C conform SPARQL endpoints in an extensive experimental evaluation.}
        }
    
  44.  Open Access 

    Achieving Business Practicability of Model-Driven Cross-Platform Apps

    Tim A. Majchrzak, Jan Ernsting, Herbert Kuchen

    Open Journal of Information Systems (OJIS), 2(2), Pages 4-15, 2015, Downloads: 1828

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194768 | GNL-LP: 1132361095 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Due to the incompatibility of mobile device platforms such as Android and iOS, apps have to be developed separately for each target platform. Cross-platform development approaches based on Web technology have significantly improved over the last years. However, since they do not lead to native apps, these frameworks are not feasible for all kinds of business apps. Moreover, the way apps are developed is cumbersome. Advanced cross-platform approaches such as MD2, which is based on model-driven development (MDSD) techniques, are a much more powerful yet less mature choice. We discuss business implications of MDSD for apps and introduce MD2 as our proposed solution to fulfill typical requirements. Moreover, we highlight a business-oriented enhancement that further increases MD2's business practicability. We generalize our findings and sketch the path towards more versatile MDSD of apps.

    BibTex:

        @Article{OJIS_2015v2i2n02_Majchrzak,
            title     = {Achieving Business Practicability of Model-Driven Cross-Platform Apps},
            author    = {Tim A. Majchrzak and
                         Jan Ernsting and
                         Herbert Kuchen},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194768},
            urn       = {urn:nbn:de:101:1-201705194768},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Due to the incompatibility of mobile device platforms such as Android and iOS, apps have to be developed separately for each target platform. Cross-platform development approaches based on Web technology have significantly improved over the last years. However, since they do not lead to native apps, these frameworks are not feasible for all kinds of business apps. Moreover, the way apps are developed is cumbersome. Advanced cross-platform approaches such as MD2, which is based on model-driven development (MDSD) techniques, are a much more powerful yet less mature choice. We discuss business implications of MDSD for apps and introduce MD2 as our proposed solution to fulfill typical requirements. Moreover, we highlight a business-oriented enhancement that further increases MD2's business practicability. We generalize our findings and sketch the path towards more versatile MDSD of apps.}
        }
    
  45.  Open Access 

    Statistical Machine Learning in Brain State Classification using EEG Data

    Yuezhe Li, Yuchou Chang, Hong Lin

    Open Journal of Big Data (OJBD), 1(2), Pages 19-33, 2015, Downloads: 1826, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194354 | GNL-LP: 113236051X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In this article, we discuss how to use a variety of machine learning methods, e.g. tree bagging, random forest, boost, support vector machine, and Gaussian mixture model, for building classifiers for electroencephalogram (EEG) data, which is collected from different brain states on different subjects. Also, we discuss how training data size influences misclassification rate. Moreover, the number of subjects that contributes to the training data affects misclassification rate. Furthermore, we discuss how sample entropy contributes to building a classifier. Our results show that classification based on sample entropy give the smallest misclassification rate. Moreover, two data sets were collected from one channel and seven channels respectively. The classification results of each data set show that the more channels we use, the less misclassification we have. Our results show that it is promising to build a self-adaptive classification system by using EEG data to distinguish idle from active state.

    BibTex:

        @Article{OJBD_2015v1i2n03_YuehzeLi,
            title     = {Statistical Machine Learning in Brain State Classification using EEG Data},
            author    = {Yuezhe Li and
                         Yuchou Chang and
                         Hong Lin},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {19--33},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194354},
            urn       = {urn:nbn:de:101:1-201705194354},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In this article, we discuss how to use a variety of machine learning methods, e.g. tree bagging, random forest, boost, support vector machine, and Gaussian mixture model, for building classifiers for electroencephalogram (EEG) data, which is collected from different brain states on different subjects. Also, we discuss how training data size influences misclassification rate. Moreover, the number of subjects that contributes to the training data affects misclassification rate. Furthermore, we discuss how sample entropy contributes to building a classifier. Our results show that classification based on sample entropy give the smallest misclassification rate. Moreover, two data sets were collected from one channel and seven channels respectively. The classification results of each data set show that the more channels we use, the less misclassification we have. Our results show that it is promising to build a self-adaptive classification system by using EEG data to distinguish idle from active state.}
        }
    
  46.  Open Access 

    PatTrieSort - External String Sorting based on Patricia Tries

    Sven Groppe, Dennis Heinrich, Stefan Werner, Christopher Blochwitz, Thilo Pionteck

    Open Journal of Databases (OJDB), 2(1), Pages 36-50, 2015, Downloads: 1822, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194627 | GNL-LP: 1132360889 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: External merge sort belongs to the most efficient and widely used algorithms to sort big data: As much data as fits inside is sorted in main memory and afterwards swapped to external storage as so called initial run. After sorting all the data in this way block-wise, the initial runs are merged in a merging phase in order to retrieve the final sorted run containing the completely sorted original data. Patricia tries are one of the most space-efficient ways to store strings especially those with common prefixes. Hence, we propose to use patricia tries for initial run generation in an external merge sort variant, such that initial runs can become large compared to traditional external merge sort using the same main memory size. Furthermore, we store the initial runs as patricia tries instead of lists of sorted strings. As we will show in this paper, patricia tries can be efficiently merged having a superior performance in comparison to merging runs of sorted strings. We complete our discussion with a complexity analysis as well as a comprehensive performance evaluation, where our new approach outperforms traditional external merge sort by a factor of 4 for sorting over 4 billion strings of real world data.

    BibTex:

        @Article{OJDB_2015v2i1n03_Groppe,
            title     = {PatTrieSort - External String Sorting based on Patricia Tries},
            author    = {Sven Groppe and
                         Dennis Heinrich and
                         Stefan Werner and
                         Christopher Blochwitz and
                         Thilo Pionteck},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {36--50},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194627},
            urn       = {urn:nbn:de:101:1-201705194627},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {External merge sort belongs to the most efficient and widely used algorithms to sort big data: As much data as fits inside is sorted in main memory and afterwards swapped to external storage as so called initial run. After sorting all the data in this way block-wise, the initial runs are merged in a merging phase in order to retrieve the final sorted run containing the completely sorted original data. Patricia tries are one of the most space-efficient ways to store strings especially those with common prefixes. Hence, we propose to use patricia tries for initial run generation in an external merge sort variant, such that initial runs can become large compared to traditional external merge sort using the same main memory size. Furthermore, we store the initial runs as patricia tries instead of lists of sorted strings. As we will show in this paper, patricia tries can be efficiently merged having a superior performance in comparison to merging runs of sorted strings. We complete our discussion with a complexity analysis as well as a comprehensive performance evaluation, where our new approach outperforms traditional external merge sort by a factor of 4 for sorting over 4 billion strings of real world data.}
        }
    
  47.  Open Access 

    Semantic and Web: The Semantic Part

    Sven Groppe, Paulo Rupino da Cunha

    Open Journal of Semantic Web (OJSW), 2(1), Pages 1-3, 2015, Downloads: 1811, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194864 | GNL-LP: 1132361222 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The Web is everywhere in daily life. Business is not possible any more without the fast communication through the web. The knowledge of the humans is reflected in the information accessible in the web. New challenges occur with the flood of information and electronic possibilities for the human being. The current World Wide Web enables an easy, instant access to a vast amount of online information. However, the content in the Web is typically for human consumption, and is not tailored to be machine-processed. The Semantic Web, which is intended to establish a machine-understandable web, thereby offers a promising and potential solution to mining and analyzing web content. The Semantic Web is currently changing from an emergent trend to a technology used in complex real-world applications. This part of the special issue "Semantic and Web" especially investigates how semantic technologies can help the human being to open the new possibilities of the web. The papers, which contribute more to Web technologies, are published in Open Journal of Web Technologies (OJWT).

    BibTex:

        @Article{OJSW_2015v2i1n01e_Groppe,
            title     = {Semantic and Web: The Semantic Part},
            author    = {Sven Groppe and
                         Paulo Rupino da Cunha},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {1--3},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194864},
            urn       = {urn:nbn:de:101:1-201705194864},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The Web is everywhere in daily life. Business is not possible any more without the fast communication through the web. The knowledge of the humans is reflected in the information accessible in the web. New challenges occur with the flood of information and electronic possibilities for the human being. The current World Wide Web enables an easy, instant access to a vast amount of online information. However, the content in the Web is typically for human consumption, and is not tailored to be machine-processed. The Semantic Web, which is intended to establish a machine-understandable web, thereby offers a promising and potential solution to mining and analyzing web content. The Semantic Web is currently changing from an emergent trend to a technology used in complex real-world applications. This part of the special issue "Semantic and Web" especially investigates how semantic technologies can help the human being to open the new possibilities of the web. The papers, which contribute more to Web technologies, are published in Open Journal of Web Technologies (OJWT).}
        }
    
  48.  Open Access 

    High-Dimensional Spatio-Temporal Indexing

    Mathias Menninghaus, Martin Breunig, Elke Pulvermüller

    Open Journal of Databases (OJDB), 3(1), Pages 1-20, 2016, Downloads: 1786

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194635 | GNL-LP: 1132360897 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: There exist numerous indexing methods which handle either spatio-temporal or high-dimensional data well. However, those indexing methods which handle spatio-temporal data well have certain drawbacks when confronted with high-dimensional data. As the most efficient spatio-temporal indexing methods are based on the R-tree and its variants, they face the well known problems in high-dimensional space. Furthermore, most high-dimensional indexing methods try to reduce the number of dimensions in the data being indexed and compress the information given by all dimensions into few dimensions but are not able to store now - relative data. One of the most efficient high-dimensional indexing methods, the Pyramid Technique, is able to handle high-dimensional point-data only. Nonetheless, we take this technique and extend it such that it is able to handle spatio-temporal data as well. We introduce a technique for querying in this structure with spatio-temporal queries. We compare our technique, the Spatio-Temporal Pyramid Adapter (STPA), to the RST-tree for in-memory and on-disk applications. We show that for high dimensions, the extra query-cost for reducing the dimensionality in the Pyramid Technique is clearly exceeded by the rising query-cost in the RST-tree. Concluding, we address the main drawbacks and advantages of our technique.

    BibTex:

        @Article{OJDB_2016v3i1n01_Menninghaus,
            title     = {High-Dimensional Spatio-Temporal Indexing},
            author    = {Mathias Menninghaus and
                         Martin Breunig and
                         Elke Pulverm{\"u}ller},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {1--20},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194635},
            urn       = {urn:nbn:de:101:1-201705194635},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {There exist numerous indexing methods which handle either spatio-temporal or high-dimensional data well. However, those indexing methods which handle spatio-temporal data well have certain drawbacks when confronted with high-dimensional data. As the most efficient spatio-temporal indexing methods are based on the R-tree and its variants, they face the well known problems in high-dimensional space. Furthermore, most high-dimensional indexing methods try to reduce the number of dimensions in the data being indexed and compress the information given by all dimensions into few dimensions but are not able to store now - relative data. One of the most efficient high-dimensional indexing methods, the Pyramid Technique, is able to handle high-dimensional point-data only. Nonetheless, we take this technique and extend it such that it is able to handle spatio-temporal data as well. We introduce a technique for querying in this structure with spatio-temporal queries. We compare our technique, the Spatio-Temporal Pyramid Adapter (STPA), to the RST-tree for in-memory and on-disk applications. We show that for high dimensions, the extra query-cost for reducing the dimensionality in the Pyramid Technique is clearly exceeded by the rising query-cost in the RST-tree. Concluding, we address the main drawbacks and advantages of our technique.}
        }
    
  49.  Open Access 

    Hierarchical Multi-Label Classification Using Web Reasoning for Large Datasets

    Rafael Peixoto, Thomas Hassan, Christophe Cruz, Aurélie Bertaux, Nuno Silva

    Open Journal of Semantic Web (OJSW), 3(1), Pages 1-15, 2016, Downloads: 1709, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194907 | GNL-LP: 113236129X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Extracting valuable data among large volumes of data is one of the main challenges in Big Data. In this paper, a Hierarchical Multi-Label Classification process called Semantic HMC is presented. This process aims to extract valuable data from very large data sources, by automatically learning a label hierarchy and classifying data items.The Semantic HMC process is composed of five scalable steps, namely Indexation, Vectorization, Hierarchization, Resolution and Realization. The first three steps construct automatically a label hierarchy from statistical analysis of data. This paper focuses on the last two steps which perform item classification according to the label hierarchy. The process is implemented as a scalable and distributed application, and deployed on a Big Data platform. A quality evaluation is described, which compares the approach with multi-label classification algorithms from the state of the art dedicated to the same goal. The Semantic HMC approach outperforms state of the art approaches in some areas.

    BibTex:

        @Article{OJSW_2016v3i1n01_Peixoto,
            title     = {Hierarchical Multi-Label Classification Using Web Reasoning for Large Datasets},
            author    = {Rafael Peixoto and
                         Thomas Hassan and
                         Christophe Cruz and
                         Aur\'{e}lie Bertaux and
                         Nuno Silva},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {1--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194907},
            urn       = {urn:nbn:de:101:1-201705194907},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Extracting valuable data among large volumes of data is one of the main challenges in Big Data. In this paper, a Hierarchical Multi-Label Classification process called Semantic HMC is presented. This process aims to extract valuable data from very large data sources, by automatically learning a label hierarchy and classifying data items.The Semantic HMC process is composed of five scalable steps, namely Indexation, Vectorization, Hierarchization, Resolution and Realization. The first three steps construct automatically a label hierarchy from statistical analysis of data. This paper focuses on the last two steps which perform item classification according to the label hierarchy. The process is implemented as a scalable and distributed application, and deployed on a Big Data platform. A quality evaluation is described, which compares the approach with multi-label classification algorithms from the state of the art dedicated to the same goal. The Semantic HMC approach outperforms state of the art approaches in some areas.}
        }
    
  50.  Open Access 

    Relationship between Externalized Knowledge and Evaluation in the Process of Creating Strategic Scenarios

    Teruaki Hayashi, Yukio Ohsawa

    Open Journal of Information Systems (OJIS), 2(1), Pages 29-40, 2015, Downloads: 1707, Citations: 6

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194751 | GNL-LP: 1132361079 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Social systems are changing so rapidly that it is important for humans to make decisions considering uncertainty. A scenario is information about the series of events/actions, which supports decision makers to take actions and reduce risks. We propose Action Planning for refining simple ideas into practical scenarios (strategic scenarios). Frameworks and items on Action Planning Sheets provide participants with organized constraints, to lead to creative and logical thinking for solving real issues in businesses or daily life. Communication among participants who have preset roles leads the externalization of knowledge. In this study, we set three criteria for evaluating strategic scenarios; novelty, utility, and feasibility, and examine the relationship between externalized knowledge and the evaluation values, in order to consider factors which affect the evaluations. Regarding a word contained in roles and scenarios as the smallest unit of knowledge, we calculate Relativeness between roles and scenarios. The results of our experiment suggest that the lower the relativeness of a strategic scenario, the higher the strategic scenario is evaluated in novelty. In addition, in the evaluation of utility, a scenario satisfying a covert requirement tends to be estimated higher. Moreover, we found the externalization of stakeholders may affect the realization of strategic scenarios.

    BibTex:

        @Article{OJIS_2015v2i1n03_Hayashi,
            title     = {Relationship between Externalized Knowledge and Evaluation in the Process of Creating Strategic Scenarios},
            author    = {Teruaki Hayashi and
                         Yukio Ohsawa},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {29--40},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194751},
            urn       = {urn:nbn:de:101:1-201705194751},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Social systems are changing so rapidly that it is important for humans to make decisions considering uncertainty. A scenario is information about the series of events/actions, which supports decision makers to take actions and reduce risks. We propose Action Planning for refining simple ideas into practical scenarios (strategic scenarios). Frameworks and items on Action Planning Sheets provide participants with organized constraints, to lead to creative and logical thinking for solving real issues in businesses or daily life. Communication among participants who have preset roles leads the externalization of knowledge. In this study, we set three criteria for evaluating strategic scenarios; novelty, utility, and feasibility, and examine the relationship between externalized knowledge and the evaluation values, in order to consider factors which affect the evaluations. Regarding a word contained in roles and scenarios as the smallest unit of knowledge, we calculate Relativeness between roles and scenarios. The results of our experiment suggest that the lower the relativeness of a strategic scenario, the higher the strategic scenario is evaluated in novelty. In addition, in the evaluation of utility, a scenario satisfying a covert requirement tends to be estimated higher. Moreover, we found the externalization of stakeholders may affect the realization of strategic scenarios.}
        }
    
  1.  Open Access 

    Scalable Distributed Computing Hierarchy: Cloud, Fog and Dew Computing

    Karolj Skala, Davor Davidovic, Enis Afgan, Ivan Sovic, Zorislav Sojat

    Open Journal of Cloud Computing (OJCC), 2(1), Pages 16-24, 2015, Downloads: 7132, Citations: 40

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194519 | GNL-LP: 1132360749 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The paper considers the conceptual approach for organization of the vertical hierarchical links between the scalable distributed computing paradigms: Cloud Computing, Fog Computing and Dew Computing. In this paper, the Dew Computing is described and recognized as a new structural layer in the existing distributed computing hierarchy. In the existing computing hierarchy, the Dew computing is positioned as the ground level for the Cloud and Fog computing paradigms. Vertical, complementary, hierarchical division from Cloud to Dew Computing satisfies the needs of high- and low-end computing demands in everyday life and work. These new computing paradigms lower the cost and improve the performance, particularly for concepts and applications such as the Internet of Things (IoT) and the Internet of Everything (IoE). In addition, the Dew computing paradigm will require new programming models that will efficiently reduce the complexity and improve the productivity and usability of scalable distributed computing, following the principles of High-Productivity computing.

    BibTex:

        @Article{OJCC_2015v2i1n03_Skala,
            title     = {Scalable Distributed Computing Hierarchy: Cloud, Fog and Dew Computing},
            author    = {Karolj Skala and
                         Davor Davidovic and
                         Enis Afgan and
                         Ivan Sovic and
                         Zorislav Sojat},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {16--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194519},
            urn       = {urn:nbn:de:101:1-201705194519},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The paper considers the conceptual approach for organization of the vertical hierarchical links between the scalable distributed computing paradigms: Cloud Computing, Fog Computing and Dew Computing. In this paper, the Dew Computing is described and recognized as a new structural layer in the existing distributed computing hierarchy. In the existing computing hierarchy, the Dew computing is positioned as the ground level for the Cloud and Fog computing paradigms. Vertical, complementary, hierarchical division from Cloud to Dew Computing satisfies the needs of high- and low-end computing demands in everyday life and work. These new computing paradigms lower the cost and improve the performance, particularly for concepts and applications such as the Internet of Things (IoT) and the Internet of Everything (IoE). In addition, the Dew computing paradigm will require new programming models that will efficiently reduce the complexity and improve the productivity and usability of scalable distributed computing, following the principles of High-Productivity computing.}
        }
    
  2.  Open Access 

    Which NoSQL Database? A Performance Overview

    Veronika Abramova, Jorge Bernardino, Pedro Furtado

    Open Journal of Databases (OJDB), 1(2), Pages 17-24, 2014, Downloads: 12759, Citations: 24

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194607 | GNL-LP: 1132360862 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: NoSQL data stores are widely used to store and retrieve possibly large amounts of data, typically in a key-value format. There are many NoSQL types with different performances, and thus it is important to compare them in terms of performance and verify how the performance is related to the database type. In this paper, we evaluate five most popular NoSQL databases: Cassandra, HBase, MongoDB, OrientDB and Redis. We compare those databases in terms of query performance, based on reads and updates, taking into consideration the typical workloads, as represented by the Yahoo! Cloud Serving Benchmark. This comparison allows users to choose the most appropriate database according to the specific mechanisms and application needs.

    BibTex:

        @Article{OJDB-v1i2n02_Abramova,
            title     = {Which NoSQL Database? A Performance Overview},
            author    = {Veronika Abramova and
                         Jorge Bernardino and
                         Pedro Furtado},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {17--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194607},
            urn       = {urn:nbn:de:101:1-201705194607},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {NoSQL data stores are widely used to store and retrieve possibly large amounts of data, typically in a key-value format. There are many NoSQL types with different performances, and thus it is important to compare them in terms of performance and verify how the performance is related to the database type. In this paper, we evaluate five most popular NoSQL databases: Cassandra, HBase, MongoDB, OrientDB and Redis. We compare those databases in terms of query performance, based on reads and updates, taking into consideration the typical workloads, as represented by the Yahoo! Cloud Serving Benchmark. This comparison allows users to choose the most appropriate database according to the specific mechanisms and application needs.}
        }
    
  3.  Open Access 

    An Introductory Approach to Risk Visualization as a Service

    Victor Chang

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 1-9, 2014, Downloads: 2048, Citations: 13

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194429 | GNL-LP: 1132360595 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper introduces the Risk Visualization as a Service (RVaaS) and presents the motivation, rationale, methodology, Cloud APIs used, operations and examples of using RVaaS. Risks can be calculated within seconds and presented in the form of Visualization to ensure that unexploited areas are ex-posed. RVaaS operates in two phases. The first phase includes the risk modeling in Black Scholes Model (BSM), creating 3D Visualization and Analysis. The second phase consists of calculating key derivatives such as Delta and Theta for financial modeling. Risks presented in visualization allow the potential investors and stakeholders to keep track of the status of risk with regard to time, prices and volatility. Our approach can improve accuracy and performance. Results in experiments show that RVaaS can perform up to 500,000 simulations and complete all simulations within 24 seconds for time steps of up to 50. We also introduce financial stock market analysis (FSMA) that can fully blend with RVaaS and demonstrate two examples that can help investors make better decision based on the pricing and market volatility information. RVaaS provides a structured way to deploy low cost, high quality risk assessment and support real-time calculations.

    BibTex:

        @Article{OJCC-v1i1n01_Chang,
            title     = {An Introductory Approach to Risk Visualization as a Service},
            author    = {Victor Chang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--9},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194429},
            urn       = {urn:nbn:de:101:1-201705194429},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper introduces the Risk Visualization as a Service (RVaaS) and presents the motivation, rationale, methodology, Cloud APIs used, operations and examples of using RVaaS. Risks can be calculated within seconds and presented in the form of Visualization to ensure that unexploited areas are ex-posed. RVaaS operates in two phases. The first phase includes the risk modeling in Black Scholes Model (BSM), creating 3D Visualization and Analysis. The second phase consists of calculating key derivatives such as Delta and Theta for financial modeling. Risks presented in visualization allow the potential investors and stakeholders to keep track of the status of risk with regard to time, prices and volatility. Our approach can improve accuracy and performance. Results in experiments show that RVaaS can perform up to 500,000 simulations and complete all simulations within 24 seconds for time steps of up to 50. We also introduce financial stock market analysis (FSMA) that can fully blend with RVaaS and demonstrate two examples that can help investors make better decision based on the pricing and market volatility information. RVaaS provides a structured way to deploy low cost, high quality risk assessment and support real-time calculations.}
        }
    
  4.  Open Access 

    Block-level De-duplication with Encrypted Data

    Pasquale Puzio, Refik Molva, Melek Önen, Sergio Loureiro

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 10-18, 2014, Downloads: 3656, Citations: 12

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194448 | GNL-LP: 1132360617 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Deduplication is a storage saving technique which has been adopted by many cloud storage providers such as Dropbox. The simple principle of deduplication is that duplicate data uploaded by different users are stored only once. Unfortunately, deduplication is not compatible with encryption. As a scheme that allows deduplication of encrypted data segments, we propose ClouDedup, a secure and efficient storage service which guarantees blocklevel deduplication and data confidentiality at the same time. ClouDedup strengthens convergent encryption by employing a component that implements an additional encryption operation and an access control mechanism. We also propose to introduce an additional component which is in charge of providing a key management system for data blocks together with the actual deduplication operation. We show that the overhead introduced by these new components is minimal and does not impact the overall storage and computational costs.

    BibTex:

        @Article{OJCC-v1i1n02_Puzio,
            title     = {Block-level De-duplication with Encrypted Data},
            author    = {Pasquale Puzio and
                         Refik Molva and
                         Melek {\"O}nen and
                         Sergio Loureiro},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {10--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194448},
            urn       = {urn:nbn:de:101:1-201705194448},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Deduplication is a storage saving technique which has been adopted by many cloud storage providers such as Dropbox. The simple principle of deduplication is that duplicate data uploaded by different users are stored only once. Unfortunately, deduplication is not compatible with encryption. As a scheme that allows deduplication of encrypted data segments, we propose ClouDedup, a secure and efficient storage service which guarantees blocklevel deduplication and data confidentiality at the same time. ClouDedup strengthens convergent encryption by employing a component that implements an additional encryption operation and an access control mechanism. We also propose to introduce an additional component which is in charge of providing a key management system for data blocks together with the actual deduplication operation. We show that the overhead introduced by these new components is minimal and does not impact the overall storage and computational costs.}
        }
    
  5.  Open Access 

    Modelling the Integrated QoS for Wireless Sensor Networks with Heterogeneous Data Traffic

    Syarifah Ezdiani, Adnan Al-Anbuky

    Open Journal of Internet Of Things (OJIOT), 1(1), Pages 1-15, 2015, Downloads: 2562, Citations: 10

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244946 | GNL-LP: 1130621979 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The future of Internet of Things (IoT) is envisaged to consist of a high amount of wireless resource-constrained devices connected to the Internet. Moreover, a lot of novel real-world services offered by IoT devices are realized by wireless sensor networks (WSNs). Integrating WSN to the Internet has therefore brought forward the requirements of an end-to-end quality of service (QoS) guarantee. In this paper, the QoS requirements for the WSN-Internet integration are investigated by first distinguishing the Internet QoS from the WSN QoS. Next, this study emphasizes on WSN applications that involve traffic with different levels of importance, thus the way realtime traffic and delay-tolerant traffic are handled to guarantee QoS in the network is studied. Additionally, an overview of the integration strategies is given, and the delay-tolerant network (DTN) gateway, being one of the desirable approaches for integrating WSNs to the Internet, is discussed. Next, the implementation of the service model is presented, by considering both traffic prioritization and service differentiation. Based on the simulation results in OPNET Modeler, it is observed that real-time traffic achieve low bound delay while delay-tolerant traffic experience a lower packet dropped, hence indicating that the needs of real-time and delay-tolerant traffic can be better met by treating both packet types differently. Furthermore, a vehicular network is used as an example case to describe the applicability of the framework in a real IoT application environment, followed by a discussion on the future work of this research.

    BibTex:

        @Article{OJIOT_2015v1i1n02_Syarifah,
            title     = {Modelling the Integrated QoS for Wireless Sensor Networks with Heterogeneous Data Traffic},
            author    = {Syarifah Ezdiani and
                         Adnan Al-Anbuky},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {1--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244946},
            urn       = {urn:nbn:de:101:1-201704244946},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The future of Internet of Things (IoT) is envisaged to consist of a high amount of wireless resource-constrained devices connected to the Internet. Moreover, a lot of novel real-world services offered by IoT devices are realized by wireless sensor networks (WSNs). Integrating WSN to the Internet has therefore brought forward the requirements of an end-to-end quality of service (QoS) guarantee. In this paper, the QoS requirements for the WSN-Internet integration are investigated by first distinguishing the Internet QoS from the WSN QoS. Next, this study emphasizes on WSN applications that involve traffic with different levels of importance, thus the way realtime traffic and delay-tolerant traffic are handled to guarantee QoS in the network is studied. Additionally, an overview of the integration strategies is given, and the delay-tolerant network (DTN) gateway, being one of the desirable approaches for integrating WSNs to the Internet, is discussed. Next, the implementation of the service model is presented, by considering both traffic prioritization and service differentiation. Based on the simulation results in OPNET Modeler, it is observed that real-time traffic achieve low bound delay while delay-tolerant traffic experience a lower packet dropped, hence indicating that the needs of real-time and delay-tolerant traffic can be better met by treating both packet types differently. Furthermore, a vehicular network is used as an example case to describe the applicability of the framework in a real IoT application environment, followed by a discussion on the future work of this research.}
        }
    
  6.  Open Access 

    The Potential of Printed Electronics and Personal Fabrication in Driving the Internet of Things

    Paulo Rosa, António Câmara, Cristina Gouveia

    Open Journal of Internet Of Things (OJIOT), 1(1), Pages 16-36, 2015, Downloads: 3811, Citations: 9

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244933 | GNL-LP: 1130621448 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In the early nineties, Mark Weiser, a chief scientist at the Xerox Palo Alto Research Center (PARC), wrote a series of seminal papers that introduced the concept of Ubiquitous Computing. Within this vision, computers and others digital technologies are integrated seamlessly into everyday objects and activities, hidden from our senses whenever not used or needed. An important facet of this vision is the interconnectivity of the various physical devices, which creates an Internet of Things. With the advent of Printed Electronics, new ways to link the physical and digital worlds became available. Common printing technologies, such as screen, flexography, and inkjet printing, are now starting to be used not only to mass-produce extremely thin, flexible and cost effective electronic circuits, but also to introduce electronic functionality into objects where it was previously unavailable. In turn, the growing accessibility to Personal Fabrication tools is leading to the democratization of the creation of technology by enabling end-users to design and produce their own material goods according to their needs. This paper presents a survey of commonly used technologies and foreseen applications in the field of Printed Electronics and Personal Fabrication, with emphasis on the potential to drive the Internet of Things.

    BibTex:

        @Article{OJIOT_2015v1i1n03_Rosa,
            title     = {The Potential of Printed Electronics and Personal Fabrication in Driving the Internet of Things},
            author    = {Paulo Rosa and
                         Ant\'{o}nio C\^{a}mara and
                         Cristina Gouveia},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {16--36},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244933},
            urn       = {urn:nbn:de:101:1-201704244933},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In the early nineties, Mark Weiser, a chief scientist at the Xerox Palo Alto Research Center (PARC), wrote a series of seminal papers that introduced the concept of Ubiquitous Computing. Within this vision, computers and others digital technologies are integrated seamlessly into everyday objects and activities, hidden from our senses whenever not used or needed. An important facet of this vision is the interconnectivity of the various physical devices, which creates an Internet of Things. With the advent of Printed Electronics, new ways to link the physical and digital worlds became available. Common printing technologies, such as screen, flexography, and inkjet printing, are now starting to be used not only to mass-produce extremely thin, flexible and cost effective electronic circuits, but also to introduce electronic functionality into objects where it was previously unavailable. In turn, the growing accessibility to Personal Fabrication tools is leading to the democratization of the creation of technology by enabling end-users to design and produce their own material goods according to their needs. This paper presents a survey of commonly used technologies and foreseen applications in the field of Printed Electronics and Personal Fabrication, with emphasis on the potential to drive the Internet of Things.}
        }
    
  7.  Open Access 

    Measuring and analyzing German and Spanish customer satisfaction of using the iPhone 4S Mobile Cloud service

    Victor Chang

    Open Journal of Cloud Computing (OJCC), 1(1), Pages 19-26, 2014, Downloads: 2014, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194450 | GNL-LP: 1132360633 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper presents the customer satisfaction analysis for measuring popularity in the Mobile Cloud, which is an emerging area in the Cloud and Big Data Computing. Organizational Sustainability Modeling (OSM) is the proposed method used in this research. The twelve-month of German and Spanish consumer data are used for the analysis to investigate the return and risk status associated with the ratings of customer satisfaction in the iPhone 4S Mobile Cloud services. Results show that there is a decline in the satisfaction ratings in Germany and Spain due to economic downturn and competitions in the market, which support our hypothesis. Key outputs have been explained and they confirm that all analysis and interpretations fulfill the criteria for OSM. The use of statistical and visualization method proposed by OSM can expose unexploited data and allows the stakeholders to understand the status of return and risk of their Cloud strategies easier than the use of other data analysis.

    BibTex:

        @Article{OJCC-v1i1n03_Chang,
            title     = {Measuring and analyzing German and Spanish customer satisfaction of using the iPhone 4S Mobile Cloud service},
            author    = {Victor Chang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {19--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194450},
            urn       = {urn:nbn:de:101:1-201705194450},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper presents the customer satisfaction analysis for measuring popularity in the Mobile Cloud, which is an emerging area in the Cloud and Big Data Computing. Organizational Sustainability Modeling (OSM) is the proposed method used in this research. The twelve-month of German and Spanish consumer data are used for the analysis to investigate the return and risk status associated with the ratings of customer satisfaction in the iPhone 4S Mobile Cloud services. Results show that there is a decline in the satisfaction ratings in Germany and Spain due to economic downturn and competitions in the market, which support our hypothesis. Key outputs have been explained and they confirm that all analysis and interpretations fulfill the criteria for OSM. The use of statistical and visualization method proposed by OSM can expose unexploited data and allows the stakeholders to understand the status of return and risk of their Cloud strategies easier than the use of other data analysis.}
        }
    
  8.  Open Access 

    Eventual Consistent Databases: State of the Art

    Mawahib Musa Elbushra, Jan Lindström

    Open Journal of Databases (OJDB), 1(1), Pages 26-41, 2014, Downloads: 6701, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194582 | GNL-LP: 1132360846 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: One of the challenges of cloud programming is to achieve the right balance between the availability and consistency in a distributed database. Cloud computing environments, particularly cloud databases, are rapidly increasing in importance, acceptance and usage in major applications, which need the partition-tolerance and availability for scalability purposes, but sacrifice the consistency side (CAP theorem). In these environments, the data accessed by users is stored in a highly available storage system, thus the use of paradigms such as eventual consistency became more widespread. In this paper, we review the state-of-the-art database systems using eventual consistency from both industry and research. Based on this review, we discuss the advantages and disadvantages of eventual consistency, and identify the future research challenges on the databases using eventual consistency.

    BibTex:

        @Article{OJDB-v1i1n03_Elbushra,
            title     = {Eventual Consistent Databases: State of the Art},
            author    = {Mawahib Musa Elbushra and
                         Jan Lindstr{\"o}m},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {26--41},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194582},
            urn       = {urn:nbn:de:101:1-201705194582},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {One of the challenges of cloud programming is to achieve the right balance between the availability and consistency in a distributed database. Cloud computing environments, particularly cloud databases, are rapidly increasing in importance, acceptance and usage in major applications, which need the partition-tolerance and availability for scalability purposes, but sacrifice the consistency side (CAP theorem). In these environments, the data accessed by users is stored in a highly available storage system, thus the use of paradigms such as eventual consistency became more widespread. In this paper, we review the state-of-the-art database systems using eventual consistency from both industry and research. Based on this review, we discuss the advantages and disadvantages of eventual consistency, and identify the future research challenges on the databases using eventual consistency.}
        }
    
  9.  Open Access 

    Detecting Data-Flow Errors in BPMN 2.0

    Silvia von Stackelberg, Susanne Putze, Jutta Mülle, Klemens Böhm

    Open Journal of Information Systems (OJIS), 1(2), Pages 1-19, 2014, Downloads: 2734, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017052611934 | GNL-LP: 1132836972 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Data-flow errors in BPMN 2.0 process models, such as missing or unused data, lead to undesired process executions. In particular, since BPMN 2.0 with a standardized execution semantics allows specifying alternatives for data as well as optional data, identifying missing or unused data systematically is difficult. In this paper, we propose an approach for detecting data-flow errors in BPMN 2.0 process models. We formalize BPMN process models by mapping them to Petri Nets and unfolding the execution semantics regarding data. We define a set of anti-patterns representing data-flow errors of BPMN 2.0 process models. By employing the anti-patterns, our tool performs model checking for the unfolded Petri Nets. The evaluation shows that it detects all data-flow errors identified by hand, and so improves process quality.

    BibTex:

        @Article{OJIS-2014v1i2n01_Stackelberg,
            title     = {Detecting Data-Flow Errors in BPMN 2.0},
            author    = {Silvia von Stackelberg and
                         Susanne Putze and
                         Jutta M{\"u}lle and
                         Klemens B{\"o}hm},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--19},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017052611934},
            urn       = {urn:nbn:de:101:1-2017052611934},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Data-flow errors in BPMN 2.0 process models, such as missing or unused data, lead to undesired process executions. In particular, since BPMN 2.0 with a standardized execution semantics allows specifying alternatives for data as well as optional data, identifying missing or unused data systematically is difficult. In this paper, we propose an approach for detecting data-flow errors in BPMN 2.0 process models. We formalize BPMN process models by mapping them to Petri Nets and unfolding the execution semantics regarding data. We define a set of anti-patterns representing data-flow errors of BPMN 2.0 process models. By employing the anti-patterns, our tool performs model checking for the unfolded Petri Nets. The evaluation shows that it detects all data-flow errors identified by hand, and so improves process quality.}
        }
    
  10.  Open Access 

    Definition and Categorization of Dew Computing

    Yingwei Wang

    Open Journal of Cloud Computing (OJCC), 3(1), Pages 1-7, 2016, Downloads: 2149, Citations: 8

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194546 | GNL-LP: 1132360781 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Dew computing is an emerging new research area and has great potentials in applications. In this paper, we propose a revised definition of dew computing. The new definition is: Dew computing is an on-premises computer software-hardware organization paradigm in the cloud computing environment where the on-premises computer provides functionality that is independent of cloud services and is also collaborative with cloud services. The goal of dew computing is to fully realize the potentials of on-premises computers and cloud services. This definition emphasizes two key features of dew computing: independence and collaboration. Furthermore, we propose a group of dew computing categories. These categories may inspire new applications.

    BibTex:

        @Article{OJCC_2016v3i1n02_YingweiWang,
            title     = {Definition and Categorization of Dew Computing},
            author    = {Yingwei Wang},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {1--7},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194546},
            urn       = {urn:nbn:de:101:1-201705194546},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Dew computing is an emerging new research area and has great potentials in applications. In this paper, we propose a revised definition of dew computing. The new definition is: Dew computing is an on-premises computer software-hardware organization paradigm in the cloud computing environment where the on-premises computer provides functionality that is independent of cloud services and is also collaborative with cloud services. The goal of dew computing is to fully realize the potentials of on-premises computers and cloud services. This definition emphasizes two key features of dew computing: independence and collaboration. Furthermore, we propose a group of dew computing categories. These categories may inspire new applications.}
        }
    
  11.  Open Access 

    Designing a Benchmark for the Assessment of Schema Matching Tools

    Fabien Duchateau, Zohra Bellahsene

    Open Journal of Databases (OJDB), 1(1), Pages 3-25, 2014, Downloads: 2082, Citations: 7

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194573 | GNL-LP: 1132360838 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Over the years, many schema matching approaches have been developed to discover correspondences between schemas. Although this task is crucial in data integration, its evaluation, both in terms of matching quality and time performance, is still manually performed. Indeed, there is no common platform which gathers a collection of schema matching datasets to fulfil this goal. Another problem deals with the measuring of the post-match effort, a human cost that schema matching approaches aim at reducing. Consequently, we propose XBenchMatch, a schema matching benchmark with available datasets and new measures to evaluate this manual post-match effort and the quality of integrated schemas. We finally report the results obtained by different approaches, namely COMA++, Similarity Flooding and YAM. We show that such a benchmark is required to understand the advantages and failures of schema matching approaches. Therefore, it could help an end-user to select a schema matching tool which covers his/her needs.

    BibTex:

        @Article{OJDB-v1i1n02_Duchateau,
            title     = {Designing a Benchmark for the Assessment of Schema Matching Tools},
            author    = {Fabien Duchateau and
                         Zohra Bellahsene},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {3--25},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194573},
            urn       = {urn:nbn:de:101:1-201705194573},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Over the years, many schema matching approaches have been developed to discover correspondences between schemas. Although this task is crucial in data integration, its evaluation, both in terms of matching quality and time performance, is still manually performed. Indeed, there is no common platform which gathers a collection of schema matching datasets to fulfil this goal. Another problem deals with the measuring of the post-match effort, a human cost that schema matching approaches aim at reducing. Consequently, we propose XBenchMatch, a schema matching benchmark with available datasets and new measures to evaluate this manual post-match effort and the quality of integrated schemas. We finally report the results obtained by different approaches, namely COMA++, Similarity Flooding and YAM. We show that such a benchmark is required to understand the advantages and failures of schema matching approaches. Therefore, it could help an end-user to select a schema matching tool which covers his/her needs.}
        }
    
  12.  Open Access 

    MapReduce-based Solutions for Scalable SPARQL Querying

    José M. Giménez-Garcia, Javier D. Fernández, Miguel A. Martínez-Prieto

    Open Journal of Semantic Web (OJSW), 1(1), Pages 1-18, 2014, Downloads: 2616, Citations: 6

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194824 | GNL-LP: 1132361168 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The use of RDF to expose semantic data on the Web has seen a dramatic increase over the last few years. Nowadays, RDF datasets are so big and rconnected that, in fact, classical mono-node solutions present significant scalability problems when trying to manage big semantic data. MapReduce, a standard framework for distributed processing of great quantities of data, is earning a place among the distributed solutions facing RDF scalability issues. In this article, we survey the most important works addressing RDF management and querying through diverse MapReduce approaches, with a focus on their main strategies, optimizations and results.

    BibTex:

        @Article{OJSW-v1i1n02_Garcia,
            title     = {MapReduce-based Solutions for Scalable SPARQL Querying},
            author    = {Jos\'{e} M. Gim\'{e}nez-Garcia and
                         Javier D. Fern\'{a}ndez and
                         Miguel A. Mart\'{i}nez-Prieto},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194824},
            urn       = {urn:nbn:de:101:1-201705194824},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The use of RDF to expose semantic data on the Web has seen a dramatic increase over the last few years. Nowadays, RDF datasets are so big and rconnected that, in fact, classical mono-node solutions present significant scalability problems when trying to manage big semantic data. MapReduce, a standard framework for distributed processing of great quantities of data, is earning a place among the distributed solutions facing RDF scalability issues. In this article, we survey the most important works addressing RDF management and querying through diverse MapReduce approaches, with a focus on their main strategies, optimizations and results.}
        }
    
  13.  Open Access 

    Relationship between Externalized Knowledge and Evaluation in the Process of Creating Strategic Scenarios

    Teruaki Hayashi, Yukio Ohsawa

    Open Journal of Information Systems (OJIS), 2(1), Pages 29-40, 2015, Downloads: 1707, Citations: 6

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194751 | GNL-LP: 1132361079 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Social systems are changing so rapidly that it is important for humans to make decisions considering uncertainty. A scenario is information about the series of events/actions, which supports decision makers to take actions and reduce risks. We propose Action Planning for refining simple ideas into practical scenarios (strategic scenarios). Frameworks and items on Action Planning Sheets provide participants with organized constraints, to lead to creative and logical thinking for solving real issues in businesses or daily life. Communication among participants who have preset roles leads the externalization of knowledge. In this study, we set three criteria for evaluating strategic scenarios; novelty, utility, and feasibility, and examine the relationship between externalized knowledge and the evaluation values, in order to consider factors which affect the evaluations. Regarding a word contained in roles and scenarios as the smallest unit of knowledge, we calculate Relativeness between roles and scenarios. The results of our experiment suggest that the lower the relativeness of a strategic scenario, the higher the strategic scenario is evaluated in novelty. In addition, in the evaluation of utility, a scenario satisfying a covert requirement tends to be estimated higher. Moreover, we found the externalization of stakeholders may affect the realization of strategic scenarios.

    BibTex:

        @Article{OJIS_2015v2i1n03_Hayashi,
            title     = {Relationship between Externalized Knowledge and Evaluation in the Process of Creating Strategic Scenarios},
            author    = {Teruaki Hayashi and
                         Yukio Ohsawa},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {29--40},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194751},
            urn       = {urn:nbn:de:101:1-201705194751},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Social systems are changing so rapidly that it is important for humans to make decisions considering uncertainty. A scenario is information about the series of events/actions, which supports decision makers to take actions and reduce risks. We propose Action Planning for refining simple ideas into practical scenarios (strategic scenarios). Frameworks and items on Action Planning Sheets provide participants with organized constraints, to lead to creative and logical thinking for solving real issues in businesses or daily life. Communication among participants who have preset roles leads the externalization of knowledge. In this study, we set three criteria for evaluating strategic scenarios; novelty, utility, and feasibility, and examine the relationship between externalized knowledge and the evaluation values, in order to consider factors which affect the evaluations. Regarding a word contained in roles and scenarios as the smallest unit of knowledge, we calculate Relativeness between roles and scenarios. The results of our experiment suggest that the lower the relativeness of a strategic scenario, the higher the strategic scenario is evaluated in novelty. In addition, in the evaluation of utility, a scenario satisfying a covert requirement tends to be estimated higher. Moreover, we found the externalization of stakeholders may affect the realization of strategic scenarios.}
        }
    
  14.  Open Access 

    Fuzzy Color Space for Apparel Coordination

    Pakizar Shamoi, Atsushi Inoue, Hiroharu Kawanaka

    Open Journal of Information Systems (OJIS), 1(2), Pages 20-28, 2014, Downloads: 2079, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194710 | GNL-LP: 1132360994 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Human perception of colors constitutes an important part in color theory. The applications of color science are truly omnipresent, and what impression colors make on human plays a vital role in them. In this paper, we offer the novel approach for color information representation and processing using fuzzy sets and logic theory, which is extremely useful in modeling human impressions. Specifically, we use fuzzy mathematics to partition the gamut of feasible colors in HSI color space based on standard linguistic tags. The proposed method can be useful in various image processing applications involving query processing. We demonstrate its effectivity in the implementation of a framework for the apparel online shopping coordination based on a color scheme. It deserves attention, since there is always some uncertainty inherent in the description of apparels.

    BibTex:

        @Article{OJIS_2014v1i2n02_Shamoi,
            title     = {Fuzzy Color Space for Apparel Coordination},
            author    = {Pakizar Shamoi and
                         Atsushi Inoue and
                         Hiroharu Kawanaka},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {20--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194710},
            urn       = {urn:nbn:de:101:1-201705194710},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Human perception of colors constitutes an important part in color theory. The applications of color science are truly omnipresent, and what impression colors make on human plays a vital role in them. In this paper, we offer the novel approach for color information representation and processing using fuzzy sets and logic theory, which is extremely useful in modeling human impressions. Specifically, we use fuzzy mathematics to partition the gamut of feasible colors in HSI color space based on standard linguistic tags. The proposed method can be useful in various image processing applications involving query processing. We demonstrate its effectivity in the implementation of a framework for the apparel online shopping coordination based on a color scheme. It deserves attention, since there is always some uncertainty inherent in the description of apparels.}
        }
    
  15.  Open Access 

    Accurate Distance Estimation between Things: A Self-correcting Approach

    Ho-sik Cho, Jianxun Ji, Zili Chen, Hyuncheol Park, Wonsuk Lee

    Open Journal of Internet Of Things (OJIOT), 1(2), Pages 19-27, 2015, Downloads: 6454, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244959 | GNL-LP: 1130622525 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper suggests a method to measure the physical distance between an IoT device (a Thing) and a mobile device (also a Thing) using BLE (Bluetooth Low-Energy profile) interfaces with smaller distance errors. BLE is a well-known technology for the low-power connectivity and suitable for IoT devices as well as for the proximity with the range of several meters. Apple has already adopted the technique and enhanced it to provide subdivided proximity range levels. However, as it is also a variation of RSS-based distance estimation, Apple's iBeacon could only provide immediate, near or far status but not a real and accurate distance. To provide more accurate distance using BLE, this paper introduces additional self-correcting beacon to calibrate the reference distance and mitigate errors from environmental factors. By adopting self-correcting beacon for measuring the distance, the average distance error shows less than 10% within the range of 1.5 meters. Some considerations are presented to extend the range to be able to get more accurate distances.

    BibTex:

        @Article{OJIOT_2015v1i2n03_Cho,
            title     = {Accurate Distance Estimation between Things: A Self-correcting Approach},
            author    = {Ho-sik Cho and
                         Jianxun Ji and
                         Zili Chen and
                         Hyuncheol Park and
                         Wonsuk Lee},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {19--27},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244959},
            urn       = {urn:nbn:de:101:1-201704244959},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper suggests a method to measure the physical distance between an IoT device (a Thing) and a mobile device (also a Thing) using BLE (Bluetooth Low-Energy profile) interfaces with smaller distance errors. BLE is a well-known technology for the low-power connectivity and suitable for IoT devices as well as for the proximity with the range of several meters. Apple has already adopted the technique and enhanced it to provide subdivided proximity range levels. However, as it is also a variation of RSS-based distance estimation, Apple's iBeacon could only provide immediate, near or far status but not a real and accurate distance. To provide more accurate distance using BLE, this paper introduces additional self-correcting beacon to calibrate the reference distance and mitigate errors from environmental factors. By adopting self-correcting beacon for measuring the distance, the average distance error shows less than 10\% within the range of 1.5 meters. Some considerations are presented to extend the range to be able to get more accurate distances.}
        }
    
  16.  Open Access 

    Distributed Join Approaches for W3C-Conform SPARQL Endpoints

    Sven Groppe, Dennis Heinrich, Stefan Werner

    Open Journal of Semantic Web (OJSW), 2(1), Pages 30-52, 2015, Downloads: 1835, Citations: 5

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194910 | GNL-LP: 1132361303 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: Currently many SPARQL endpoints are freely available and accessible without any costs to users: Everyone can submit SPARQL queries to SPARQL endpoints via a standardized protocol, where the queries are processed on the datasets of the SPARQL endpoints and the query results are sent back to the user in a standardized format. As these distributed execution environments for semantic big data (as intersection of semantic data and big data) are freely accessible, the Semantic Web is an ideal playground for big data research. However, when utilizing these distributed execution environments, questions about the performance arise. Especially when several datasets (locally and those residing in SPARQL endpoints) need to be combined, distributed joins need to be computed. In this work we give an overview of the various possibilities of distributed join processing in SPARQL endpoints, which follow the SPARQL specification and hence are "W3C conform". We also introduce new distributed join approaches as variants of the Bitvector-Join and combination of the Semi- and Bitvector-Join. Finally we compare all the existing and newly proposed distributed join approaches for W3C conform SPARQL endpoints in an extensive experimental evaluation.

    BibTex:

        @Article{OJSW_2015v2i1n04_Groppe,
            title     = {Distributed Join Approaches for W3C-Conform SPARQL Endpoints},
            author    = {Sven Groppe and
                         Dennis Heinrich and
                         Stefan Werner},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {30--52},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194910},
            urn       = {urn:nbn:de:101:1-201705194910},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Currently many SPARQL endpoints are freely available and accessible without any costs to users: Everyone can submit SPARQL queries to SPARQL endpoints via a standardized protocol, where the queries are processed on the datasets of the SPARQL endpoints and the query results are sent back to the user in a standardized format. As these distributed execution environments for semantic big data (as intersection of semantic data and big data) are freely accessible, the Semantic Web is an ideal playground for big data research. However, when utilizing these distributed execution environments, questions about the performance arise. Especially when several datasets (locally and those residing in SPARQL endpoints) need to be combined, distributed joins need to be computed. In this work we give an overview of the various possibilities of distributed join processing in SPARQL endpoints, which follow the SPARQL specification and hence are "W3C conform". We also introduce new distributed join approaches as variants of the Bitvector-Join and combination of the Semi- and Bitvector-Join. Finally we compare all the existing and newly proposed distributed join approaches for W3C conform SPARQL endpoints in an extensive experimental evaluation.}
        }
    
  17.  Open Access 

    An Efficient Approach for Cost Optimization of the Movement of Big Data

    Prasad Teli, Manoj V. Thomas, K. Chandrasekaran

    Open Journal of Big Data (OJBD), 1(1), Pages 4-15, 2015, Downloads: 2390, Citations: 4

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194335 | GNL-LP: 113236048X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: With the emergence of cloud computing, Big Data has caught the attention of many researchers in the area of cloud computing. As the Volume, Velocity and Variety (3 Vs) of big data are growing exponentially, dealing with them is a big challenge, especially in the cloud environment. Looking at the current trend of the IT sector, cloud computing is mainly used by the service providers to host their applications. A lot of research has been done to improve the network utilization of WAN (Wide Area Network) and it has achieved considerable success over the traditional LAN (Local Area Network) techniques. While dealing with this issue, the major questions of data movement such as from where to where this big data will be moved and also how the data will be moved, have been overlooked. As various applications generating the big data are hosted in geographically distributed data centers, they individually collect large volume of data in the form of application data as well as the logs. This paper mainly focuses on the challenge of moving big data from one data center to other. We provide an efficient algorithm for the optimization of cost in the movement of the big data from one data center to another for offline environment. This approach uses the graph model for data centers in the cloud and results show that the adopted mechanism provides a better solution to minimize the cost for data movement.

    BibTex:

        @Article{OJBD_2015v1i1n02_Teli,
            title     = {An Efficient Approach for Cost Optimization of the Movement of Big Data},
            author    = {Prasad Teli and
                         Manoj V. Thomas and
                         K. Chandrasekaran},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194335},
            urn       = {urn:nbn:de:101:1-201705194335},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {With the emergence of cloud computing, Big Data has caught the attention of many researchers in the area of cloud computing. As the Volume, Velocity and Variety (3 Vs) of big data are growing exponentially, dealing with them is a big challenge, especially in the cloud environment. Looking at the current trend of the IT sector, cloud computing is mainly used by the service providers to host their applications. A lot of research has been done to improve the network utilization of WAN (Wide Area Network) and it has achieved considerable success over the traditional LAN (Local Area Network) techniques. While dealing with this issue, the major questions of data movement such as from where to where this big data will be moved and also how the data will be moved, have been overlooked. As various applications generating the big data are hosted in geographically distributed data centers, they individually collect large volume of data in the form of application data as well as the logs. This paper mainly focuses on the challenge of moving big data from one data center to other. We provide an efficient algorithm for the optimization of cost in the movement of the big data from one data center to another for offline environment. This approach uses the graph model for data centers in the cloud and results show that the adopted mechanism provides a better solution to minimize the cost for data movement.}
        }
    
  18.  Open Access 

    Context-Dependent Testing of Applications for Mobile Devices

    Tim A. Majchrzak, Matthias Schulte

    Open Journal of Web Technologies (OJWT), 2(1), Pages 27-39, 2015, Downloads: 1562, Citations: 4

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291390 | GNL-LP: 1133021646 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Applications propel the versatility of mobile devices. Apps enable the realization of new ideas and greatly contribute to the proliferation of mobile computing. Unfortunately, software quality of apps often is low. This at least partly can be attributed to problems with testing them. However, it is not a lack of techniques or tools that make app testing cumbersome. Rather, frequent context changes have to be dealt with. Mobile devices most notably move: network parameters such as latency and usable bandwidth change, along with data read from sensors such as GPS coordinates. Additionally, usage patterns vary. To address context changes in testing, we propose a novel concept. It is based on identifying blocks of code between which context changes are possible. It helps to greatly reduce complexity. Besides introducing our concept, we present a use case, show its application and benefits, and discuss challenges.

    BibTex:

        @Article{OJWT_2015v2i1n04_Majchrzak,
            title     = {Context-Dependent Testing of Applications for Mobile Devices},
            author    = {Tim A. Majchrzak and
                         Matthias Schulte},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {27--39},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291390},
            urn       = {urn:nbn:de:101:1-201705291390},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Applications propel the versatility of mobile devices. Apps enable the realization of new ideas and greatly contribute to the proliferation of mobile computing. Unfortunately, software quality of apps often is low. This at least partly can be attributed to problems with testing them. However, it is not a lack of techniques or tools that make app testing cumbersome. Rather, frequent context changes have to be dealt with. Mobile devices most notably move: network parameters such as latency and usable bandwidth change, along with data read from sensors such as GPS coordinates. Additionally, usage patterns vary. To address context changes in testing, we propose a novel concept. It is based on identifying blocks of code between which context changes are possible. It helps to greatly reduce complexity. Besides introducing our concept, we present a use case, show its application and benefits, and discuss challenges.}
        }
    
  19.  Open Access 

    Perceived Sociability of Use and Individual Use of Social Networking Sites - A Field Study of Facebook Use in the Arctic

    Juhani Iivari

    Open Journal of Information Systems (OJIS), 1(1), Pages 23-53, 2014, Downloads: 2559, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194708 | GNL-LP: 1132360978 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper investigates determinants of individual use of social network sites (SNSs). It introduces a new construct, Perceived Sociability of Use (PSOU), to explain the use of such computer mediated communication applications. Based on a field study of 113 Facebook users it shows that PSOU in the sense of maintaining social contacts is a significant predictor of Perceived Benefits (PB), Perceived Enjoyment (PE), attitude toward use and intention to use. Inspired by Benbasat and Barki, this paper also attempts to answer questions "what makes the system useful", "what makes the system enjoyable to use" and "what makes the system sociable to use". As a consequence it pays special focus on systems characteristics of IT applications as potential predictors of PSOU, PB and PE, introducing seven such designable qualities (user-to-user interactivity, user identifiability, system quality, information quality, usability, user-to-system interactivity, and aesthetics). The results indicate that especially satisfaction with user-to-user interactivity is a significant determinant of PSOU, and that satisfactions with six of these seven designable qualities have significant paths in the proposed nomological network.

    BibTex:

        @Article{OJIS-v1i1n03_Iivari,
            title     = {Perceived Sociability of Use and Individual Use of Social Networking Sites - A Field Study of Facebook Use in the Arctic},
            author    = {Juhani Iivari},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {23--53},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194708},
            urn       = {urn:nbn:de:101:1-201705194708},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper investigates determinants of individual use of social network sites (SNSs). It introduces a new construct, Perceived Sociability of Use (PSOU), to explain the use of such computer mediated communication applications. Based on a field study of 113 Facebook users it shows that PSOU in the sense of maintaining social contacts is a significant predictor of Perceived Benefits (PB), Perceived Enjoyment (PE), attitude toward use and intention to use. Inspired by Benbasat and Barki, this paper also attempts to answer questions "what makes the system useful", "what makes the system enjoyable to use" and "what makes the system sociable to use". As a consequence it pays special focus on systems characteristics of IT applications as potential predictors of PSOU, PB and PE, introducing seven such designable qualities (user-to-user interactivity, user identifiability, system quality, information quality, usability, user-to-system interactivity, and aesthetics). The results indicate that especially satisfaction with user-to-user interactivity is a significant determinant of PSOU, and that satisfactions with six of these seven designable qualities have significant paths in the proposed nomological network.}
        }
    
  20.  Open Access 

    SIWeb: understanding the Interests of the Society through Web data Analysis

    Marco Furini, Simone Montangero

    Open Journal of Web Technologies (OJWT), 1(1), Pages 1-14, 2014, Downloads: 2059, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291334 | GNL-LP: 1133021522 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The high availability of user-generated contents in the Web scenario represents a tremendous asset for understanding various social phenomena. Methods and commercial products that exploit the widespread use of the Web as a way of conveying personal opinions have been proposed, but a critical thinking is that these approaches may produce a partial, or distorted, understanding of the society, because most of them focus on definite scenarios, use specific platforms, base their analysis on the sole magnitude of data, or treat the different Web resources with the same importance. In this paper, we present SIWeb (Social Interests through Web Analysis), a novel mechanism designed to measure the interest the society has on a topic (e.g., a real world phenomenon, an event, a person, a thing). SIWeb is general purpose (it can be applied to any decision making process), cross platforms (it uses the entire Webspace, from social media to websites, from tags to reviews), and time effective (it measures the time correlatio between the Web resources). It uses fractal analysis to detect the temporal relations behind all the Web resources (e.g., Web pages, RSS, newsgroups, etc.) that talk about a topic and combines this number with the temporal relations to give an insight of the the interest the society has about a topic. The evaluation of the proposal shows that SIWeb might be helpful in decision making processes as it reflects the interests the society has on a specific topic.

    BibTex:

        @Article{OJWT-v1i1n01_Furini,
            title     = {SIWeb: understanding the Interests of the Society through Web data Analysis},
            author    = {Marco Furini and
                         Simone Montangero},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {1--14},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291334},
            urn       = {urn:nbn:de:101:1-201705291334},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The high availability of user-generated contents in the Web scenario represents a tremendous asset for understanding various social phenomena. Methods and commercial products that exploit the widespread use of the Web as a way of conveying personal opinions have been proposed, but a critical thinking is that these approaches may produce a partial, or distorted, understanding of the society, because most of them focus on definite scenarios, use specific platforms, base their analysis on the sole magnitude of data, or treat the different Web resources with the same importance. In this paper, we present SIWeb (Social Interests through Web Analysis), a novel mechanism designed to measure the interest the society has on a topic (e.g., a real world phenomenon, an event, a person, a thing). SIWeb is general purpose (it can be applied to any decision making process), cross platforms (it uses the entire Webspace, from social media to websites, from tags to reviews), and time effective (it measures the time correlatio between the Web resources). It uses fractal analysis to detect the temporal relations behind all the Web resources (e.g., Web pages, RSS, newsgroups, etc.) that talk about a topic and combines this number with the temporal relations to give an insight of the the interest the society has about a topic. The evaluation of the proposal shows that SIWeb might be helpful in decision making processes as it reflects the interests the society has on a specific topic.}
        }
    
  21.  Open Access 

    Causal Consistent Databases

    Mawahib Musa Elbushra, Jan Lindström

    Open Journal of Databases (OJDB), 2(1), Pages 17-35, 2015, Downloads: 3713, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194619 | GNL-LP: 1132360870 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Many consistency criteria have been considered in databases and the causal consistency is one of them. The causal consistency model has gained much attention in recent years because it provides ordering of relative operations. The causal consistency requires that all writes, which are potentially causally related, must be seen in the same order by all processes. The causal consistency is a weaker criteria than the sequential consistency, because there exists an execution, which is causally consistent but not sequentially consistent, however all executions satisfying the sequential consistency are also causally consistent. Furthermore, the causal consistency supports non-blocking operations; i.e. processes may complete read or write operations without waiting for global computation. Therefore, the causal consistency overcomes the primary limit of stronger criteria: communication latency. Additionally, several application semantics are precisely captured by the causal consistency, e.g. collaborative tools. In this paper, we review the state-of-the-art of causal consistent databases, discuss the features, functionalities and applications of the causal consistency model, and systematically compare it with other consistency models. We also discuss the implementation of causal consistency databases and identify limitations of the causal consistency model.

    BibTex:

        @Article{OJDB_2015v2i1n02_Elbushra,
            title     = {Causal Consistent Databases},
            author    = {Mawahib Musa Elbushra and
                         Jan Lindstr{\"o}m},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {17--35},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194619},
            urn       = {urn:nbn:de:101:1-201705194619},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Many consistency criteria have been considered in databases and the causal consistency is one of them. The causal consistency model has gained much attention in recent years because it provides ordering of relative operations. The causal consistency requires that all writes, which are potentially causally related, must be seen in the same order by all processes. The causal consistency is a weaker criteria than the sequential consistency, because there exists an execution, which is causally consistent but not sequentially consistent, however all executions satisfying the sequential consistency are also causally consistent. Furthermore, the causal consistency supports non-blocking operations; i.e. processes may complete read or write operations without waiting for global computation. Therefore, the causal consistency overcomes the primary limit of stronger criteria: communication latency. Additionally, several application semantics are precisely captured by the causal consistency, e.g. collaborative tools. In this paper, we review the state-of-the-art of causal consistent databases, discuss the features, functionalities and applications of the causal consistency model, and systematically compare it with other consistency models. We also discuss the implementation of causal consistency databases and identify limitations of the causal consistency model.}
        }
    
  22.  Open Access 

    An Analytical Model of Multi-Core Multi-Cluster Architecture (MCMCA)

    Norhazlina Hamid, Robert John Walters, Gary B. Wills

    Open Journal of Cloud Computing (OJCC), 2(1), Pages 4-15, 2015, Downloads: 1928, Citations: 3

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194487 | GNL-LP: 1132360692 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Multi-core clusters have emerged as an important contribution in computing technology for provisioning additional processing power in high performance computing and communications. Multi-core architectures are proposed for their capability to provide higher performance without increasing heat and power usage, which is the main concern in a single-core processor. This paper introduces analytical models of a new architecture for large-scale multi-core clusters to improve the communication performance within the interconnection network. The new architecture will be based on a multi - cluster architecture containing clusters of multi-core processors.

    BibTex:

        @Article{OJCC_2015v2i1n02_Hamid,
            title     = {An Analytical Model of Multi-Core Multi-Cluster Architecture (MCMCA)},
            author    = {Norhazlina Hamid and
                         Robert John Walters and
                         Gary B. Wills},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194487},
            urn       = {urn:nbn:de:101:1-201705194487},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Multi-core clusters have emerged as an important contribution in computing technology for provisioning additional processing power in high performance computing and communications. Multi-core architectures are proposed for their capability to provide higher performance without increasing heat and power usage, which is the main concern in a single-core processor. This paper introduces analytical models of a new architecture for large-scale multi-core clusters to improve the communication performance within the interconnection network. The new architecture will be based on a multi - cluster architecture containing clusters of multi-core processors.}
        }
    
  23.  Open Access 

    Sensing as a Service: Secure Wireless Sensor Network Infrastructure Sharing for the Internet of Things

    Cintia B. Margi, Renan C. A. Alves, Johanna Sepulveda

    Open Journal of Internet Of Things (OJIOT), 3(1), Pages 91-102, 2017, Downloads: 569, Citations: 3

    Special Issue: Proceedings of the International Workshop on Very Large Internet of Things (VLIoT 2017) in conjunction with the VLDB 2017 Conference in Munich, Germany.

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017080613467 | GNL-LP: 1137820209 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Internet of Things (IoT) andWireless Sensor Networks (WSN) are composed of devices capable of sensing/actuation, communication and processing. They are valuable technology for the development of applications in several areas, such as environmental, industrial and urban monitoring and processes controlling. Given the challenges of different protocols and technologies used for communication, resource constrained devices nature, high connectivity and security requirements for the applications, the main challenges that need to be addressed include: secure communication between IoT devices, network resource management and the protected implementation of the security mechanisms. In this paper, we present a secure Software-Defined Networking (SDN) based framework that includes: communication protocols, node task programming middleware, communication and computation resource management features and security services. The communication layer for the constrained devices considers IT-SDN as its basis. Concerning security, we address the main services, the type of algorithms to achieve them, and why their secure implementation is needed. Lastly, we showcase how the Sensing as a Service paradigm could enable WSN usage in more environments.

    BibTex:

        @Article{OJIOT_2017v3i1n08_Margi,
            title     = {Sensing as a Service: Secure Wireless Sensor Network Infrastructure Sharing for the Internet of Things},
            author    = {Cintia B. Margi and
                         Renan C. A. Alves and
                         Johanna Sepulveda},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2017},
            volume    = {3},
            number    = {1},
            pages     = {91--102},
            note      = {Special Issue: Proceedings of the International Workshop on Very Large Internet of Things (VLIoT 2017) in conjunction with the VLDB 2017 Conference in Munich, Germany.},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017080613467},
            urn       = {urn:nbn:de:101:1-2017080613467},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Internet of Things (IoT) andWireless Sensor Networks (WSN) are composed of devices capable of sensing/actuation, communication and processing. They are valuable technology for the development of applications in several areas, such as environmental, industrial and urban monitoring and processes controlling. Given the challenges of different protocols and technologies used for communication, resource constrained devices nature, high connectivity and security requirements for the applications, the main challenges that need to be addressed include: secure communication between IoT devices, network resource management and the protected implementation of the security mechanisms. In this paper, we present a secure Software-Defined Networking (SDN) based framework that includes: communication protocols, node task programming middleware, communication and computation resource management features and security services. The communication layer for the constrained devices considers IT-SDN as its basis. Concerning security, we address the main services, the type of algorithms to achieve them, and why their secure implementation is needed. Lastly, we showcase how the Sensing as a Service paradigm could enable WSN usage in more environments.}
        }
    
  24.  Open Access 

    Semantic Blockchain to Improve Scalability in the Internet of Things

    Michele Ruta, Floriano Scioscia, Saverio Ieva, Giovanna Capurso, Eugenio Di Sciascio

    Open Journal of Internet Of Things (OJIOT), 3(1), Pages 46-61, 2017, Downloads: 1575, Citations: 3

    Special Issue: Proceedings of the International Workshop on Very Large Internet of Things (VLIoT 2017) in conjunction with the VLDB 2017 Conference in Munich, Germany.

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017080613488 | GNL-LP: 1137820225 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Generally scarce computational and memory resource availability is a well known problem for the IoT, whose intrinsic volatility makes complex applications unfeasible. Noteworthy efforts in overcoming unpredictability (particularly in case of large dimensions) are the ones integrating Knowledge Representation technologies to build the so-called Semantic Web of Things (SWoT). In spite of allowed advanced discovery features, transactions in the SWoT still suffer from not viable trust management strategies. Given its intrinsic characteristics, blockchain technology appears as interesting from this perspective: a semantic resource/service discovery layer built upon a basic blockchain infrastructure gains a consensus validation. This paper proposes a novel Service-Oriented Architecture (SOA) based on a semantic blockchain for registration, discovery, selection and payment. Such operations are implemented as smart contracts, allowing distributed execution and trust. Reported experiments early assess the sustainability of the proposal.

    BibTex:

        @Article{OJIOT_2017v3i1n05_Ruta,
            title     = {Semantic Blockchain to Improve Scalability in the Internet of Things},
            author    = {Michele Ruta and
                         Floriano Scioscia and
                         Saverio Ieva and
                         Giovanna Capurso and
                         Eugenio Di Sciascio},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2017},
            volume    = {3},
            number    = {1},
            pages     = {46--61},
            note      = {Special Issue: Proceedings of the International Workshop on Very Large Internet of Things (VLIoT 2017) in conjunction with the VLDB 2017 Conference in Munich, Germany.},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017080613488},
            urn       = {urn:nbn:de:101:1-2017080613488},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Generally scarce computational and memory resource availability is a well known problem for the IoT, whose intrinsic volatility makes complex applications unfeasible. Noteworthy efforts in overcoming unpredictability (particularly in case of large dimensions) are the ones integrating Knowledge Representation technologies to build the so-called Semantic Web of Things (SWoT). In spite of allowed advanced discovery features, transactions in the SWoT still suffer from not viable trust management strategies. Given its intrinsic characteristics, blockchain technology appears as interesting from this perspective: a semantic resource/service discovery layer built upon a basic blockchain infrastructure gains a consensus validation. This paper proposes a novel Service-Oriented Architecture (SOA) based on a semantic blockchain for registration, discovery, selection and payment. Such operations are implemented as smart contracts, allowing distributed execution and trust. Reported experiments early assess the sustainability of the proposal.}
        }
    
  25.  Open Access 

    Pattern-sensitive Time-series Anonymization and its Application to Energy-Consumption Data

    Stephan Kessler, Erik Buchmann, Thorben Burghardt, Klemens Böhm

    Open Journal of Information Systems (OJIS), 1(1), Pages 3-22, 2014, Downloads: 2648, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194696 | GNL-LP: 113236096X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Time series anonymization is an important problem. One prominent example of time series are energy consumption records, which might reveal details of the daily routine of a household. Existing privacy approaches for time series, e.g., from the field of trajectory anonymization, assume that every single value of a time series contains sensitive information and reduce the data quality very much. In contrast, we consider time series where it is combinations of tuples that represent personal information. We propose (n; l; k)-anonymity, geared to anonymization of time-series data with minimal information loss, assuming that an adversary may learn a few data points. We propose several heuristics to obtain (n; l; k)-anonymity, and we evaluate our approach both with synthetic and real data. Our experiments confirm that it is sufficient to modify time series only moderately in order to fulfill meaningful privacy requirements.

    BibTex:

        @Article{OJIS-v1i1n02_Kessler,
            title     = {Pattern-sensitive Time-series Anonymization and its Application to Energy-Consumption Data},
            author    = {Stephan Kessler and
                         Erik Buchmann and
                         Thorben Burghardt and
                         Klemens B{\"o}hm},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {3--22},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194696},
            urn       = {urn:nbn:de:101:1-201705194696},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Time series anonymization is an important problem. One prominent example of time series are energy consumption records, which might reveal details of the daily routine of a household. Existing privacy approaches for time series, e.g., from the field of trajectory anonymization, assume that every single value of a time series contains sensitive information and reduce the data quality very much. In contrast, we consider time series where it is combinations of tuples that represent personal information. We propose (n; l; k)-anonymity, geared to anonymization of time-series data with minimal information loss, assuming that an adversary may learn a few data points. We propose several heuristics to obtain (n; l; k)-anonymity, and we evaluate our approach both with synthetic and real data. Our experiments confirm that it is sufficient to modify time series only moderately in order to fulfill meaningful privacy requirements.}
        }
    
  26.  Open Access 

    Evaluation of Node Failures in Cloud Computing Using Empirical Data

    Abdulelah Alwabel, Robert John Walters, Gary B. Wills

    Open Journal of Cloud Computing (OJCC), 1(2), Pages 15-24, 2014, Downloads: 2172, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194435 | GNL-LP: 1132360609 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Cloud has emerged as a new computing paradigm that promises to move into computing-as-utility era. Desktop Cloud is a new type of Cloud computing introduced to further achieve this ambition with an aim to reduce costs. It merges two computing models: Cloud computing and volunteer computing. The aim of Desktop Cloud is to provide Cloud services out of infrastructure that is not made for this purpose, like PCs and laptops. Such computing resources lead to a high level of volatility as a result of the fact that they can leave without prior knowledge. This paper studies the impact of node failures using evaluation metrics based on real data collected from public archive to simulate failure events in the infrastructure of a Desktop Cloud. The contribution of this paper is: (i) analysing the failure events, (ii) proposing metrics to evaluate Desktop Clouds, and (iii) evaluating several VM allocation mechanisms in the presence of node failures.

    BibTex:

        @Article{OJCC-2014v1i2n02_Alwabel,
            title     = {Evaluation of Node Failures in Cloud Computing Using Empirical Data},
            author    = {Abdulelah Alwabel and
                         Robert John Walters and
                         Gary B. Wills},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {15--24},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194435},
            urn       = {urn:nbn:de:101:1-201705194435},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Cloud has emerged as a new computing paradigm that promises to move into computing-as-utility era. Desktop Cloud is a new type of Cloud computing introduced to further achieve this ambition with an aim to reduce costs. It merges two computing models: Cloud computing and volunteer computing. The aim of Desktop Cloud is to provide Cloud services out of infrastructure that is not made for this purpose, like PCs and laptops. Such computing resources lead to a high level of volatility as a result of the fact that they can leave without prior knowledge. This paper studies the impact of node failures using evaluation metrics based on real data collected from public archive to simulate failure events in the infrastructure of a Desktop Cloud. The contribution of this paper is: (i) analysing the failure events, (ii) proposing metrics to evaluate Desktop Clouds, and (iii) evaluating several VM allocation mechanisms in the presence of node failures.}
        }
    
  27.  Open Access 

    Concept Design for Creating Essential Hypothesis, Rules, and Goals: Toward a Data Marketplace

    Jun Nakamura, Masahiko Teramoto

    Open Journal of Information Systems (OJIS), 2(2), Pages 16-26, 2015, Downloads: 1619, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194774 | GNL-LP: 1132361117 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The abductive reasoning model has been discussed in the context of business strategy. However, this model seems unrealistic for applications in the real business world considering the unpredictable, competitive business environment. This study improves the model by formulating an experimental case study through a web-based workplace for generating product ideas. We discuss the possible embodiment of product ideas as the basis for configuring features through the use of dynamic quality function deployment. The entire concept design process is proposed as a blueprint for building a data marketplace.

    BibTex:

        @Article{OJIS_2015v2i2n03_Nakamura,
            title     = {Concept Design for Creating Essential Hypothesis, Rules, and Goals: Toward a Data Marketplace},
            author    = {Jun Nakamura and
                         Masahiko Teramoto},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {16--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194774},
            urn       = {urn:nbn:de:101:1-201705194774},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The abductive reasoning model has been discussed in the context of business strategy. However, this model seems unrealistic for applications in the real business world considering the unpredictable, competitive business environment. This study improves the model by formulating an experimental case study through a web-based workplace for generating product ideas. We discuss the possible embodiment of product ideas as the basis for configuring features through the use of dynamic quality function deployment. The entire concept design process is proposed as a blueprint for building a data marketplace.}
        }
    
  28.  Open Access 

    BEAUFORD: A Benchmark for Evaluation of Formalisation of Definitions in OWL

    Cheikh Kacfah Emani, Catarina Ferreira Da Silva, Bruno Fiés, Parisa Ghodous

    Open Journal of Semantic Web (OJSW), 2(1), Pages 4-15, 2015, Downloads: 1187, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194879 | GNL-LP: 1132361257 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In this paper we present BEAUFORD, a benchmark for methods which aim to provide formal expressions of concepts using the natural language (NL) definition of these concepts. Adding formal expressions of concepts to a given ontology allows reasoners to infer more useful pieces of information or to detect inconsistencies in this given ontology. To the best of our knowledge, BEAUFORD is the first benchmark to tackle this ontology enrichment problem. BEAUFORD allows the breaking down of a given formalisation approach by identifying its key features. In addition, BEAUFORD provides strong mechanisms to evaluate efficiently an approach even in case of ambiguity which is a major challenge in formalisation of NL resources. Indeed, BEAUFORD takes into account the fact that a given NL phrase can be formalised in many ways. Hence, it proposes a suitable specification to represent these multiple formalisations. Taking advantage of this specification, BEAUFORD redefines classical precision and recall and introduces other metrics to take into account the fact that there is not only one unique way to formalise a definition. Finally, BEAUFORD comprises a well-suited dataset to concretely judge of the efficiency of methods of formalisation. Using BEAUFORD, current approaches of formalisation of definitions can be compared accurately using a suitable gold standard.

    BibTex:

        @Article{OJSW_2015v2i1n02_Kachfah,
            title     = {BEAUFORD: A Benchmark for Evaluation of Formalisation of Definitions in OWL},
            author    = {Cheikh Kacfah Emani and
                         Catarina Ferreira Da Silva and
                         Bruno Fi\'{e}s and
                         Parisa Ghodous},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194879},
            urn       = {urn:nbn:de:101:1-201705194879},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In this paper we present BEAUFORD, a benchmark for methods which aim to provide formal expressions of concepts using the natural language (NL) definition of these concepts. Adding formal expressions of concepts to a given ontology allows reasoners to infer more useful pieces of information or to detect inconsistencies in this given ontology. To the best of our knowledge, BEAUFORD is the first benchmark to tackle this ontology enrichment problem. BEAUFORD allows the breaking down of a given formalisation approach by identifying its key features. In addition, BEAUFORD provides strong mechanisms to evaluate efficiently an approach even in case of ambiguity which is a major challenge in formalisation of NL resources. Indeed, BEAUFORD takes into account the fact that a given NL phrase can be formalised in many ways. Hence, it proposes a suitable specification to represent these multiple formalisations. Taking advantage of this specification, BEAUFORD redefines classical precision and recall and introduces other metrics to take into account the fact that there is not only one unique way to formalise a definition. Finally, BEAUFORD comprises a well-suited dataset to concretely judge of the efficiency of methods of formalisation. Using BEAUFORD, current approaches of formalisation of definitions can be compared accurately using a suitable gold standard.}
        }
    
  29.  Open Access 

    Big Data in the Cloud: A Survey

    Pedro Caldeira Neves, Jorge Bernardino

    Open Journal of Big Data (OJBD), 1(2), Pages 1-18, 2015, Downloads: 3316, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194365 | GNL-LP: 1132360528 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Big Data has become a hot topic across several business areas requiring the storage and processing of huge volumes of data. Cloud computing leverages Big Data by providing high storage and processing capabilities and enables corporations to consume resources in a pay-as-you-go model making clouds the optimal environment for storing and processing huge quantities of data. By using virtualized resources, Cloud can scale very easily, be highly available and provide massive storage capacity and processing power. This paper surveys existing databases models to store and process Big Data within a Cloud environment. Particularly, we detail the following traditional NoSQL databases: BigTable, Cassandra, DynamoDB, HBase, Hypertable, and MongoDB. The MapReduce framework and its developments Apache Spark, HaLoop, Twister, and other alternatives such as Apache Giraph, GraphLab, Pregel and MapD - a novel platform that uses GPU processing to accelerate Big Data processing - are also analyzed. Finally, we present two case studies that demonstrate the successful use of Big Data within Cloud environments and the challenges that must be addressed in the future.

    BibTex:

        @Article{OJBD_2015v1i2n02_Neves,
            title     = {Big Data in the Cloud: A Survey},
            author    = {Pedro Caldeira Neves and
                         Jorge Bernardino},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194365},
            urn       = {urn:nbn:de:101:1-201705194365},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Big Data has become a hot topic across several business areas requiring the storage and processing of huge volumes of data. Cloud computing leverages Big Data by providing high storage and processing capabilities and enables corporations to consume resources in a pay-as-you-go model making clouds the optimal environment for storing and processing huge quantities of data. By using virtualized resources, Cloud can scale very easily, be highly available and provide massive storage capacity and processing power. This paper surveys existing databases models to store and process Big Data within a Cloud environment. Particularly, we detail the following traditional NoSQL databases: BigTable, Cassandra, DynamoDB, HBase, Hypertable, and MongoDB. The MapReduce framework and its developments Apache Spark, HaLoop, Twister, and other alternatives such as Apache Giraph, GraphLab, Pregel and MapD - a novel platform that uses GPU processing to accelerate Big Data processing - are also analyzed. Finally, we present two case studies that demonstrate the successful use of Big Data within Cloud environments and the challenges that must be addressed in the future.}
        }
    
  30.  Open Access 

    Cooperative Hybrid Cloud Intermediaries - Making Cloud Sourcing Feasible for Small and Medium-sized Enterprises

    Till Haselmann, Gottfried Vossen, Stuart Dillon

    Open Journal of Cloud Computing (OJCC), 2(2), Pages 4-20, 2015, Downloads: 970, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194494 | GNL-LP: 1132360714 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: "The cloud" is widely advertised as a silver bullet for many IT-related challenges of small and medium-sized enterprises (SMEs). While it can potentially have a number of attractive benefits, many SMEs refrain from using cloud sourcing and cloud services because of high upfront costs for building the appropriate knowledge in the enterprise, for searching and screening of possible cloud service providers, and for mastering the intricate legal issues related to outsourcing sensitive data. This paper presents the concept of hybrid cloud intermediaries, an approach that can address many of the prevailing issues. With the aid of empirical findings from a cross-nation study of cloud adoption in SMEs for context, we describe the concept in detail and show conceivable variants, including a comprehensive cross-perspective consolidated model of cloud intermediary value-creation. Subsequently, we analyze the benefits of such a hybrid cloud intermediary for addressing cloud adoption issues in SMEs, and suggest suitable governance structures based on the cooperative paradigm. The resulting entity - a cooperative hybrid cloud intermediary or, more concisely, co-op cloud - is discussed in detail showing both feasible scenarios and limitations for SMEs that would like to engage in a cloud-sourcing.

    BibTex:

        @Article{OJCC_2015v2i2n02_Haselmann,
            title     = {Cooperative Hybrid Cloud Intermediaries - Making Cloud Sourcing Feasible for Small and Medium-sized Enterprises},
            author    = {Till Haselmann and
                         Gottfried Vossen and
                         Stuart Dillon},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {4--20},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194494},
            urn       = {urn:nbn:de:101:1-201705194494},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {"The cloud" is widely advertised as a silver bullet for many IT-related challenges of small and medium-sized enterprises (SMEs). While it can potentially have a number of attractive benefits, many SMEs refrain from using cloud sourcing and cloud services because of high upfront costs for building the appropriate knowledge in the enterprise, for searching and screening of possible cloud service providers, and for mastering the intricate legal issues related to outsourcing sensitive data. This paper presents the concept of hybrid cloud intermediaries, an approach that can address many of the prevailing issues. With the aid of empirical findings from a cross-nation study of cloud adoption in SMEs for context, we describe the concept in detail and show conceivable variants, including a comprehensive cross-perspective consolidated model of cloud intermediary value-creation. Subsequently, we analyze the benefits of such a hybrid cloud intermediary for addressing cloud adoption issues in SMEs, and suggest suitable governance structures based on the cooperative paradigm. The resulting entity - a cooperative hybrid cloud intermediary or, more concisely, co-op cloud - is discussed in detail showing both feasible scenarios and limitations for SMEs that would like to engage in a cloud-sourcing.}
        }
    
  31.  Open Access 

    Detecting Vital Documents in Massive Data Streams

    Shun Kawahara, Kazuhiro Seki, Kuniaki Uehara

    Open Journal of Web Technologies (OJWT), 2(1), Pages 16-26, 2015, Downloads: 1330, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291373 | GNL-LP: 113302159X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Existing knowledge bases, includingWikipedia, are typically written and maintained by a group of voluntary editors. Meanwhile, numerous web documents are being published partly due to the popularization of online news and social media. Some of the web documents, called "vital documents", contain novel information that should be taken into account in updating articles of the knowledge bases. However, it is practically impossible for the editors to manually monitor all the relevant web documents. Consequently, there is a considerable time lag between an edit to knowledge base and the publication dates of such vital documents. This paper proposes a realtime detection framework of web documents containing novel information flowing in massive document streams. The framework consists of twostep filter using statistical language models. Further, the framework is implemented on the distributed and faulttolerant realtime computation system, Apache Storm, in order to process the large number of web documents. On a publicly available web document data set, the TREC KBA Stream Corpus, the validity of the proposed framework is demonstrated in terms of the detection performance and processing time.

    BibTex:

        @Article{OJWT_2015v2i1n03_Kawahara,
            title     = {Detecting Vital Documents in Massive Data Streams},
            author    = {Shun Kawahara and
                         Kazuhiro Seki and
                         Kuniaki Uehara},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {16--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291373},
            urn       = {urn:nbn:de:101:1-201705291373},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Existing knowledge bases, includingWikipedia, are typically written and maintained by a group of voluntary editors. Meanwhile, numerous web documents are being published partly due to the popularization of online news and social media. Some of the web documents, called "vital documents", contain novel information that should be taken into account in updating articles of the knowledge bases. However, it is practically impossible for the editors to manually monitor all the relevant web documents. Consequently, there is a considerable time lag between an edit to knowledge base and the publication dates of such vital documents. This paper proposes a realtime detection framework of web documents containing novel information flowing in massive document streams. The framework consists of twostep filter using statistical language models. Further, the framework is implemented on the distributed and faulttolerant realtime computation system, Apache Storm, in order to process the large number of web documents. On a publicly available web document data set, the TREC KBA Stream Corpus, the validity of the proposed framework is demonstrated in terms of the detection performance and processing time.}
        }
    
  32.  Open Access 

    Statistical Machine Learning in Brain State Classification using EEG Data

    Yuezhe Li, Yuchou Chang, Hong Lin

    Open Journal of Big Data (OJBD), 1(2), Pages 19-33, 2015, Downloads: 1826, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194354 | GNL-LP: 113236051X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In this article, we discuss how to use a variety of machine learning methods, e.g. tree bagging, random forest, boost, support vector machine, and Gaussian mixture model, for building classifiers for electroencephalogram (EEG) data, which is collected from different brain states on different subjects. Also, we discuss how training data size influences misclassification rate. Moreover, the number of subjects that contributes to the training data affects misclassification rate. Furthermore, we discuss how sample entropy contributes to building a classifier. Our results show that classification based on sample entropy give the smallest misclassification rate. Moreover, two data sets were collected from one channel and seven channels respectively. The classification results of each data set show that the more channels we use, the less misclassification we have. Our results show that it is promising to build a self-adaptive classification system by using EEG data to distinguish idle from active state.

    BibTex:

        @Article{OJBD_2015v1i2n03_YuehzeLi,
            title     = {Statistical Machine Learning in Brain State Classification using EEG Data},
            author    = {Yuezhe Li and
                         Yuchou Chang and
                         Hong Lin},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {19--33},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194354},
            urn       = {urn:nbn:de:101:1-201705194354},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In this article, we discuss how to use a variety of machine learning methods, e.g. tree bagging, random forest, boost, support vector machine, and Gaussian mixture model, for building classifiers for electroencephalogram (EEG) data, which is collected from different brain states on different subjects. Also, we discuss how training data size influences misclassification rate. Moreover, the number of subjects that contributes to the training data affects misclassification rate. Furthermore, we discuss how sample entropy contributes to building a classifier. Our results show that classification based on sample entropy give the smallest misclassification rate. Moreover, two data sets were collected from one channel and seven channels respectively. The classification results of each data set show that the more channels we use, the less misclassification we have. Our results show that it is promising to build a self-adaptive classification system by using EEG data to distinguish idle from active state.}
        }
    
  33.  Open Access 

    Data Transfers in Hadoop: A Comparative Study

    Ujjal Marjit, Kumar Sharma, Puspendu Mandal

    Open Journal of Big Data (OJBD), 1(2), Pages 34-46, 2015, Downloads: 2528, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194373 | GNL-LP: 1132360536 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Hadoop is an open source framework for processing large amounts of data in distributed computing environment. It plays an important role in processing and analyzing the Big Data. This framework is used for storing data on large clusters of commodity hardware. Data input and output to and from Hadoop is an indispensable action for any data processing job. At present, many tools have been evolved for importing and exporting Data in Hadoop. In this article, some commonly used tools for importing and exporting data have been emphasized. Moreover, a state-of-the-art comparative study among the various tools has been made. With this study, it has been decided that where to use one tool over the other with emphasis on the data transfer to and from Hadoop system. This article also discusses about how Hadoop handles backup and disaster recovery along with some open research questions in terms of Big Data transfer when dealing with cloud-based services.

    BibTex:

        @Article{OJBD_2015v1i2n04_UjjalMarjit,
            title     = {Data Transfers in Hadoop: A Comparative Study},
            author    = {Ujjal Marjit and
                         Kumar Sharma and
                         Puspendu Mandal},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {34--46},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194373},
            urn       = {urn:nbn:de:101:1-201705194373},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Hadoop is an open source framework for processing large amounts of data in distributed computing environment. It plays an important role in processing and analyzing the Big Data. This framework is used for storing data on large clusters of commodity hardware. Data input and output to and from Hadoop is an indispensable action for any data processing job. At present, many tools have been evolved for importing and exporting Data in Hadoop. In this article, some commonly used tools for importing and exporting data have been emphasized. Moreover, a state-of-the-art comparative study among the various tools has been made. With this study, it has been decided that where to use one tool over the other with emphasis on the data transfer to and from Hadoop system. This article also discusses about how Hadoop handles backup and disaster recovery along with some open research questions in terms of Big Data transfer when dealing with cloud-based services.}
        }
    
  34.  Open Access 

    A Semantic Question Answering Framework for Large Data Sets

    Marta Tatu, Mithun Balakrishna, Steven Werner, Tatiana Erekhinskaya, Dan Moldovan

    Open Journal of Semantic Web (OJSW), 3(1), Pages 16-31, 2016, Downloads: 2099, Citations: 2

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194921 | GNL-LP: 1132361338 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Traditionally, the task of answering natural language questions has involved a keyword-based document retrieval step, followed by in-depth processing of candidate answer documents and paragraphs. This post-processing uses semantics to various degrees. In this article, we describe a purely semantic question answering (QA) framework for large document collections. Our high-precision approach transforms the semantic knowledge extracted from natural language texts into a language-agnostic RDF representation and indexes it into a scalable triplestore. In order to facilitate easy access to the information stored in the RDF semantic index, a user's natural language questions are translated into SPARQL queries that return precise answers back to the user. The robustness of this framework is ensured by the natural language reasoning performed on the RDF store, by the query relaxation procedures, and the answer ranking techniques. The improvements in performance over a regular free text search index-based question answering engine prove that QA systems can benefit greatly from the addition and consumption of deep semantic information.

    BibTex:

        @Article{OJSW_2016v3i1n02_Tatu,
            title     = {A Semantic Question Answering Framework for Large Data Sets},
            author    = {Marta Tatu and
                         Mithun Balakrishna and
                         Steven Werner and
                         Tatiana Erekhinskaya and
                         Dan Moldovan},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2016},
            volume    = {3},
            number    = {1},
            pages     = {16--31},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194921},
            urn       = {urn:nbn:de:101:1-201705194921},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Traditionally, the task of answering natural language questions has involved a keyword-based document retrieval step, followed by in-depth processing of candidate answer documents and paragraphs. This post-processing uses semantics to various degrees. In this article, we describe a purely semantic question answering (QA) framework for large document collections. Our high-precision approach transforms the semantic knowledge extracted from natural language texts into a language-agnostic RDF representation and indexes it into a scalable triplestore. In order to facilitate easy access to the information stored in the RDF semantic index, a user's natural language questions are translated into SPARQL queries that return precise answers back to the user. The robustness of this framework is ensured by the natural language reasoning performed on the RDF store, by the query relaxation procedures, and the answer ranking techniques. The improvements in performance over a regular free text search index-based question answering engine prove that QA systems can benefit greatly from the addition and consumption of deep semantic information.}
        }
    
  35.  Open Access 

    Integrating Human Factors and Semantic Mark-ups in Adaptive Interactive Systems

    Marios Belk, Panagiotis Germanakos, Efi Papatheocharous, Panayiotis Andreou, George Samaras

    Open Journal of Web Technologies (OJWT), 1(1), Pages 15-26, 2014, Downloads: 1919, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-2017052611313 | GNL-LP: 113283600X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper focuses on incorporating individual differences in cognitive processing and semantic mark-ups in the context of adaptive interactive systems. In particular, a semantic Web-based adaptation framework is proposed that enables Web content providers to enrich content and functionality of Web environments with semantic mark-ups. The Web content is created using a Web authoring tool and is further processed and reconstructed by an adaptation mechanism based on cognitive factors of users. Main aim of this work is to investigate the added value of personalising content and functionality of Web environments based on the unique cognitive characteristics of users. Accordingly, a user study has been conducted that entailed a psychometric-based survey for extracting the users' cognitive characteristics, combined with a real usage scenario of an existing commercial Web environment that was enriched with semantic mark-ups and personalised based on different adaptation effects. The paper provides interesting insights in the design and development of adaptive interactive systems based on cognitive factors and semantic mark-ups.

    BibTex:

        @Article{OJWT-v1i1n02_Belk,
            title     = {Integrating Human Factors and Semantic Mark-ups in Adaptive Interactive Systems},
            author    = {Marios Belk and
                         Panagiotis Germanakos and
                         Efi Papatheocharous and
                         Panayiotis Andreou and
                         George Samaras},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {1},
            pages     = {15--26},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-2017052611313},
            urn       = {urn:nbn:de:101:1-2017052611313},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper focuses on incorporating individual differences in cognitive processing and semantic mark-ups in the context of adaptive interactive systems. In particular, a semantic Web-based adaptation framework is proposed that enables Web content providers to enrich content and functionality of Web environments with semantic mark-ups. The Web content is created using a Web authoring tool and is further processed and reconstructed by an adaptation mechanism based on cognitive factors of users. Main aim of this work is to investigate the added value of personalising content and functionality of Web environments based on the unique cognitive characteristics of users. Accordingly, a user study has been conducted that entailed a psychometric-based survey for extracting the users' cognitive characteristics, combined with a real usage scenario of an existing commercial Web environment that was enriched with semantic mark-ups and personalised based on different adaptation effects. The paper provides interesting insights in the design and development of adaptive interactive systems based on cognitive factors and semantic mark-ups.}
        }
    
  36.  Open Access 

    P-LUPOSDATE: Using Precomputed Bloom Filters to Speed Up SPARQL Processing in the Cloud

    Sven Groppe, Thomas Kiencke, Stefan Werner, Dennis Heinrich, Marc Stelzner, Le Gruenwald

    Open Journal of Semantic Web (OJSW), 1(2), Pages 25-55, 2014, Downloads: 2632, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194858 | GNL-LP: 1132361214 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: Increasingly data on the Web is stored in the form of Semantic Web data. Because of today's information overload, it becomes very important to store and query these big datasets in a scalable way and hence in a distributed fashion. Cloud Computing offers such a distributed environment with dynamic reallocation of computing and storing resources based on needs. In this work we introduce a scalable distributed Semantic Web database in the Cloud. In order to reduce the number of (unnecessary) intermediate results early, we apply bloom filters. Instead of computing bloom filters, a time-consuming task during query processing as it has been done traditionally, we precompute the bloom filters as much as possible and store them in the indices besides the data. The experimental results with data sets up to 1 billion triples show that our approach speeds up query processing significantly and sometimes even reduces the processing time to less than half.

    BibTex:

        @Article{OJSW-v1i2n02_Groppe,
            title     = {P-LUPOSDATE: Using Precomputed Bloom Filters to Speed Up SPARQL Processing in the Cloud},
            author    = {Sven Groppe and
                         Thomas Kiencke and
                         Stefan Werner and
                         Dennis Heinrich and
                         Marc Stelzner and
                         Le Gruenwald},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {25--55},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194858},
            urn       = {urn:nbn:de:101:1-201705194858},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Increasingly data on the Web is stored in the form of Semantic Web data. Because of today's information overload, it becomes very important to store and query these big datasets in a scalable way and hence in a distributed fashion. Cloud Computing offers such a distributed environment with dynamic reallocation of computing and storing resources based on needs. In this work we introduce a scalable distributed Semantic Web database in the Cloud. In order to reduce the number of (unnecessary) intermediate results early, we apply bloom filters. Instead of computing bloom filters, a time-consuming task during query processing as it has been done traditionally, we precompute the bloom filters as much as possible and store them in the indices besides the data. The experimental results with data sets up to 1 billion triples show that our approach speeds up query processing significantly and sometimes even reduces the processing time to less than half.}
        }
    
  37.  Open Access 

    A Comparative Evaluation of Current HTML5 Web Video Implementations

    Martin Hoernig, Andreas Bigontina, Bernd Radig

    Open Journal of Web Technologies (OJWT), 1(2), Pages 1-9, 2014, Downloads: 12687, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291328 | GNL-LP: 1133021514 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: HTML5 video is the upcoming standard for playing videos on the World Wide Web. Although its specification has not been fully adopted yet, all major browsers provide the HTML5 video element and web developers already rely on its functionality. But there are differences between implementations and inaccuracies that trouble the web developer community. To help to improve the current situation we draw a comparison between the most important web browsers. We focus on the event mechanism, since it is essential for interacting with the video element. Furthermore, we compare the seeking accuracy, which is relevant for more specialized applications. Our tests reveal varieties of differences between browser interfaces and show that even simple software solutions may still need third-party plugins in today's browsers.

    BibTex:

        @Article{OJWT-v1i2n01_Hoernig,
            title     = {A Comparative Evaluation of Current HTML5 Web Video Implementations},
            author    = {Martin Hoernig and
                         Andreas Bigontina and
                         Bernd Radig},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--9},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291328},
            urn       = {urn:nbn:de:101:1-201705291328},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {HTML5 video is the upcoming standard for playing videos on the World Wide Web. Although its specification has not been fully adopted yet, all major browsers provide the HTML5 video element and web developers already rely on its functionality. But there are differences between implementations and inaccuracies that trouble the web developer community. To help to improve the current situation we draw a comparison between the most important web browsers. We focus on the event mechanism, since it is essential for interacting with the video element. Furthermore, we compare the seeking accuracy, which is relevant for more specialized applications. Our tests reveal varieties of differences between browser interfaces and show that even simple software solutions may still need third-party plugins in today's browsers.}
        }
    
  38.  Open Access 

    A Self-Optimizing Cloud Computing System for Distributed Storage and Processing of Semantic Web Data

    Sven Groppe, Johannes Blume, Dennis Heinrich, Stefan Werner

    Open Journal of Cloud Computing (OJCC), 1(2), Pages 1-14, 2014, Downloads: 2743, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194478 | GNL-LP: 113236065X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Clouds are dynamic networks of common, off-the-shell computers to build computation farms. The rapid growth of databases in the context of the semantic web requires efficient ways to store and process this data. Using cloud technology for storing and processing Semantic Web data is an obvious way to overcome difficulties in storing and processing the enormously large present and future datasets of the Semantic Web. This paper presents a new approach for storing Semantic Web data, such that operations for the evaluation of Semantic Web queries are more likely to be processed only on local data, instead of using costly distributed operations. An experimental evaluation demonstrates the performance improvements in comparison to a naive distribution of Semantic Web data.

    BibTex:

        @Article{OJCC-v1i2n01_Groppe,
            title     = {A Self-Optimizing Cloud Computing System for Distributed Storage and Processing of Semantic Web Data},
            author    = {Sven Groppe and
                         Johannes Blume and
                         Dennis Heinrich and
                         Stefan Werner},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {1--14},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194478},
            urn       = {urn:nbn:de:101:1-201705194478},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Clouds are dynamic networks of common, off-the-shell computers to build computation farms. The rapid growth of databases in the context of the semantic web requires efficient ways to store and process this data. Using cloud technology for storing and processing Semantic Web data is an obvious way to overcome difficulties in storing and processing the enormously large present and future datasets of the Semantic Web. This paper presents a new approach for storing Semantic Web data, such that operations for the evaluation of Semantic Web queries are more likely to be processed only on local data, instead of using costly distributed operations. An experimental evaluation demonstrates the performance improvements in comparison to a naive distribution of Semantic Web data.}
        }
    
  39.  Open Access 

    Getting Indexed by Bibliographic Databases in the Area of Computer Science

    Arne Kusserow, Sven Groppe

    Open Journal of Web Technologies (OJWT), 1(2), Pages 10-27, 2014, Downloads: 2823, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291343 | GNL-LP: 1133021557 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Every author and publisher is interested in adding their publications to the widely used bibliographic databases freely accessible in the world wide web: This ensures the visibility of their publications and hence of the published research. However, the inclusion requirements of publications in the bibliographic databases are heterogeneous even on the technical side. This survey paper aims in shedding light on the various data formats, protocols and technical requirements of getting indexed by widely used bibliographic databases in the area of computer science and provides hints for maximal database inclusion. Furthermore, we point out the possibilities to utilize the data of bibliographic databases, and describes some personal and institutional research repository systems with special regard to the support of inclusion in bibliographic databases.

    BibTex:

        @Article{OJWT_2014v1i2n02_Kusserow,
            title     = {Getting Indexed by Bibliographic Databases in the Area of Computer Science},
            author    = {Arne Kusserow and
                         Sven Groppe},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2014},
            volume    = {1},
            number    = {2},
            pages     = {10--27},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291343},
            urn       = {urn:nbn:de:101:1-201705291343},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Every author and publisher is interested in adding their publications to the widely used bibliographic databases freely accessible in the world wide web: This ensures the visibility of their publications and hence of the published research. However, the inclusion requirements of publications in the bibliographic databases are heterogeneous even on the technical side. This survey paper aims in shedding light on the various data formats, protocols and technical requirements of getting indexed by widely used bibliographic databases in the area of computer science and provides hints for maximal database inclusion. Furthermore, we point out the possibilities to utilize the data of bibliographic databases, and describes some personal and institutional research repository systems with special regard to the support of inclusion in bibliographic databases.}
        }
    
  40.  Open Access 

    Using Nuisance Telephone Denial of Service to Combat Online Sex Trafficking

    Ross A. Malaga

    Open Journal of Information Systems (OJIS), 2(1), Pages 1-8, 2015, Downloads: 2414, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194736 | GNL-LP: 1132361036 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Over the past few years, sex trafficking has been linked to online classified ads sites such as Craigslist.com and Backpage.com. However, to date technology-based solutions have not been used to attack classified ad sites or the advertisers. This paper proposes and tests a new approach to combating online sex trafficking promulgated via online classified ad sites - nuisance telephone denial of service (TDoS) attacks on the advertisers. The method of attack is described and implications are discussed.

    BibTex:

        @Article{OJIS_2015v2i1n01_Malaga,
            title     = {Using Nuisance Telephone Denial of Service to Combat Online Sex Trafficking},
            author    = {Ross A. Malaga},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {1--8},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194736},
            urn       = {urn:nbn:de:101:1-201705194736},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Over the past few years, sex trafficking has been linked to online classified ads sites such as Craigslist.com and Backpage.com. However, to date technology-based solutions have not been used to attack classified ad sites or the advertisers. This paper proposes and tests a new approach to combating online sex trafficking promulgated via online classified ad sites - nuisance telephone denial of service (TDoS) attacks on the advertisers. The method of attack is described and implications are discussed.}
        }
    
  41.  Open Access 

    IT Governance Practices for Electric Utilities: Insights from Brazil and Europe

    Paulo Rupino da Cunha, Luiz Mauricio Martins, Antão Moura, António Dias de Figueiredo

    Open Journal of Information Systems (OJIS), 2(1), Pages 9-28, 2015, Downloads: 2152, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194743 | GNL-LP: 1132361044 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: We propose a framework of 14 IT governance practices tailored for the electric utilities sector. They were selected and ranked as "essential", "important", or "good" by top executives and IT staff from two multi-billion dollar companies - one in Brazil and another in Europe - from a generic set of 83 collected in the literature and in the field. Our framework addresses a need of electric utilities for which specific guidance was lacking. We have also uncovered a significant impact of social issues in IT governance, whose depth seems to be missing in the current research. As a byproduct of our work, the larger generic framework from which we have departed and the tailoring method that we have proposed can be used to customize the generic framework to different industries.

    BibTex:

        @Article{OJIS_2015v2i1n02_Cunha,
            title     = {IT Governance Practices for Electric Utilities: Insights from Brazil and Europe},
            author    = {Paulo Rupino da Cunha and
                         Luiz Mauricio Martins and
                         Ant\~{a}o Moura and
                         Ant\'{o}nio Dias de Figueiredo},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {9--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194743},
            urn       = {urn:nbn:de:101:1-201705194743},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {We propose a framework of 14 IT governance practices tailored for the electric utilities sector. They were selected and ranked as "essential", "important", or "good" by top executives and IT staff from two multi-billion dollar companies - one in Brazil and another in Europe - from a generic set of 83 collected in the literature and in the field. Our framework addresses a need of electric utilities for which specific guidance was lacking. We have also uncovered a significant impact of social issues in IT governance, whose depth seems to be missing in the current research. As a byproduct of our work, the larger generic framework from which we have departed and the tailoring method that we have proposed can be used to customize the generic framework to different industries.}
        }
    
  42.  Open Access 

    Cognitive Spam Recognition Using Hadoop and Multicast-Update

    Mukund YR, Sunil Sandeep Nayak, K. Chandrasekaran

    Open Journal of Big Data (OJBD), 1(1), Pages 16-28, 2015, Downloads: 1918, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194340 | GNL-LP: 1132360498 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In today's world of exponentially growing technology, spam is a very common issue faced by users on the internet. Spam not only hinders the performance of a network, but it also wastes space and time, and causes general irritation and presents a multitude of dangers - of viruses, malware, spyware and consequent system failure, identity theft, and other cyber criminal activity. In this context, cognition provides us with a method to help improve the performance of the distributed system. It enables the system to learn what it is supposed to do for different input types as different classifications are made over time and this learning helps it increase its accuracy as time passes. Each system on its own can only do so much learning, because of the limited sample set of inputs that it gets to process. However, in a network, we can make sure that every system knows the different kinds of inputs available and learns what it is supposed to do with a better success rate. Thus, distribution and combination of this cognition across different components of the network leads to an overall improvement in the performance of the system. In this paper, we describe a method to make machines cognitively label spam using Machine Learning and the Naive Bayesian approach. We also present two possible methods of implementation - using a MapReduce Framework (hadoop), and also using messages coupled with a multicast-send based network - with their own subtypes, and the pros and cons of each. We finally present a comparative analysis of the two main methods and provide a basic idea about the usefulness of the two in various different scenarios.

    BibTex:

        @Article{OJBD_2015v1i1n03_YR,
            title     = {Cognitive Spam Recognition Using Hadoop and Multicast-Update},
            author    = {Mukund YR and
                         Sunil Sandeep Nayak and
                         K. Chandrasekaran},
            journal   = {Open Journal of Big Data (OJBD)},
            issn      = {2365-029X},
            year      = {2015},
            volume    = {1},
            number    = {1},
            pages     = {16--28},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194340},
            urn       = {urn:nbn:de:101:1-201705194340},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In today's world of exponentially growing technology, spam is a very common issue faced by users on the internet. Spam not only hinders the performance of a network, but it also wastes space and time, and causes general irritation and presents a multitude of dangers - of viruses, malware, spyware and consequent system failure, identity theft, and other cyber criminal activity. In this context, cognition provides us with a method to help improve the performance of the distributed system. It enables the system to learn what it is supposed to do for different input types as different classifications are made over time and this learning helps it increase its accuracy as time passes. Each system on its own can only do so much learning, because of the limited sample set of inputs that it gets to process. However, in a network, we can make sure that every system knows the different kinds of inputs available and learns what it is supposed to do with a better success rate. Thus, distribution and combination of this cognition across different components of the network leads to an overall improvement in the performance of the system. In this paper, we describe a method to make machines cognitively label spam using Machine Learning and the Naive Bayesian approach. We also present two possible methods of implementation - using a MapReduce Framework (hadoop), and also using messages coupled with a multicast-send based network - with their own subtypes, and the pros and cons of each. We finally present a comparative analysis of the two main methods and provide a basic idea about the usefulness of the two in various different scenarios.}
        }
    
  43.  Open Access 

    Evidential Sensor Data Fusion in a Smart City Environment

    Aditya Gaur, Bryan W. Scotney, Gerard P. Parr, Sally I. McClean

    Open Journal of Internet Of Things (OJIOT), 1(2), Pages 1-18, 2015, Downloads: 2799, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201704244969 | GNL-LP: 113062319X | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Wireless sensor networks have increasingly become contributors of very large amounts of data. The recent deployment of wireless sensor networks in Smart City infrastructures have led to very large amounts of data being generated each day across a variety of domains, with applications including environmental monitoring, healthcare monitoring and transport monitoring. The information generated through the wireless sensor nodes has made possible the visualization of a Smart City environment for better living. The Smart City offers intelligent infrastructure and cogitative environment for the elderly and other people living in the Smart society. Different types of sensors are present that help in monitoring inhabitants' behaviour and their interaction with real world objects. To take advantage of the increasing amounts of data, there is a need for new methods and techniques for effective data management and analysis, to generate information that can assist in managing the resources intelligently and dynamically. Through this research a Smart City ontology model is proposed, which addresses the fusion process related to uncertain sensor data using semantic web technologies and Dempster-Shafer uncertainty theory. Based on the information handling methods, such as Dempster-Shafer theory (DST), an equally weighted sum operator and maximization operation, a higher level of contextual information is inferred from the low-level sensor data fusion process. In addition, the proposed ontology model helps in learning new rules that can be used in defining new knowledge in the Smart City system.

    BibTex:

        @Article{OJIOT_2015v1i2n02_Gaur,
            title     = {Evidential Sensor Data Fusion in a Smart City Environment},
            author    = {Aditya Gaur and
                         Bryan W. Scotney and
                         Gerard P. Parr and
                         Sally I. McClean},
            journal   = {Open Journal of Internet Of Things (OJIOT)},
            issn      = {2364-7108},
            year      = {2015},
            volume    = {1},
            number    = {2},
            pages     = {1--18},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201704244969},
            urn       = {urn:nbn:de:101:1-201704244969},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Wireless sensor networks have increasingly become contributors of very large amounts of data. The recent deployment of wireless sensor networks in Smart City infrastructures have led to very large amounts of data being generated each day across a variety of domains, with applications including environmental monitoring, healthcare monitoring and transport monitoring. The information generated through the wireless sensor nodes has made possible the visualization of a Smart City environment for better living. The Smart City offers intelligent infrastructure and cogitative environment for the elderly and other people living in the Smart society. Different types of sensors are present that help in monitoring inhabitants' behaviour and their interaction with real world objects. To take advantage of the increasing amounts of data, there is a need for new methods and techniques for effective data management and analysis, to generate information that can assist in managing the resources intelligently and dynamically. Through this research a Smart City ontology model is proposed, which addresses the fusion process related to uncertain sensor data using semantic web technologies and Dempster-Shafer uncertainty theory. Based on the information handling methods, such as Dempster-Shafer theory (DST), an equally weighted sum operator and maximization operation, a higher level of contextual information is inferred from the low-level sensor data fusion process. In addition, the proposed ontology model helps in learning new rules that can be used in defining new knowledge in the Smart City system.}
        }
    
  44.  Open Access 

    Model of Creative Thinking Process on Analysis of Handwriting by Digital Pen

    Kenshin Ikegami, Yukio Ohsawa

    Open Journal of Information Systems (OJIS), 2(2), Pages 27-39, 2015, Downloads: 1244, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194781 | GNL-LP: 1132361125 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: In order to perceive infrequent events as hints for new ideas, it is desired to know and model the process of creating and refining ideas. In this paper, we address this modeling problem experimentally. Firstly, we focus on the relation between thinking time and writing time in handwriting. We observe two types of patterns; one group takes longer time in thinking and shorter in writing, the other takes longer in writing and shorter in thinking. The group having spends longer in writing has shorter time span from one sentence to another than the other group. Backtracking, i.e., the event that participants return back to their former sheet and modify opinions, is observed more often in the group of longer writing than the other group. In addition, participants in this backtracking group gets higher scores for their ideas on sheets than those in the no-backtracking group. We propose a model of creative thinking by applying Operations of Structure of Intellect. It is inferred that the group of longer writing conducts a series of thinking flow, including divergent thinking, convergent thinking and evaluation. In contrast, the group of longer thinking tends to conduct the two different thinking flow: divergent thinking and evaluation; convergent thinking and evaluation. For making creative ideas, we conduct divergent thinking without evaluation and created a large number of ideas. We conclude that the rotations of divergent thinking, convergent thinking and evaluation increase the frequency of "backtracking" and make the ideas more logical ones.

    BibTex:

        @Article{OJIS_2015v2i2n04_Ikegami,
            title     = {Model of Creative Thinking Process on Analysis of Handwriting by Digital Pen},
            author    = {Kenshin Ikegami and
                         Yukio Ohsawa},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {27--39},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194781},
            urn       = {urn:nbn:de:101:1-201705194781},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {In order to perceive infrequent events as hints for new ideas, it is desired to know and model the process of creating and refining ideas. In this paper, we address this modeling problem experimentally. Firstly, we focus on the relation between thinking time and writing time in handwriting. We observe two types of patterns; one group takes longer time in thinking and shorter in writing, the other takes longer in writing and shorter in thinking. The group having spends longer in writing has shorter time span from one sentence to another than the other group. Backtracking, i.e., the event that  participants return back to their former sheet and modify opinions, is observed more often in the group of longer writing than the other group. In addition, participants in this backtracking group gets higher scores for their ideas on sheets than those in the no-backtracking group. We propose a model of creative thinking by applying Operations of Structure of Intellect. It is inferred that the group of longer writing conducts a series of thinking flow, including divergent thinking, convergent thinking and evaluation. In contrast, the group of longer thinking tends to conduct the two different thinking flow: divergent thinking and evaluation; convergent thinking and evaluation. For making creative ideas, we conduct divergent thinking without evaluation and created a large number of ideas. We conclude that the rotations of divergent thinking, convergent thinking and evaluation increase the frequency of "backtracking" and make the ideas more logical ones.}
        }
    
  45.  Open Access 

    Ontology Evolution Using Ontology Templates

    Miroslav Blasko, Petr Kremen, Zdenek Kouba

    Open Journal of Semantic Web (OJSW), 2(1), Pages 16-29, 2015, Downloads: 1011, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194898 | GNL-LP: 1132361281 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Evolving ontologies by domain experts is difficult and typically cannot be performed without the assistance of an ontology engineer. This process takes long time and often recurrent modeling errors have to be resolved. This paper proposes a technique for creating controlled ontology evolution scenarios that ensure consistency of the possible ontology evolution and give guarrantees to the domain expert that his/her updates do not cause inconsistency. We introduce ontology templates that formalize the notion of controlled evolution and define ontology template consistency checking service together with a consistency checking algorithm. We prove correctness and demonstate the practical use of the techniques in two scenarios.

    BibTex:

        @Article{OJSW_2015v2i1n03_Blasko,
            title     = {Ontology Evolution Using Ontology Templates},
            author    = {Miroslav Blasko and
                         Petr Kremen and
                         Zdenek Kouba},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {16--29},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194898},
            urn       = {urn:nbn:de:101:1-201705194898},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Evolving ontologies by domain experts is difficult and typically cannot be performed without the assistance of an ontology engineer. This process takes long time and often recurrent modeling errors have to be resolved. This paper proposes a technique for creating controlled ontology evolution scenarios that ensure consistency of the possible ontology evolution and give guarrantees to the domain expert that his/her updates do not cause inconsistency. We introduce ontology templates that formalize the notion of controlled evolution and define ontology template consistency checking service together with a consistency checking algorithm. We prove correctness and demonstate the practical use of the techniques in two scenarios.}
        }
    
  46.  Open Access 

    A Toulmin's Framework-Based Method for Design Argumentation of Cyber-Physical Systems

    Noriyuki Kushiro, Ryoichi Torikai, Shodai Matsuda, Kunio Takahara

    Open Journal of Information Systems (OJIS), 2(2), Pages 40-55, 2015, Downloads: 1043, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194809 | GNL-LP: 1132361141 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The design of cyber-physical systems (CPS) is a promising domain, where the data market is expected to soon penetrate. When engineers focus on only a particular part of data (whether intentionally or not) for establishing a design hypothesis, the design hypothesis may also be supported by data sets in the market. Therefore, the validity of such a design hypothesis cannot be evaluated by the data itself, and can only be accepted by the robustness of the logic behind the design argumentation. Although the validation of the design logic is significant, cognitive aspects (which people have spontaneously) disturb the design argumentation reasoning. Therefore, a design method that overcomes the cognitive aspects is indispensable for the CPS designers. This work proposes a CPS design method using the interaction between logic and data sets with a logic visualization tool, and applies the proposed method to the design of a diagnosis system for semiconductor manufacture. The capability of the proposed method is also discussed and analyzed in this paper.

    BibTex:

        @Article{OJIS_2015v2i2n05_Kushiro,
            title     = {A Toulmin's Framework-Based Method for Design Argumentation of Cyber-Physical Systems},
            author    = {Noriyuki Kushiro and
                         Ryoichi Torikai and
                         Shodai Matsuda and
                         Kunio Takahara},
            journal   = {Open Journal of Information Systems (OJIS)},
            issn      = {2198-9281},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {40--55},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194809},
            urn       = {urn:nbn:de:101:1-201705194809},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The design of cyber-physical systems (CPS) is a promising domain, where the data market is expected to soon penetrate. When engineers focus on only a particular part of data (whether intentionally or not) for establishing a design hypothesis, the design hypothesis may also be supported by data sets in the market. Therefore, the validity of such a design hypothesis cannot be evaluated by the data itself, and can only be accepted by the robustness of the logic behind the design argumentation. Although the validation of the design logic is significant, cognitive aspects (which people have spontaneously) disturb the design argumentation reasoning. Therefore, a design method that overcomes the cognitive aspects is indispensable for the CPS designers. This work proposes a CPS design method using the interaction between logic and data sets with a logic visualization tool, and applies the proposed method to the design of a diagnosis system for semiconductor manufacture. The capability of the proposed method is also discussed and analyzed in this paper.}
        }
    
  47.  Open Access 

    PatTrieSort - External String Sorting based on Patricia Tries

    Sven Groppe, Dennis Heinrich, Stefan Werner, Christopher Blochwitz, Thilo Pionteck

    Open Journal of Databases (OJDB), 2(1), Pages 36-50, 2015, Downloads: 1822, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194627 | GNL-LP: 1132360889 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Presentation: Video

    Abstract: External merge sort belongs to the most efficient and widely used algorithms to sort big data: As much data as fits inside is sorted in main memory and afterwards swapped to external storage as so called initial run. After sorting all the data in this way block-wise, the initial runs are merged in a merging phase in order to retrieve the final sorted run containing the completely sorted original data. Patricia tries are one of the most space-efficient ways to store strings especially those with common prefixes. Hence, we propose to use patricia tries for initial run generation in an external merge sort variant, such that initial runs can become large compared to traditional external merge sort using the same main memory size. Furthermore, we store the initial runs as patricia tries instead of lists of sorted strings. As we will show in this paper, patricia tries can be efficiently merged having a superior performance in comparison to merging runs of sorted strings. We complete our discussion with a complexity analysis as well as a comprehensive performance evaluation, where our new approach outperforms traditional external merge sort by a factor of 4 for sorting over 4 billion strings of real world data.

    BibTex:

        @Article{OJDB_2015v2i1n03_Groppe,
            title     = {PatTrieSort - External String Sorting based on Patricia Tries},
            author    = {Sven Groppe and
                         Dennis Heinrich and
                         Stefan Werner and
                         Christopher Blochwitz and
                         Thilo Pionteck},
            journal   = {Open Journal of Databases (OJDB)},
            issn      = {2199-3459},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {36--50},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194627},
            urn       = {urn:nbn:de:101:1-201705194627},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {External merge sort belongs to the most efficient and widely used algorithms to sort big data: As much data as fits inside is sorted in main memory and afterwards swapped to external storage as so called initial run. After sorting all the data in this way block-wise, the initial runs are merged in a merging phase in order to retrieve the final sorted run containing the completely sorted original data. Patricia tries are one of the most space-efficient ways to store strings especially those with common prefixes. Hence, we propose to use patricia tries for initial run generation in an external merge sort variant, such that initial runs can become large compared to traditional external merge sort using the same main memory size. Furthermore, we store the initial runs as patricia tries instead of lists of sorted strings. As we will show in this paper, patricia tries can be efficiently merged having a superior performance in comparison to merging runs of sorted strings. We complete our discussion with a complexity analysis as well as a comprehensive performance evaluation, where our new approach outperforms traditional external merge sort by a factor of 4 for sorting over 4 billion strings of real world data.}
        }
    
  48.  Open Access 

    Why Is This Link Here? Identifying Academic Web Interlinking Motivations in Nigerian Universities

    Anthony Nwohiri

    Open Journal of Web Technologies (OJWT), 2(1), Pages 4-15, 2015, Downloads: 900, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705291363 | GNL-LP: 1133021581 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: This paper investigates the university websites of Nigeria, Africa's most populous nation. Its aim is to identify motivations why authors embed outbound hyperlinks on these websites. A classification scheme for academic web interlinking motivations was applied to over 5,000 hyperlinks pointing from the websites of 107 Nigerian universities. Classifying the motivations based on studying the source and target pages is a big challenge, especially due to the following three reasons: there could be many possible reasons available; guessing the true intentions of link creators could be a difficult task; multiple link creation motivations could exist. The pioneer application of Pearson's chi-square test of independence offers a better picture of motivations. The chi-square test identifies the significant differences in interlinking motivations, which are peculiar to Nigerian universities of a particular category (federal, state and private universities). The study is a stepping stone toward further research on feasibility of findings in other developing countries. Results obtained from this research will be of great use for academic webpage developers and web authors, and will modify their work towards improving the use of hyperlinks as one of the major communication tools on the Web.

    BibTex:

        @Article{OJWT_2015v2i1n02_Nwohir,
            title     = {Why Is This Link Here? Identifying Academic Web Interlinking Motivations in Nigerian Universities},
            author    = {Anthony Nwohiri},
            journal   = {Open Journal of Web Technologies (OJWT)},
            issn      = {2199-188X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {4--15},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705291363},
            urn       = {urn:nbn:de:101:1-201705291363},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {This paper investigates the university websites of Nigeria, Africa's most populous nation. Its aim is to identify motivations why authors embed outbound hyperlinks on these websites. A classification scheme for academic web interlinking motivations was applied to over 5,000 hyperlinks pointing from the websites of 107 Nigerian universities. Classifying the motivations based on studying the source and target pages is a big challenge, especially due to the following three reasons: there could be many possible reasons available; guessing the true intentions of link creators could be a difficult task; multiple link creation motivations could exist. The pioneer application of Pearson's chi-square test of independence offers a better picture of motivations. The chi-square test identifies the significant differences in interlinking motivations, which are peculiar to Nigerian universities of a particular category (federal, state and private universities). The study is a stepping stone toward further research on feasibility of findings in other developing countries. Results obtained from this research will be of great use for academic webpage developers and web authors, and will modify their work towards improving the use of hyperlinks as one of the major communication tools on the Web.}
        }
    
  49.  Open Access 

    A Trust-Based Approach for Management of Dynamic QoS Violations in Cloud Federation Environments

    Manoj V. Thomas, K. Chandrasekaran

    Open Journal of Cloud Computing (OJCC), 2(2), Pages 21-43, 2015, Downloads: 1153, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194523 | GNL-LP: 1132360765 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: Cloud Federation is an emerging technology where Cloud Service Providers (CSPs) offering specialized services to customers collaborate in order to reap the real benefits of Cloud Computing. When a CSP in the Cloud Federation runs out of resources, it can get the required resources from other partners in the federation. Normally, there will be QoS agreements between the partners in the federation for the resource sharing. In this paper, we propose a trust based mechanism for the management of dynamic QoS violations, when one CSP requests resources from another CSP in the federation. In this work, we have implemented the partner selection process, when one CSP does not have enough resources, using the Analytic Hierarchy Process (AHP) and the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) methods, and also considering the trust values of various CSPs in the federation. We have also implemented the Single Sign-On (SSO) authentication in the cloud federation using the Fully Hashed Menezes-Qu-Vanstone (FHMQV) protocol and AES-256 algorithm. The proposed trust-based approach is used to dynamically manage the QoS violations among the partners in the federation. We have implemented the proposed approach using the CloudSim toolkit, and the analysis of the results are also given.

    BibTex:

        @Article{OJCC_2015v2i2n03_Thomas,
            title     = {A Trust-Based Approach for Management of Dynamic QoS Violations in Cloud Federation Environments},
            author    = {Manoj V. Thomas and
                         K. Chandrasekaran},
            journal   = {Open Journal of Cloud Computing (OJCC)},
            issn      = {2199-1987},
            year      = {2015},
            volume    = {2},
            number    = {2},
            pages     = {21--43},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194523},
            urn       = {urn:nbn:de:101:1-201705194523},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {Cloud Federation is an emerging technology where Cloud Service Providers (CSPs) offering specialized services to customers collaborate in order to reap the real benefits of Cloud Computing. When a CSP in the Cloud Federation runs out of resources, it can get the required resources from other partners in the federation. Normally, there will be QoS agreements between the partners in the federation for the resource sharing. In this paper, we propose a trust based mechanism for the management of dynamic QoS violations, when one CSP requests resources from another CSP in the federation. In this work, we have implemented the partner selection process, when one CSP does not have enough resources, using the Analytic Hierarchy Process (AHP) and the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) methods, and also considering the trust values of various CSPs in the federation. We have also implemented the Single Sign-On (SSO) authentication in the cloud federation using the Fully Hashed Menezes-Qu-Vanstone (FHMQV) protocol and AES-256 algorithm. The proposed trust-based approach is used to dynamically manage the QoS violations among the partners in the federation. We have implemented the proposed approach using the CloudSim toolkit, and the analysis of the results are also given.}
        }
    
  50.  Open Access 

    Semantic and Web: The Semantic Part

    Sven Groppe, Paulo Rupino da Cunha

    Open Journal of Semantic Web (OJSW), 2(1), Pages 1-3, 2015, Downloads: 1811, Citations: 1

    Full-Text: pdf | URN: urn:nbn:de:101:1-201705194864 | GNL-LP: 1132361222 | Meta-Data: tex xml rdf rss | Show/Hide Abstract | Show/Hide BibTex

    Abstract: The Web is everywhere in daily life. Business is not possible any more without the fast communication through the web. The knowledge of the humans is reflected in the information accessible in the web. New challenges occur with the flood of information and electronic possibilities for the human being. The current World Wide Web enables an easy, instant access to a vast amount of online information. However, the content in the Web is typically for human consumption, and is not tailored to be machine-processed. The Semantic Web, which is intended to establish a machine-understandable web, thereby offers a promising and potential solution to mining and analyzing web content. The Semantic Web is currently changing from an emergent trend to a technology used in complex real-world applications. This part of the special issue "Semantic and Web" especially investigates how semantic technologies can help the human being to open the new possibilities of the web. The papers, which contribute more to Web technologies, are published in Open Journal of Web Technologies (OJWT).

    BibTex:

        @Article{OJSW_2015v2i1n01e_Groppe,
            title     = {Semantic and Web: The Semantic Part},
            author    = {Sven Groppe and
                         Paulo Rupino da Cunha},
            journal   = {Open Journal of Semantic Web (OJSW)},
            issn      = {2199-336X},
            year      = {2015},
            volume    = {2},
            number    = {1},
            pages     = {1--3},
            url       = {http://nbn-resolving.de/urn:nbn:de:101:1-201705194864},
            urn       = {urn:nbn:de:101:1-201705194864},
            publisher = {RonPub},
            bibsource = {RonPub},
            abstract = {The Web is everywhere in daily life. Business is not possible any more without the fast communication through the web. The knowledge of the humans is reflected in the information accessible in the web. New challenges occur with the flood of information and electronic possibilities for the human being. The current World Wide Web enables an easy, instant access to a vast amount of online information. However, the content in the Web is typically for human consumption, and is not tailored to be machine-processed. The Semantic Web, which is intended to establish a machine-understandable web, thereby offers a promising and potential solution to mining and analyzing web content. The Semantic Web is currently changing from an emergent trend to a technology used in complex real-world applications. This part of the special issue "Semantic and Web" especially investigates how semantic technologies can help the human being to open the new possibilities of the web. The papers, which contribute more to Web technologies, are published in Open Journal of Web Technologies (OJWT).}
        }