Sei sulla pagina 1di 14

AN INCREMENTAL AND

DISSEMINATED SUPPOSITION
FOR SWAPPING MAP REDUCE
USING ONTOLOGIES

Aim and Objective

The main objective of the proposed system is to optimize wed data


resource allocation using Hadoop slot allocation.

A large volume of Semantic Web data performance balancing and slot


PreScheduling.

ABSTRACT

The large amount of Semantic Web data and its fast growth pose a significant
computational challenge in performing efficient and scalable reasoning.

The resources of single machines are no longer sufficient and we are required
to distribute the process to improve performance Constructing transfer
inference forest and effective assertion triples.

A large volume of Semantic Web data, the fast growth of ontology bases has
brought significant challenges in performing efficient and scalable reasoning.

Distributed reasoning methods are thus required to improve the scalability


and performance of inferences.

evaluated our system using very large real-world datasets (Bio2RDF, LLD,
LDSR) and the LUBM synthetic benchmark, scaling up to 100 billion triples.

EXISTING SYSTEM

A centralized architecture executed on a single machine or local server when


dealing with large datasets.

Distributed reasoning approaches executed on multiple computing nodes have


thus emerged to improve the scalability and speed of inferences.

The general execution constraint that map tasks are executed before reduce
tasks.

Disadvantage

The data volume of RDF closure is ordinarily larger than original RDF
data.

The storage of RDF closure is thus not a small amount and the query on it
takes nontrivial time.

The data volume increases and the ontology base is updated, these
methods.

generally the ontology size is more than original data size

PROPOSED SYSTEM

The choice of Map Reduce is motivated by the fact that it can limit data
exchange and alleviate load balancing problems by dynamically scheduling jobs
on computing nodes.

The incremental RDF triples more efficiently, we present two novel concepts,
transfer inference forest and effective assertion triples.

largely reduce the storage and simplify the reasoning process.

Advantage

Linear scalability, automatic failover support, and convenient backup of


Map Reduce jobs Distributed data on the web make it difficult to acquire
appropriate triples for appropriate inferences

Well leverage the old and new data to minimize the updating time and
reduce the reasoning time when facing big RDF datasets

Technical's and algorithm


Existing Algorithm

PWS (Personalized Web search )

Bench Mark

Proposed Algorithm

RSVM

Bio2RDF

Hadoop Biosensing K means Over Map Reduce

SYSTEM REQUIREMENTS
SOFTWARE REQUIREMENTS

Operating system

Windows XP/7.

Coding Language

JAVA/J2EE

IDE

Implementation Tools :

Hadoop

Tools Support

Cygwin 2.28, VM Multiplayer

Netbeans 7.4
:

HARDWARE REQUIREMENTS

System

Dual Core 2.5 GHz.

Hard Disk

Ram

160 GB.
2GB

System Architecture

List of Modules

Data Preprocessing

Imputation Process

HDFS Upload

Data Upload

Execution Path Optimization

Job Execution

Multi Execution Path

Evaluation

Preprocessing

In this module we are going to perform two


processes. Collect the data set values to reduce
the irrelevant content to the web sites.

Data set values to split the before slot allocation


and after allocation.

Data values to proceed the minimum data to


map the task.

Imputation Process

In this module we are going to perform two processes.


Slot allocation Slot pre-scheduling process.

slot allocation process we are going allocate the slot


based on dynamic Hadoop slot allocation optimization
mechanism.

slot pre-scheduling process we are going to improve the


data locality. Slot Pre-Scheduling technique that can
improve the data locality while having no negative
impact on the fairness of Map-Reduce jobs.

Thank You

Potrebbero piacerti anche