Professional Documents
Culture Documents
ISSN : 2319-6890(Online)
2347-5013(Print)
Innovative Research Publications (IRP) is a fast growing international academic publisher that publishes
International Journals in the fields of Engineering, Science, Management. IRP is establishing a distinctive
and independent profile in the international arena. Our publications are distinctive for their relevance to
the target groups and for their stimulating contribution to R&D. Our Journals are the products of dynamic
International Journal of
interchange between Scientists, authors, publisher and designer.
Objectives:
Engineering Research
·Publishing National and Internationals Journals, Magazine, Books and others in online version as well
as print version to provide high quality and high standard publications in National and International
Journals
(IJER)
·Organizing technical events i.e. Seminars, workshop, conferences and symposia etc. to expose knowledge
of researchers Volume
Volume 4 Issue
Volume 43issue Special
3 3Issue 2115
96
Volume
Volume 343Issue
Issue 107
Volume 4 Issue Special
Volume Issue
·Collaborating with educational and research organizations to expand awareness about R&D Volume
Volume
Volume 4 4&
Issue
Volume
Volume
May 19 & 20,
May 19 Issue
Volume
Volume Issue
3
Special
23Issue
Volume
2015
20, 1856 43
3 Issue
Issue
Issue 24
8
3 Issue
2015
Feb. 01,Oct.
012015
July
March 01,
01,
Nov. 20, 2014
2014
2014
2015
Jan. 01,01, 2015
·Helping to financial weak researchers to promote their researches at world level Volume
Sept.4 Issue
01,
August
April 3
2014
01,
01,20152014
April
June
May01,
March 01, 2015 2014
2014
01, 2014
Our Journals
1. International Journal of Scientific Engineering and Technology
ISSN : 2277-1581
Subject : Science, Engineering, Management and Agriculture Engineering
Last Date for submitting paper : 10th of each month
Web : www.ijset.com, Email : editor@ijset.com
A National
2nd Conference
International on on
Conference "Recent Advances
Convergent in Chemical
Innovative Engineering"
Technologies (ICCIT-2015)
Innovative Research Publications GreenChem-15, on OnMarch 20, 2015
Gulmohar, Bhopal M.P. India, Contant No.:+91-9752135004 International Journal of Engineering Research
May 19 & 20, 2015
Organized By
Web : www.irpindiia.org, Email : info@irpindia.org Web : www.ijer.in, Email
Organized
Department of Chemical Engg, by :Yavatmal
JDIET,
editor@ijer.in
(M.S) India
Contant No.:+91-9752135004
Cambridge Institute of Technology, K.R. Puram, Bangalore
Editorial Board
Editor in Chief
Dr. R. K. Singh,
Professor and Head, Department of Electronics and Communication,
KNIT Sultanpur U.P., India
Managing Editor
Mr. J. K. Singh, Managing Editor
Innovative Research Publications, Bhopal M.P. India
Advisory Board
1. Dr. Asha Sharma, Jodhpur, Rajasthan, India
2. Dr. Subhash Chander Dubey, Jammu India
3. Dr. Rajeev Jain, Jabalpur M.P. India
4. Dr. C P Paul, Indore M.P. India
5. Dr. S. Satyanarayana, Guntur, A.P, India.
Organizing Committee
List of Contents
S.No. Manuscript Detail Page No.
Simulation and Analysis of an Energy efficient protocol Ad-LEACH for Smart Home
1. Wireless Sensor Network 336-339
Keywords—Wireless Sensor Network, Cluster, Energy II. LEACH (LOW ENERGY ADAPTIVE CLUSTERING
Efficient, Smart Home HIERARCHY)
The first hierarchical cluster-based routing protocol for
I. INTRODUCTION wireless sensor network is LEACH[1]. This protocol divides the
A wireless sensor Network(WSN) is made up of large nodes in the network into clusters. As shown in Fig 2 each cluster
number of small sensors with low-power transceivers. It helps to has a Cluster Head (CH). This dedicated CH node have extra
privileges. It is responsible for creating and manipulating a
gather data in a different environment. Each sensor collects data
TDMA (Time division multiple access) schedule[7]. The
and sends it through the network to a single processing centre, aggregated data send from nodes to the BS using CDMA (Code
base station (BS). These collected data helps to decide the division multiple access) by CH. Other than the CHs the
features of the environment or to detect a state of an object in a remaining nodes in a network are cluster members. LEACH
network. As shown in Fig 1, in a smart home(SH) protocol is divided into rounds. Each round consists of two
environment[6] each sensor deployed devices are considered as phases:
nodes in WSN. The status of the device information is sensed by Set-up Phase
the sensor. Then SH-WSN passes the sensed information to the
o Advertisement Phase
BS. The collected information are used to know the current
scenario of the smart home. It plays a great role in controlling the o Cluster Set-up Phase
smart home. Steady Phase
Each node in the WSN spends energy to transmit collected o Schedule Creation
data to CH. Each CH spends energy to receive data from all the o Data Transmission
nodes in the cluster, to aggregate the collected data and transmit
to the BS. The network protocol plays a vital role in the data A. Setup Phase
communication. Since WSN consume energy for the Each node independently decides that it can become a CH or
communication, it is major and critical task to identify the energy not. When the node became CH, it decides based on the node
efficient protocol. There are many protocols are proposed for served as a CH for the last time. There is a more chance to
WSN. Among those LEACH(Low Energy Adaptive Clustering become CH for a node that hasn't been a CH for long time than
Hierarchy) protocol helps to save energy in the smaller WSN like the nodes that have been a CH recently. During the
advertisement phase, an advertisement packet helps CH to
smart home.
informs their neighborhood that they become CHs. Non-CH Hierarchical Cluster-Based Routing - HRC[3] generate
nodes receive this CH information with the strongest signal energy-efficient clusters in a sensor network. It generates the CH
strength. By sending their IDs, the member nodes inform the set. Members of a cluster set are selected as a CH using round-
CH that they become a member of that cluster. All the CHs
robin technique. The formed clusters are continued for a short
communicate to the cluster members using TDMA. Now the
CH knows the number of member nodes and their Ids in the period of time called round. A round consists two phases namely
cluster. Based on all messages received within the cluster, the called an election phase and a data transfer phase. In the first
CH creates a TDMA schedule, pick a CDMA code randomly, phase, the sensor nodes generate the cluster with the head-set. In
and broadcast the TDMA table to cluster members. After that the second phase, the members in the head-set transmit the
steady-state phase begins. collected data to the BS as per their turn. The HCR protocol is
more energy efficient than the traditional cluster-based routing
techniques for continuous monitoring applications.
Enhanced-LEACH - En-LEACH [4] is the another version
of LEACH protocol. Probability method is used to select the
cluster CH. The formula used for selecting cluster head is
Cluster Head = Energy of the node / Energy of the Cluster
Objective of En-LEACH
Fig 2. LEACH Protocol
o To handle cluster-head failure
B. Steady-State Phase o To account for the non-uniform and dynamic
Actual data transmission begins in this phase. Nodes starts residual energy of the nodes
sending their data to the CH during their allocated TDMA slot.
After implementation, the result shows the first node death is
Minimal amount of energy is used in this transmission. All the
almost two times later than the LEACH. Last node death occurs
other non-CH node's radio can be turned off until the nodes are
much later than LEACH.
allocated TDMA slot. This minimize the energy dissipation in
these nodes. The CH aggregates the received data and sends it to Ad-LEACH is the new approach for the wireless sensor
the BS. network. In this approach, CH is elected based on two criteria (i)
node's residual energy (ii) node should not serve as a CH
DISADVANTAGE OF LEACH recently. The higher priority node will be elected as a CH. After
this, the next priority node will be considered as CH-Rep(Cluster
LEACH protocol selects the CH randomly without
Head Representative). In LEACH protocol, CH loses more
considering energy consumption. In this protocol, a node with
energy than the normal node because it spends energy to receive
the less energy has the same priority to become CH as the node
data from all the cluster node and to transmit to the base station
with the more energy. This results, all nodes die soon that leads
which is far away. CH spends less energy to transmit since it
the network failure fast. Since LEACH has the drawback, many
transmits single base station and spend more energy to receive
researchers have been done to make this protocol perform
since it receives from many nodes. Here CH-Rep helps CH to
better[5].
save energy. It receives data from all the nodes in a cluster,
aggregate it and transmit to the CH. Now the CH transmit to the
III. LITERATURE REVIEW base station. In Ad-LEACH, CH spends less energy than in
Cluster Based Routing Protocol-CBRP[2] is a distributive LEACH since it receives data from only one node CH-Rep.
energy efficient protocol for data gathering in wireless sensor
network. This protocol elects CH only based on a node’s own IV. IMPLEMENTATION
residual energy. After the CH selection, CBRP establishes a
Nowadays, research in the area of low-energy radio is a great
spanning tree over all of the CHs. Only the root node of the
challenge for researchers. There are distinct theories about the
spanning tree can communicate with the sink node by single-hop
radio model and energy dissipation in the transmit and receive
communication. The energy consumption by the nodes in the
mode. Fig 3 shows the radio energy dissipation model[1] used in
network for all communication is calculated by the free space
our work.
model. CBRP proved that the energy saved extremely and
extends the network lifetime.
I. RESULT
Ad-LEACH protocol is simulated using MATLAB tool. The
sensor network consists of 100 nodes that scatter randomly in 100 Figure 5 Network Lifetime
X 100 of a square field. The table I shows all parameters used to
implement the Ad-LEACH protocol. In this simulation, all node
begins with 20J of energy. TABLE. II Comparative Study of CBRP, HCR, En-LEACH &
Ad-LEACH
En- Ad-
Features CBRP HCR
LEACH LEACH
One from node's
Node with High
Selection Head-Set Probability residual
Residual
of CH by Round method energy
Energy
Robin &node
Troubleshooter: Solution Finder for log errors from Multiple Solution Sources
SmitaB Patil Dr. D R Shashikumar
M.Tech Student Deptt of CSE, CiTech, Bangalure, Karnataka India
smitapatil.july90@gmail.com
Abstract : Currently, one of the challenges for team- independently and would help the team to correlate the data
Escalation is to have a unified tool through all the Knowledge available with other knowledge sources. The result with respect
Base can be searched together. With this tool-Troubleshooter, to the different solution sources are display on different tabs of
we would be able to facilitate the team to pull the information the tabbed pane where each tabs will display names of solution
from all the knowledge sources such as, engineering KB, sources so that it is easy for the user to know from where the
Documentationand JIRA or even from external Google. solutions are fetched.
Further, in Engineering Knowledge base, we have multiple
sources like bug tracking system, Forum and even customer The tool Troubleshooter also provided with list of
searchable knowledge base. With this proposed tool- standard java errors and also some frequently occurring errors
“Troubleshooter: Solution Finder for Log Errors”, we will be with respect to the different products- SRM, NCM/UIM, ViPR,
integrating all the knowledge sources under one umbrella SMARTS and Watch4Net.By providing bunch of errors
through which the team can search through. This would Troubleshooter helps the user to directly select an error from the
reduce the time taken to search all knowledge sources provided list and perform search operation instead of typing
independently and would help the team to correlate the data error manually and perform search operation.
available with other knowledge sources. Additionally, this tool
The tool is very useful for the user to find solutions
also has the capacity to search based on the product selected
from different solution sources at a time as there is no such
and thus would help in filtering the result which are essentially
application which provides solutions from multiple solution
for the selected product.
sources at time this tool helps them for the same. Initially this
tool will authenticate the user and never ask again for the
Index Terms : Search Solutions in single solution sources, authentication. But without this tool one need to authenticate by
Search Solutions in Multiple solution sources. History for
all solutions sources individually irrespective of other solution
Search Results, Product Based Search.
sources. Without this tool the user needs to give his/her
username and password manually for the multiple times
I. INTRODUCTION whenever solution sources needs an authentication this is one
overhead work for the user.
The escalation team in industryis used to examine the log files
manually for an error. Once the employee comes up with errors
in the log fie, his/her next target is to find solutions for the II. Existing System
obtained error(s) with respective to the products.
There are many systems exists to find the errors from log files
With proposed system-Troubleshooter, we would be for examples Xpolog, Log Analyzer, Event Log Analyzer,
able to facilitate the team to pull the information from all the Piwik, nxlog and Octopussy. These log analyzing systems that
knowledge sources such as, engineering KB, Documentation and are available in the market are used to analyze a log file and to
JIRA or even from external Google. Further, in Engineering find the errors in the log file, but there is no existing system that
Knowledge base, we have multiple sources like bug tracking finds the solutions for the errors-Obtained after log search. By
system, Forum and even customer searchable knowledge base. which we can conclude that there is no existing system that
With this proposed tool-“Troubleshooter: Solution Finder for provides an efficient way for finding solutions from different
Log Errors”, we will be integrating all the knowledge sources solution sources within same framework. But there are
under one umbrella through which the team can search through. Knowledge sources from which user can get solutions
individually irrespective of each other Knowledge Bases.
The tool-Troubleshooter: Solution Finder for log errors
from Multiple Solution Sources provides a solution search from III. Solution search in multiple solution sources
different solution sources with respect to the products like SRM,
NCM/UIM, ViPR, SMARTS and Watch4Net. The tool “Troubleshooter: Solution Finder for log errors from
Multiple Solution Sources” - The objective of this tool is to
The Troubleshooter will also provide a search history provide an efficient way of finding solutions for log errors from
for the user. With search history one can directly go back to the single or multiple solution sources and display solutions within
previous search operations. The GUI of the tool is very simple the same framework- One can easilycorrelate the data available
and user friendly. Troubleshooter: Solution Finder for log errors with all the knowledge sourcesandfind the most relevant solution
from Multiple Solution Sources is a standalone tool so it needs among all the errors obtained from multiple solution sources for
to install on each user’s personal computer. This tool would an errors-Errors obtained by log file search.
reduce the time taken to search all knowledge sources
This tool provides the user to perform search in single or Obtained errors from the log files search are given as
multiple solution sources- Knowledge Base, Clear Quest, JIRA, input to this tool- Troubleshooter: Solution Finder for log errors
Forum and Documentation. We will be integrating all the from Multiple Solution Sources. Then the tool asks the user to
knowledge sources under one umbrella through which the team input the name of solution source(s) and name of the product
can search through. This would reduce the time taken to search based on user requirements. Once after getting proper input from
all knowledge sources independently and would help the team to the user the tool searches for given error in specified solution for
correlate the data available with other knowledge sources. the specified product based on user requirements. After solution
Additionally, this tool also has the capacity to search based on search the tool displays output from different solution sources on
the product selected and thus would help in filtering the result different tabs of display.
which are essentially for the selected product-SRM, NCM,
ViPR, SMARTS and Watch4Net. IV. Search History
The Tool will also provide a search history for the user, History contains an information such as Product selected-the
in case if the user facing same errors again and again very type of product that user has been selected, Searched for-The
frequently that time user can use the history to go back to his/her query for which user want to find the solution from multiple
previous search operations to get relevant solution instead of solution sources with respect to an product, Solution Sources-
searching solution from multiple solution sources for the same Name of solution source(s) where user want to search a query
error again. The GUI of the tool is very simple and user friendly, and Date and Time of search process.
the tool tips are provided for all the components of the tool so
that one can easily operate this tool. This tool is standalone so it V. Results
needs to install on each user‘s personal computer. This tool
reduces workload of user by giving an option to search the error A. Initial Screen
for solutions in multiple solution sources. The result with respect
to the different solution sources are display on different tabs of
the tabbed pane where each tabs will display names of solution
sources so that it is easy for the user to get from where the
solutions are fetched and tool will also provide products filtered
search results- The search is based on particular product based
on user requirement.
The tool also provided with some standard java errors
and also the bunch of errors with respect to the products- SRM,
NCM/UIM, ViPR, SMARTS and Watch4Net. This helps the
user to directly select an error from the provided list and perform
search operation instead of typing error manually and perform
search operation.
The tool is very useful for the user to find solutions
from different solution sources because it integrate all the
knowledge sources under one umbrella through which the team
can search through. Initially this tool will authenticate the user
and never ask again for the authentication. But without this tool Fig 2: Initial Screen of the Tool-Troubleshooter: Solution Finder
one need to authenticate by all solutions sources individually for log errors from Multiple Solution Sources
with irrespective of other solution sources. Without this tool the
user needs to give his/her username and password manually for This snapshot shows how an initial page looks. Initial
the multiple times whenever a solution source needs an page contains knowledge base search where user need to select
authentication this is again an overhead for the user. Knowledge sources and products based on his/her requirements
and post the query in the text field. Query can be entered
manually by user or can be selected from standard java errors list
or can be selected from the product based errors that the tool
provided.
display names of solution sources so that it is easy for the user to solutions from different solution sources are displayed on
know from where the solutions are fetched. One can see the different tabs of tabbed pane on display window. Where each
different tabs with the knowledge base name in initial screen fig tabs will display names of solution sources so that it is easy for
2. the user to get from where the solutions are fetched.
REFERENCES
i. http://www.xpolog.com/
ii. https://eventloganalyzer.codeplex.com/
iii. http://en.wikipedia.org/wiki/Piwik
iv. https://toolbox.googleapps.com/apps/loganalyzer/
v. http://en.wikipedia.org/wiki/JIRA
vi. http://www.google.com/custom?q
Fig 3: vii. http://nxlog-ce.sourceforge.net/
History for the searched results. viii. http://sourceforge.net/projects/syslog-analyzer/
ix. https://confluence.atlassian.com/display/JIRA/JIRA +Requirements
C. Solutions from Solution Sources.
Fig 4:
Solutions from CQ Knowledge source with respect to ViPR
product for the query selected from product based error.
sowmyaklrsv@gmail.com, rana.krishnakumar@citech.edu.in
ABSTRACT- Environmental Monitoring Systems and Sensors Agriculture environments are complex system where significant
systems have increased in importance over the years. However, changes in one environmental factor could have an adverse
increases in measurement points mean increases in installation effect on another. Environmental factors can affect survival and
and maintenance cost. Not to mention, the measurement points growth, in particular with regards to germination, sprouting,
once they have been built and installed, can be tedious to flowering and fruit development. They can also indicate
relocate in the future. Therefore, the purpose of this Master’s increased risk of disease and be used for prediction of upcoming
thesis is to present a project called “A web enabled wireless changes in the environment. It is therefore of particular interest
sensor network system for precision agriculture application to monitor these environmental factors in particular for any
using Internet of Things” which is capable of intelligently control and management systems that might be implemented.
monitoring agricultural conditions in a pre-programmed Temperature, humidity, pollution sensor, soil moisture, are the
manner. The proposed system consists of three stations: Sensor variables that are of interest to growers. Manual collection of
Node, Router, and Server. To allow for better monitoring of the data for desired factors can be sporadic, not continuous and
climate condition in an agricultural environment such as field produce variations from incorrect measurement taking. This can
or greenhouse, the sensor station is equipped with several cause difficulty in controlling these important factors. Sensor
sensor elements such as Temperature, humidity, pollution networks have been deployed for a wide variety of applications
sensor, and soil moisture. The communication between the and awareness has increased with regards to implementing
sensor node and the server is achieved via wireless ZigBee technology into an agricultural environment. Sensors are
modules. The overall system architecture shows advantages in becoming the solution to many existing problems in industries
cost, size, flexibility and power. It is believed that the outcomes with their ability to operate in a wide range of environments.
of the project allow for opportunities to perform further
Sensor nodes can reduce the time and effort required to monitor
research and development of a ZigBee based Wireless Sensor
an environment. This method reduces the risk of information
Network that is a portable and flexible type of sensing system
being lost or misplaced. It would also allow placement in critical
for an Agricultural Environment.
locations without the need to place personnel at risk. Monitoring
Index Terms – Internet of Things, Precision Agriculture, Zigbee systems can permit quicker response time to adverse factors and
based Wireless Sensor Network. conditions, better quality control to produce and lower labour
cost. The utilization of this technology would allow for remote
1. INTRODUCTION measurements of factors such as temperature, humidity, soil
moisture, pollution.
Agriculture products are dependent upon environmental factors
where plant growth and development are largely affected by the In Agriculture field of studies the sensing devices are mainly
conditions experienced. Similarly diseases that occur due to necessary for two intensions,
environmental factors can cause plant growth to be significantly
i. Sense and communicate with actuators
affected.
ii.To sense the parameter and send the information to remote
Agriculture environments such as fields and greenhouses allow
base station for expert analysis.
growers to produce plants with an emphasis on agriculture yield
and productivity. In addition, it also provides the possibility to In this paper study an attempt has been made to develop
grow plants in environments previously not suited for the task. KrishiSense.A Web enabled WSN system for agriculture
In particular, the use of greenhouse provides plants with application using IOT integrate the open Geospatial Consortium
protection from the harsh weather conditions, diseases and a specified Sensor Web Enablement standards on the sensing
controlled environment. system thereby enabling the interoperability between different
standardized sensing devices.KrishiSense is an interconnection
5. CONCLUSION
REFERNCES
cluster head selection minimizes the consumption of energy in characterized through novel stabilitynotions and shows which
WSN. transmission scheme is employed and which cluster nodes are
chosen to collaborate with the clusterhead. Extensive simulation
2. Literature Survey results are provided to demonstrate theeffectiveness of their
proposed game model and algorithm.
Our work is motivated by a number of prior works
related to clustering in wireless sensor network. Some of them Dali Wei et al [2] have proposed a distributed clustering
are analyzed here. algorithm, Energy-efficient Clustering (EC) that determines
suitable clustersizes depending on the hop distance to the data
OzlemDurmazIncel et al [17] have proposed a method for sink, while achieving approximate equalization of node lifetimes
Fast Data Collectionin Tree-Based Wireless Sensor Networks in and reducedenergy consumption levels. They additionally
their work, they explored andevaluated a number of different proposed a simpleenergy-efficient multi hop data collection
techniques using realistic simulation modelsunder the many-to- protocol to evaluatethe effectiveness of EC and calculate the
one communication paradigm known as converge cast. They end-to-end energyconsumption of this protocol; yet EC is
first consider time scheduling on a single frequency channel with suitable for any datacollection protocol that focuses on energy
the aim ofminimizing the number of time slots required conservation. Performanceresults demonstrate that EC extends
(schedule length) to complete a converge cast. Next, they network lifetimeand achieves energy equalization more
combined scheduling with transmission power control tomitigate effectively than two well-knownclustering algorithms, HEED
the effects of interference, and show that while power control and UCR.
helpsin reducing the schedule length under a single frequency,
scheduling transmissions using multiple frequencies is more Otgonchimeg Buyanjargal and Youngmi Kwon [22] have
efficient. They gave lower bounds on the schedule length when proposed a modified algorithm of Low Energy Adaptive
interference is completely eliminated, and propose algorithms Clustering Hierarchy (LEACH) protocol which is a well-known
that achieve these bounds. They also evaluate theperformance of energy efficient clustering algorithm for WSNs. Their modified
various channel assignment methods and find empirically that protocol called “Adaptive and Energy Efficient Clustering
formoderate size networks of about 100 nodes, the use of multi Algorithm for Event-Driven Application in Wireless Sensor
frequency scheduling can suffice to eliminate most of the Networks (AEEC)” is aimed at prolonging the lifetime of a
interference. Then, the data collectionrate no longer remains sensor network by balancing energy usage of the nodes. AEEC
limited by interference but by the topology of therouting tree. To makes the nodes with more residual energy have more chances
this end, they constructed degree-constrained spanning treesand to be selected as cluster head. Also, they used elector nodes
capacitated minimal spanning trees, and show significant which took the responsibility of collecting energy information of
improvement inscheduling performance over different the nearest sensor nodes and selecting the cluster head. They
deployment densities. Lastly, Theyevaluated the impact of compared the performance of their AEEC algorithm with the
different interference and channel models on the schedule LEACH protocol using simulations.
length. Dilip Kumar et al. [4] have studied the impact of
Guoliang Xing et al [8], have proposed a rendezvous-based heterogeneity of nodes in terms of their energy in wireless
data collectionapproach in which a subset of nodes serve as sensor networks that are hierarchically clustered. They have
rendezvous points that buffer andaggregate data originated from assumed that a percentage of the population of sensor nodes is
sources and transfer to the base station when itarrives. This equipped with the additional energy resources. They also
approach combines the advantages of controlled mobility and in- assumed that the sensor nodes were randomly distributed and
network data caching and can achieve a desirable balance were not mobile, the coordinates of the sink and the dimensions
between network energysaving and data collection delay. They of the sensor field were known. Homogeneous clustering
proposed efficient rendezvous designalgorithms with provable protocols assume that all the sensor nodes were equipped with
performance bounds for mobile base stations withvariable and the same amount of energy and as a result, they cannot take the
fixed tracks, respectively. The effectiveness of their advantage of the presence of node heterogeneity. Adapting this
approachwas validated through both theoretical analysis and approach, they have introduced an energy efficient
extensive simulations. heterogeneous clustered scheme for wireless sensor networks
based on weighted election probabilities of each node to become
Dan Wu et al [3], have proposed a method to focus on how a cluster head according to the residual energy in each node.
to select a proper transmission scheme, with the goal of Finally, the simulation results demonstrated that their proposed
improving the energy efficiency, e.g., prolonging the network heterogeneous clustering approach was more effective in
lifetime. In particular, they model the transmission scheme prolonging the network life-time compared with LEACH.
selection problem as a nontransferablecoalition formation game,
with the characteristic function based on the networklifetime. Xiang Min et al. [20] have presented the clustering algorithm
Then, a simple algorithm based on a merge-and-split rule and the mainly takes into account reducing the total energy consumption
Pareto order is proposed to form coalition groups among with optimum parameters. By optimizing the one-hop distance
individual sensor nodes.The resulting coalitional structure is and the clustering angle, all nodes are divided into static clusters
with different sizes, which was maintained the connectivity and References
reduce the energy consumption for inter-cluster communication. Buttyan, L. and Holczer, T., “Perfectly anonymous data aggregation in wireless
sensor networks Mobile Adhoc and Sensor Systems (MASS), IEEE 7th
Besides, with continuous working mechanisms for cluster head
International Conference on, pp. 513-528, 2010.
acting as the local control center, the frequency of cluster head i. Dali Wei, Yichao Jin, SerdarVural, Klaus Moessner, Rahim Tafazolli,
updating was reduced, and the energy consumption for the new "An Energy Efficient Clustering Solution for Wireless Sensor Networks", IEEE
cluster head set-up was reduced. With the clustering algorithm, Transactions on Wireless Communications, VOL. 10, NO. 11, 2011
ii. Dan Wu, YuemingCai, Jinlong Wang, " A Coalition Formation
the total energy consumption for inter-cluster and intra-cluster
Framework forTransmission Scheme Selection in Wireless Sensor Networks",
communications was reduced. The simulation results show that IEEE Transactions OnVehicular Technology, VOL. 60, No. 6 2011.
the system time is extended effectively. iii. Dilip Kumar, Trilok C. Aseri, R.B. Patel, “EEHC: Energy efficient
heterogeneous clustered scheme for wireless sensor networks”, Computer
3. Proposed Methodology Communications, Vol. 32, pp. 662-667, 2009.
iv. D. Tian, and N. D. Georganas,” A Node Scheduling Scheme for
Energy Conservation in Large Wireless Sensor Networks” Thesis,
In 2005, Karaboga proposed an Artificial Bee Colony Multimedia Communications Research Laboratory, School of Information
(ABC), which is based on a particular intelligent behavior of Technology and Engineering, University of Ottawa, 2002.
honeybee swarms.ABC is developed based on inspecting the v. Ewa Hansen, Jonas Neander, Mikael Nolin, Mats Björkman ,
"Efficient Cluster Formation for Sensor Networks ", march 2006, “MRTC report
behaviors of real bees on finding nectar and sharing the ISSN 1404-3041 ISRN MDH-MRTC-199/2006-1- SE, Mälardalen Real-Time
information of food sources to the bees in the hive. Research Centre, MälardaleUniversity, March, 2006"
vi. G. Pei and C. Chien,”Low Power TDMA in Large Wireless Sensor
Agents in ABC are The Employed Bee, The Onlooker Bee and Networks”, Military Communications Conference, vol.1, pp.347–351, 2001.
The Scout. vii. Guoliang Xing, M Minming Li, Tian Wang, WeijiaJia and Jun Huang,
"Efficient Rendezvous Algorithms for Mobility Enabled Wireless Sensor
The Employed bees: It stays on a food source and provides Networks", IEEE Transactions On Mobile Computing, VOL. 11, No. 1 2012.
the neighborhood of the source in its memory. viii. Hnin Yu Shwe, JIANG Xiao-hong, Susumu Horiguchi, “Energy
saving in wireless sensor networks”, Journal of Communication and Computer
The Onlooker bees: It gets the information of food sources Volume 6, No.5, 2009
ix. I.F Akyildiz, W. Su, Y. Sankarasubramaniam, E. Cayirci, “A survey
from the employed bees in the hive and select one of the on sensor networks. IEEE Communications Magazine”,pp.102–114, 2002.
food source to gathers the nectar. x. J. Kulik, W. Heinzelman, and H. Balakrishnan, “Negotiation-based
protocols for disseminating information in wireless sensor networks,” Wireless
The Scout: It is responsible for finding new food, the new Networks, vol. 8, no. 2/3, pp. 69–185, 2002.
nectar, and sources. xi. J.M. McCune,”Adaptability in sensor networks” Undergraduate
Thesis in Computer Engineering, University of Virginia, April 2003.
Procedure of ABC: xii. K. Intae and R. Poovendran,“ Maximizing static network lifetime of
Initialize (Move the scouts). wireless broadcast ad hoc networks,” in Proceedings of the IEEE International
Move the onlookers. Conference on Communications pp. 2256–2261.11–15, 2003.
xiii. Liyang Yu , Neng Wang , Wei Zhang and Chunlei Zheng,”GROUP: a
Move the scouts only if the counters of the employed bees hit Grid-clustering Routing Protocol for Wireless Sensor Networks”, In proceedings
the limit. of Wireless Communications, Networking and Mobile Computing, pp. 1 - 5 ,
Update the memory 2006.
Check the termination condition xiv. M. Gerla, T. Kwon, and G. Pei,” On Demand Routing in Large Ad
Hoc Wireless Networks with Passive Clustering”, proceedings of IEEE Wireless
Thus ABC procedure optimizes by choosing the cluster head and Communications and Networking Confernce, pp.100-105.
then gives efficient sensor readings with less number of nodes xv. Mohammad Zeynali, Leili Mohammad Khanli and Amir Mollanejad
and reduced energy consumption. “TBRP: Novel Tree Based Routing Protocol in Wireless Sensor Network”,
International Journal of Grid and Distributed Computing, Vol. 2, No. 4, 2009.
xvi. OzlemDurmazIncel, Amitabha Ghosh, Bhaskar Krishnamachari, and
4. Objectives Krishnakant Chintalapudi, "Fast Data Collection in Tree-Based Wireless Sensor
Networks", IEEE Transactions On Mobile Computing, Vol. 11, No. 1 2012.
xvii. W. Heinzelman, A. Chandrakasan, and H. Balakrishnan,” Energy-
A study over the recent techniques for developing a Efficient Communication Protocol for Wireless Microsensor Networks”, Maui,
cluster head selection technique for wireless sensor network. Hawaii, In Proceedings of the 33rd International Con-ference on System
Sciences, 2000.
Developing an optimization technique for the selection xviii. W. Ye, J. Heidemann, and D. Estrin, “An Energy-Efficient MAC
of cluster head for dynamic environment in wireless sensor Protocol for Wireless Sensor Networks,” in Proceedings of IEEE INFOCOM,
network to achieve energy efficient aggregation of sensor pp. 1567-1576, 2002.
readings from cluster head to Base station (BS). xix. Xiang Min, Shi Wei-ren, Jiang Chang-jiang and Zhang Ying, “a
Energy efficient clustering algorithm for maximizing lifetime of wireless sensor
Analysis of the proposed technique using various simulations set networks”, AEU- International journal of electronics and communications,
up with different existing techniques. vol.64, no. 4, pp. 289-298, 2010.
xx. Xianghui Wang and Guoyin Zhang, “DECP: A Distributed Election
5. Possible outcome Clustering Protocol for Heterogeneous Wireless Sensor Networks”,
computational science, vol. 4489/2007, pp. 105-108, 2007.
The expecting outcome of the paper is the remarkable reduction xxi. [22]. Ye, M.; Li, C.F.; Chen, G.; Wu, J.,” EECS: An Energy
of energy consumption because of the dynamic and efficient Efficient Clustering Scheme in Wireless Sensor Networks” In Proceedings of
optimization technique for the selection of cluster head in sensor the IEEE International Performance Computing and Communications
networks. Conference, pp.535-540, 2005
IV. CONCLUSION
Abstract: Opportunistic data forwarding has become a hot Figure 1 shows an example of mobile ad-hoc network and its
topic in the multihop wireless networking. Opportunistic data communication technology. As shown in Figure, an ad hoc
forwarding is not used in mobile ad hoc networks (MANETs) network might consist of several home-computing devices which
due to the lack of an efficient lightweight proactive strong includes laptops, cellular phones etc. Each node can
source routing scheme. Proactive Source Routing uses communicate directly with any other node that resides within its
Breadth First Spanning Trees (BFSTs) and maintains more transmission range. The node needs to use intermediate nodes to
network topology information to facilitate source routing. It relay the messages hop by hop to communicate with nodes that
overhead is much smaller than traditional DV-based protocols, reside beyond this range.
link state (LS)-based routing protocols and reactive source Opportunistic data forwarding utilizes the broadcast
routing protocols but the computational and memory overhead nature of wireless communication links [iv] and data packets are
involved in maintaining BFSTs to reach every node in the handled in a multihop wireless network. In traditional IP
denser networks is high. In this paper Zone-based Proactive forwarding, the intermediate nodes looks up a forwarding table
Source Routing Protocol is proposed. Zone routing protocol to find a dedicated next hop, but Opportunistic data forwarding
(ZRP) uses partition based routing. It uses Source routing broadcasts the data packet and allows potentially multiple
inside the zone and on-demand routing outside the zone. The downstream nodes to act on the packet. One of the initial works
advantages of both proactive and zone based routing protocols on opportunistic data forwarding is selective diversity
is combined by this approach. The simulations shows that the forwarding by Larsson [v]. In this paper the transmitter sends the
Z-PSR, zone based proactive source routing protocol performs packet to multiple receivers and selects best forwarder from
better compared to PSR. these receivers which successfully receives the data and requests
Keywords: PSR, BFST, Link State, Source routing, Ad-hoc the selected node to forward the data. The overhead in this
Network. approach is more and it should be reduced before it can be
implemented in practical networks. This issue was addressed in
I. INTRODUCTION the seminal work on ExOR [vi], which outlines a solution at the
Mobile ad-hoc network (MANET) is a self-organized and self- link and network layers. In ExOR, all the nodes in the work are
configurable wireless Communication network. It represents enabled to overhear all packets on the air and therefore, more
complex distributed systems that contains various wireless number of nodes can potentially forward a packet, provided that
mobile nodes which can freely move and dynamically self- all the nodes should be included in the forwarder list which is
organize into arbitrary and temporary ad-hoc network carried by the packet. The contention feature of the medium-
topologies. It allows people and devices to seamlessly access-control (MAC) sublayer effectively utilized and hence the
internetwork in areas without pre-existing communication forwarder which is very much closer to the destination will
infrastructure, e.g., battlefield communications, emergency access the medium. Therefore, the MAC sublayer can determine
operations, disaster recovery environments. A great deal of the actual next-hop forwarder to utilize the long-haul
research results have been published since its early days in transmissions in a better way.
1980s [i]. The salient research challenges in this area are A lightweight proactive source routing(PSR) protocol
link access control, security, end-to-end transfer and providing is proposed to facilitate opportunistic data forwarding in
support for real-time multimedia streaming [ii]. In the research MANETs. In this protocol, each node maintains a breadth-first
on MANETs, the network layer has received a considerable search spanning tree of the network rooted at itself. This routing
amount of attention. Hence large number of routing protocols information is periodically exchanged among neighbouring
with differing objectives for various specific needs have been nodes for updated network topology information. And hence
proposed in this network [iii]. PSR allows a node to have full-path information to all other
nodes in the network. The communication cost is only linear to
the number of the nodes. Thus, it supports both source routing
and conventional IP forwarding. But the computational and
memory overhead involved in maintaining the BFSTs to reach
every node in the denser networks will be high.
In this paper, Z-PSR (Zone based proactive source
routing protocol is proposed which is lightweight, source
routed, uses Breadth First Spanning Trees and is based on
Fig 1: Mobile Ad-hoc Network PSR[vii] and ZRP[viii].
at each node
The objectives of the Zone-based proactive source 40
routing protocol are as follows
30 Z-PSR
1. Develop a routing protocol which minimizes the computation
PSR
overhead in searching for a route. 20
2. The protocol should reduce the memory occupied by each
BFST. 10
3. The protocol should find route to the destination with 0
minimum delay. 1 2 3 4 5 6 7 8 9 10111213141516
4. The minimize energy consumption compared to the existing
PSR protocol. Node Id
The following steps are taken, in order to meet the Fig.5 Length of BFST at each node
above objectives.
1. Each node will maintain a BFST of its one hop or two hop Below Fig.6 Shows the Packet delivery ratio, for both
neighbours only, as opposed to PSR where every node needs to PSR and Z-PSR, Z-PSR maintains a 99.9% delivery ratio.
maintain BFST to reach every other node in the network.
2. Whether to maintain one hop or two hop neighbours BFST is
decided based on parameter radius. If radius = 1, maintain BFST
1.2
to reach one hop neighbours. If Radius =2, maintain BFST to
reach two hop neighbours and so on. The Simulations in this
1
paper has used radius 2.
3. When a node needs to send data to its one hop or two hop
Packet delivery ratio
ABSTRACT: In a private database query system a client issues without learning what the query was or even how many records
queries to a database server and obtains the results without match the query. The client learns nothing else about the
learning anything else about the database and without the database contents.
server learning the query. In this work we develop tools for Unfortunately, being a generalization of SPIR, private database
implementing private database queries using homomorphic queries is subject to all the same inherent inefficiency constraints
encryption (HE), that is, using an encryption system that as SPIR, making the design of practical schemes for private
supports only limited computations on encrypted data. We database queries a challenging task. In this work we explore the
show that a polynomial encoding of the database enables an use of homomorphic encryption (SWHE) [3] for the design of
efficient implementation of several different query types using private database query protocols. In particular, we show that
only low-degree computations on cipher texts. Specifically, we certain polynomial encodings of the database let us implement
study two separate settings that offer different interesting query types using only homomorphiccomputations
privacy/efficiency tradeoffs. In the basic client-server setting, involving low-degree polynomials. There are now several
we show that additive homomorphisms are sufficient to encryp- tion schemes that efficiently support the low-degree
implement conjunction and threshold queries. We obtain homomorphiccomputations on encrypted data that we need [4,
further efficiency improvements using an additive system that 5].
also supports a single homomorphic multiplication on In this work we consider two different settings. The first is the
ciphertexts. This implementation hides all aspects of the traditional, two-party, client- server setting. In this setting the
client’s query from the server, and reveals nothing to the client server has the database, the client has a query, and we seek a
on non-matching records. To improve performance further we protocol that gives the client all (and only) those records that
turn to the “Isolated-Box” architecture of De Cristofaro et al. match its query without the server learning what the query is. As
In that architecture the role of the database server is split mentioned above,in this setting the server must process the
between two non-colluding parties. The server encrypts and entire database for every query (or else it would learn that the
pre-processes then-record database and also prepares an unprocessed records do not match the query). Moreover the
encrypted inverted index. The server sends the encrypted server has to return to the client as much data as the number of
database and inverted index to a proxy, but keeps the records in the database, or else it would learn some information
decryption keys to itself. The client interacts with both server about the number of records that match the query.
and proxy for every query and privacy holds as long as the
server and proxy do not collude. We show that using a system To bypass these severe limitations, we consider also a different
that supports only log(n) multiplications on encrypteddata it is model in which the database server is split into two entities
possible to implement conjunctions and threshold queries (called here “server” and “proxy”), and privacy holds only so
efficiently.We implemented our protocols for the Isolated-box long as these two entities do not collude. This approach was
architecture using the ho- momorphic encryption system by taken in particular by De Cristofaro et al. [6], where they support
Brakerski, and compared it to a simpler implementation that private evaluation of a few simple query types and report
only uses Paillier’s additively homomorphic encryption system. performance very close to a non-private off-the-shelf MySQL
The implementation using some- what homomorphic system. However, the architecture of De Cristofaro et al. cannot
encryption was able to handle a query with a few thousand handle conjunctions: the client can ask for all the records with
matches out of a million-record database in just a few minutes, age=25 OR name=‘Bob’, but cannot ask for all the records with
far outperforming the implementation using additively age=25 AND name=‘Bob’. In this work we show how to
homomorphic encryption. implement conjunctions, disjunctions, and threshold queries in a
similar architecture.
Keywords: Cipher Text, Homomorphic Encryption,
Threshold, Non Colluding Parties. 1.1. Our Protocols
The protocols and tools we present in this work are
1. INTRODUCTION aimed at revealing to the client the indexes of the records that
Enabling private database queries is an important (and match its query, leaving it to a standard follow-up protocol to
hard) research problem arising in many real- world settings. The fetch the records themselves. Also, we only consider honest but
problem can be thought of as a generalization of symmetric curious security in this work. Our protocols can be enhanced to
private information retrieval (SPIR) [1, 2] where clients can handle malicious adversaries using generic tools such as [7]. It is
retrieve records by specifying complex queries. For example, the an interesting open problem to design more efficient protocols in
client may ask for the records of all people of age 25 to 29 who the malicious settings specific to the private database queries
also live in Bangalore, and the server should return these records problem.
ABSTRACT : Distributed denial of service (DDoS) is a monitor packets on the network and compare them against a
rapidly growing problem. Traditional architecture of the database of signatures or attributes from known malicious
internet is more exposed to Bandwidth distributed denial of threats. In Anomaly-based intrusion detection will monitor
service (BW-DDoS) attacks. These attacks disrupt network network traffic and compare it against base profile. The base
infrastructure operation by sending a huge number of packets profile will identify what is normal for that network and what
to cause congestion and delayed response. Attacker disrupts sort of bandwidth is used and what protocols and ports are
connectivity between client and server. According a recent used, and devices are connected to each other and alert the
survey of Akamai’s Prolexic Quarterly Global DDoS Attack user when traffic is detected which is anomalous and
Report Quarter 4 of 2014 Compared to Q1 of 2014 then find significantly different than the base profile. In this paper, we
39% increase in bandwidth-DDoS attack. In this paper, we used anomaly intrusion detection method to identify the
exposed the different types of BW-DDoS attacks on the intruder in the network.
internet and also we build an intrusion detection system to
detect DDoS attacks. II. Material and Methodology
Keywords — DDOS, Bandwidth-DDoS, Internet, We observed some information on DDoS attack statistics
Congestion. obtained in the first quarter of 2014 on networks of various
sectors in the world including financial sector networks. The
I. I nt ro d uct io n source of data is Prolexic Attack Report Q4 2014 [2] provided
The internet is a group of two or more devices or nodes or by Prolexic Technologies. The world largest and most trusted
terminals which are connected by a large number of network DDoS attack mitigation provider. Ten of the world’s largest
devices. Denial of Service (DoS) attacks is very common in the banks and the leading e-commerce companies get the services
world of internet today. A distributed denial of service (DDoS) of Prolexic to protect themselves from DDoS attacks. The
attack is a form of DoS, which uses multiple machines to range of data is based on all DDoS attacks dealt by Prolexic
prevent the legitimate use of services. Internet services are in different regions of the world. Some key information
more exposed to Bandwidth DDoS attacks. The increase of extracted from the report regarding the comparison of first
these attacks has made servers and network devices at risk. quarter of 2014 with the last quarter of 2014 is:
BW-DDoS aim to deny normal services for legitimate users by i) Total number of DDoS attacks was increased by
sending huge traffic to the machines or networks to exhaust 25%.
services, connection and the bandwidth. The BW-DDoS ii) Total number of BW-DDoS attacks increased by
attacker uses the different methods and attacking agents like 39%.
zombies. Zombies are groups of computers connected to iii) 60 to 86.5 percent of BW-DDoS attacks targeted
internet that compromised by an attacker and it can be used to the network.
perform malicious task on victim and Fig.1.1 shows the attacker iv) A decline was observed in UDP flood attacks
uses three zombies to generate high volume of malicious traffic
to network over Internet cause legitimate user unable to access 2.1. Motivation behind BW-DDoS Attacks
the services. The motivation behind the BW-DDoS attacks by personal,
social or financial benefits. Attacker may do so due to
personal revenge, getting publicity or some political
motivation. However, most BW-DDoS attacks are launched
by organized groups targeting financial websites such as
banks or stock exchange.
2.2.3. Amplification attack The simulation is implemented in the java platform. In our
Amplification attack most effectively uses the zombie’s simulation, we used some parameters to establish a proposed
bandwidth. Each packet sent by a compromised computer cause system to identify intruder in the network and how
transmission of large packets to the victim by non compromised performance of network affected by these attack? Simulation
machines. The response data must be larger than the request parameters are provided in table I. we implement router based
data in size. The larger the amplification means effective identification to send queries to the destination based on the
bandwidth consumption. bandwidth provided to the intermediate nodes. We observed
DNS Amplification Attack that attacker like zombies act like it’s having higher
Domain Name System (DNS) is a core service of the Internet. bandwidth to transfer query to destination. It leads to access
Since the DNS response packets are larger than the query traffic to network and delays the response from the network.
packet. The attackers send queries to the open DNS resolvers Table I Simulation parameters
with large size UDP messages and spoof the source IP address Number of nodes 10
as the target address. Upon receiving the query request, the Mac 802.11
DNS resolver will send back the resolution to the attack target.
Flooded by large quantities of resolution responses, the target Simulation Time 20sec
will suffer network congestion leading to the Bandwidth
distributed denial of services. Traffic Source CBR
Packet Size 512
2.3. Methods for Attack detection and mitigation
The proposed system used to identify the attacker and traffic in Dimension of area 800×600
the network. These methods includes filtering, rate limiting,
detouring method are used.
References
Prathibha
Deptt of CSE, VTU, SJBIT Bangalore, Karnataka India
pattidiggavi@gmail.com
Abstract: Many search engines are used to search anything number of objects to be examined. The IR2-tree, however, also
from anywhere , this system is used to search fast nearest inherits a drawback of signature files: false hits. That is, a
neighbor using keyword in which the existing system works signature file, due to its conservative nature, may still direct the
mainly on finding top-k nearest neighbor, whereas each node search to some objects, even though they do not have all the
has to match whole querying keywords. It does not consider the keywords. To overcome such problems involved spatial inverted
density of objects in spatial space also this are low efficient for index which will convert in tuples and find the matching
incremental queries. The existing system works on Ir2-Tree but keyword through latitude and longitude using the nearest
it is not such efficient, to overcome this the triplet form of neighbour search. Spatial keyword query typically takes a
spatial index are introduced and which is accurate and consist location and a set of keywords as input parameters and returns
of efficient responsive time. the matched objects according to certain spatial constraints and
Keywords—query keywords,NN search,IR2tree textual patterns.
B) Hybrid index
Fig.2 The circle around query object q depicts the search region
after reporting o as next nearest object. The figure above shows the comparision result of the both
existing IR2 tree and the proposed NN search algorithm in mili
We design a variant of inverted index that is optimized for seconds and prove the effieciency of the NN search.
multidimensional points, and is thus named the spatial inverted
index (SI-index).This access method successfully incorporates V. CONCLUSION
point coordinates into a conventional inverted index with small There are many applications seen for calling a search engine
extra space, owing to a delicate compact storage scheme. thats ready to with efficiency support novel varieties of
Meanwhile, an SI-index preserves the spatial locality of data abstraction queries that are integrated with keyword search. The
points, and comes with an R-tree built on every inverted list at present solutions to such queries either incur preventative space
little space overhead. As a result, it offers two competing ways consumption or are unable to provide real time answers. The
for query processing. planned system has remedied the situation by developing an
• We can (sequentially) merge multiple lists very much like access methodology referred to as the abstraction Inverted index
merging traditional inverted lists by ids. (SI-index). Not solely that the SI-index is fairly space
• Alternatively, we can also leverage the R-trees to browse the economical, however additionally its the flexibility to perform
points of all relevant lists in ascending order of their distances to keyword-augmented nearest neighbour search in time thas at the
the query point. order of dozens of milliseconds. Moreover, because the SI-
index relies on the standard technology of inverted index, its
NN Search Algorithm readily incorporable in a business search engine that applies
huge similarity, implying its immediate industrial merits.
Figure3 shows the flow diagram of the NN search in which first
it gives the latitude and longitude with keywords and range and
then it calculates the distance of nearest one and sort the REFERENCES
distances in ascending order and filter the places and match the i. D. Felipe, V. Hristidis, and N. Rishe. Keyword search on spatial
keywords with query keywords and then finally give matched databases. In Proc. of International Conference on Data Engineering (ICDE),
placed list. pages 656–665, 2008.
ii. X. Cao, L. Chen, G. Cong, C. S. Jensen, Q. Qu, A. Skovsgaard, D.
Distance Wu, and M. L. Yiu. Spatial keyword querying. In ER, pages 16–29, 2012.
Lat,lang Sorting
keyword range
calculation
Process
Filtered Places iii. G. Cong, C. S. Jensen, and D. Wu. Efficient retrieval of the top-k most
Process
relevant spatial web objects. PVLDB, 2(1):337–348, 2009.
iv. R. Hariharan, B. Hore, C. Li, and S. Mehrotra . Processing spatial
keyword (SK) queries in geographic information retrieval (GIR) systems. In
Keyword
Proc. of Scientific and Statistical Database Management(SSDBM), 2007.
Matched v. Yanwei Xu,Jihong Guan,Fengrong Li,shuigeng Zhou Scalable
matching
places list
algorithm continual top-k keyword search in relational databases. Data and knowledge
Engineering 86(2013)206-223.
Fig.3 Flow diagram of NN Search vi. Hristidis and Y. Papakonstantinou. Discover: Keyword search in
relational databases. In Proc. of Very Large Data Bases (VLDB), pages 670–
681, 2002.
IV. RESULT AND ANALYSIS I. Kamel and C. Faloutsos. Hilbert R-tree: An improved r-tree using
When look on to the comparisons between existing and proposed fractals. In Proc. of Very Large Data Bases (VLDB), pages 500–509, 1994.
system, the primary set of experiments is to check the vii. Lu, Y. Lu, and G. Cong. Reverse spatial and textual k nearest
neighbor search. In Proc. of ACM Management of Data (SIGMOD), pages 349–
performance of various mixtures of fast neighbour search and 360, 2011.
existing search methods. All methods are tested below two viii. S. Stiassny. mathematical analysis of various superimposedcoding
request patterns: information analysis and results. In additional methods. Am. Doc., 11(2):155–169, 1960.
specific the chapter particularly curious about the overall ix. D. Zhang, Y. M. Chee, A. Mondal, A. K. H. Tung, and M.
Kitsuregawa. Keyword search in spatial databases: Towards searching by
number of results and search delay during a spatial data search document. In Proc. of International Conference on Data Engineering (ICDE),
and also the average interval of an information extraction since pages 688–699, 2009.
they are the dominant factors affecting service quality
experienced by the users.
Abstract : The fundamental technique adopted in data mining method is used to construct the tree like structure and frequent
to retrieve the data from the database are apriori algorithm, fp item set for the data stored in the vertical data format.
tree, ecalt algorithm using the rule of data mining like
association rule, classification, clustering etc. The apriori There are many technique to find out the frequent item set but
algorithm traverse many times into database to generate they were having many drawbacks like it take more time to
frequent item and it take more space, where as the fp tree is construct the tree structure and they are becaming more
advantageous compare to the apriori algorithm but it will not complex to understand and it need more space to store data. So
use the memory up to mark. So to overcome the drawback of overcome this drawback, in recent years Deng and Wang gave
this algorithm, in this paper we are adopting new technique us the new technique called prepost code to generate the
called ppc-tree, this tree is constructed based on the pre -post frequent dataset, this method is based on the fp tree structure.
traverse, the prepostalgorithm it construct the tree based Here the data will be stored in the form of tree like structure.
vertical traverse of the database, it also scan the database twice The prepostcode(ppc tree) is having two step of execution. First
and construct the tree, the ppc tree look similar to fp-tree but it construct the tree like structure by traversing into data set and
tree constructed in the vertical. time and space utilized by this then using the tree structure it construct the frequent item set
algorithm is less compare to other technique,the experiment using the aprori algorithm.
show the performance, stability and scalability of the
algorithm. 2 Related work
Keywords: apriori algorithm, fp tree, ppc tree, prepost The algorithm that we are using now for mining the frequent
algorithm itemset is the combination of the apriori method and fp growth
method. The apriori methodis scanning the database and prune
1 Introduction for the frequent item set. apriori algorithm will work based on
the candidate generate and test strategy, it scan the database of
The new technique was proposed for data mining that is the n item, if item present in the database is not frequent then it
frequent item set. It was proposed by threepeople, they are generate the k candidate itemset of frequent item, then this
Agrawal, Imielinski, and Swami(1993). Since the frequent item frequent itemset will used for generating the next frequent
set is a technique it used with basic technique of datamining like item,this procedure will continue till the all frequent item
classification ,clustering, etc..And based on this technique new generate, after all frequent itemset is generated till purne the
algorithm was proposed for every application, that algorithm database. 1994), adopt the Apriori-like method.
provide more efficiency, scalability,optimal method.The
frequent item set thecnique is further classified into three group. The advantages of Apriori-like method is it provide good
performance by reducing the size of candidates. The apriori
1Technique based candidate generate-and-test strategy: this method is very expensive and here we need to scan the database
is the basic technique of dataming, here data set will generated repeatedly and then we need to check a large set of candidates in
repeatedly until all the candidates generated. Its work like first it database for matching item.
generate the one set candidate and further the first generated
candidate set is used to generate the next set of candidate set and The fp growth will store the data in database using datastructure
this set used to generate next set of candidate.This method will tree called fp tree, which will not use the candidate generation
be continued till the all candidate is generated. method and it will use the partition, divide and conquer method
to store the data. Advantages of fp tree is it will reduce the
2Technique based on divide and conquer strategy: this search space and will generate frequent time set without using
method of dataset is compressed using divide and conquer the candidate generation.
method to construct the tree like structure like fp tree and
frequent item etc…this fptree were used to understand the The ppc algorithm will work based by combining the
reduce the space and to increase the time efficiency advantages of apriori algorithm and fp growth.
2 Each node of the subtree consists of five fields namely: item- Table 1: transaction database
name, count, childNode-list, pre-order, and post-order. item –
name will specific the frequent item of that node. Count Ordered
valuesay the number of transactions presented by the portion of Id Item frequent items
the path reaching this node. childNode-listsay the number of the 1 a, c, g c, a
children of the node. pre-orderthis say about the preorder rank 2 e, a, c, b b, c, e, a
of the node. post-orderthis say about the postorder rank of the 3 f, e, c, b, i b, c, e, f
node. 4 b, f, h b, f
5 b, f, e, c, d b, c, e ,f
The difference between the fp tree and ppc tree
Obviously, the second column and the last column are
1. equivalent for Fmining frequent patterns under the given
P-tree has twofield in each node ,one is node list and another one minimum support threshold. In the last columns of Table 1, all
is header table structure its hold the connection between node infrequent items are eliminated and frequent items are listed in
which same item node ,but in ppc we don’t have that ,here we support-descending order. This ensures that the DB can be
are using preorder and post order method . efficiently represented by a compressed tree structure.
2. For Pre-Post code I generation, we traverse the PPC-tree twice by
n ppctree every node will be having the preorder filed and post preorder and postorder. After that, we get the Figure 1. In this
order field, but in fp tree we are not having this.the preoder of
6 Reference
i. Agrawal R, Srikant R. Fast algorithms for mining association
rules[C]lProc. 20th int. conf very large data bases, VLDB. 1994, 1215487-499.
ii. Deng Z H, Wang Z H, Jiang J .I. A new algorithm for fast mining
frequent itemsets using N- lists[J]. Science China Information Sciences,
Fig: ppc tree structure 2012,55(9) 2008-2030.
iii. Savasere A, Omiecinski E, Navathe S. An efficient algorithm for
mining association rules in large databases. In: The 21th International
4 Experimental result Conference on Very Large Data Bases (VLDB'95), Zurich, 1995. 432- 443.
The three computers have same configuration, CPU is AMD iv. H. Mannila, H. Toivonen, and A. Verkamo.Efficient algorithm for
discovering association rules. AAAI Workshop on Knowledge Discovery in
Athlon dual-core processor, clocked at 2.11GHz, memory size
Databases, pp. 181-192, JuL 1994.
2G. T10I4D100K and Pumsb act as experimental data. We v. Shi Yue-mei. Hu Guo-hua. A Sampling Algorithm for Mining
compare the runtime of three algorithms Pre Post, FP tree and Association Rules in Distributed Database[C]. In:2009 First International
apriori algorithm when they are performed on the two Workshop on Database Technology and Applications, 2009, 431-434.
datasets.From experimental results shown above we know the vi. Han J, Pei J, Yin Y Mining frequent patterns without candidate
generation[C]/ACM SIGMOD Record. ACM, 2000, 29(2): 1-12.
runtime will become shorter when support increases. vii. Mobasher B, Dai H, Luo T, et aL Effective personalization based on
It'sevident.also reflects performance of the parallel algorithm is association rule discovery from web usage data[C]lProceedings of the 3rd
not as good as PrePost on small dataset.The reason is each node international workshop on Web information and data management. ACM,
needs to send message to others in clusters, but delay of network 20019-15.
bandwidth is unpredictable, so I/O operation occupies main
runtime, thus affecting the performance of the algorithm.
Contrarily, Pre Post has an advantage of data localization. But
when the dataset is large, PrePost at a lower support threshold
can not be performed due to memory overflow.
Graph1: Run time using the prepost ,fp tree, apriori algorithm
Keywords—Ontology based text mining, Classification, But this is not suitable for large data because it might make
Clustering. misplacements of proposals to wrong groups due to manual
process. Misplacement of proposals can be happen for following
I. INTRODUCTION reasons. First, keywords might give an incomplete meaning
In computer science, ontology can be said as set of concepts about whole proposals. Second, keywords which are provided
that is knowledge within domain and relation between the by applicants may have misconception and also we can say
pairs of concepts. Ontology is used in various domains as a keywords will give only partial representation of proposals.
form of knowledge representation about the world. In this Third, manual grouping which is done by area expert.
project, ontology is a model for describing the world that gives
the mapping between the properties and relationship types III. BACKGROUND
which gives the close relationship between the ontology and real This project uses the concept of ontology based text mining
world. approach such as classification and clustering algorithms. The
proposed system builds the research ontology and applies the
Research project selection is an important task in government as decision tree algorithm to classify the data into the disciplines
well as private funding agencies. It is a challenge for multi using created ontology and then the resultant of classification is
process task that begins with a call for proposals by the funding helps to make clusters of similar data.
agencies. Earlier it was a manual method for classifying but
this method has extended from manual to automatically to be A. Ontology
done based on feature vector value. After submission of the Ontology has several technical advantages like flexibility and
proposals, the next step is to apply the preprocessing step like easily accommodates heterogeneous data. Nowadays,
data cleaning to remove all the stop words from the proposals. ontology has become a prominent in the research field
especially in computer science. Ontology is knowledge
The web technology has defined many stop words. By applying repository which defines the terms and concepts. And also
the preprocessing step like data cleaning, can remove all the stop represent the relationship between various concepts. It’s a tree
words from all the submitted proposals. The Obtained clean data like structure defined by author Gangemi A, 2005. Ontology
words can be considered as tokens, assigned with unique id to in this paper is created by submitting proposals which
each tokens. Then calculate the number of times the token has containing the keywords, which are the representation of overall
been repeated, which gives the frequency of tokens. Now apply project. Creating a list of keywords from specific area itself is
the frequency tokenized algorithm for calculating the inverse an area of ontology [2]. By creating this it will be easy to
document frequency text which gives number of classify the proposals into their respective area by checking
documents or proposals. By multiplying the frequency of number of times words have been appeared in paper.
text to the obtained, IDFT value, will get the feature vector
value. Finally the proposal whose feature vector value is highest B. Classification
that will appear at the top of list in descending order based on is Based on the data, input text data can be classified into
value. number of classes in classification. Various text mining
techniques are used for classification of text data such as
C. Clustering
Number of similar objects collected and grouped together is
called a cluster. Following are few of the definitions of the
cluster. number of documents to the frequency. And by
multiplying this value with frequency will get feature vector
1. A cluster is a set of entities which are like and entities value. Finally we can rank the papers based on the feature vector
from different clusters are not alike. value.
2. A cluster is an aggregation of points in the test space
such that distance between any two points in the cluster is less V. METHODOLOGY
than the distance between any point in cluster. In this paper research projects are clustered into specific area
3. Clusters are connected regions of multi dimensional using ontology of different areas. Following are the modules for
space containing high density of points separated by low density proposed system and is also shown in Figure 2.
of points.
Module 1: In the first module users have to submit the
4. Clustering means grouping of similar types of objects
into one cluster proposals. At a time five proposals can be submitted.
Proposals along with their abstract will be sent and those will be
stored on ontology.
Clustering is a technique used to make group of documents Module 2: By applying preprocessing step like data cleaning,
having similar features. Documents within cluster have we can remove all the stop words from proposals. Then
similar objects and dissimilar objects as compared to any obtained cleaned data will be given as input to next module.
other cluster. Clustering algorithms creates a vector of topics for
document and measures the weights of how well the
documents file into each cluster [9].
Abstract-The amount of content in the Log Files are domain from which the messages must be interpreted. They also
increasing drastically, manual method to find the errors in provide solution to the bugs once they are examined.
those files and fix the issue is becoming very difficult and
complex, time consuming and it is not a very efficient method Similarly the products like W4N, ViPR, NCM, ECS,
to be followed. Log Analyzer Tool overcomes all the difficulties SRM, SMARTS etc. also generates Log files which are to be
faced so far. It is highly automated with advanced analyzed by the log analyst to fix them and ensure the products
functionalities which are not provided by the other tools. Log are bug free in order to keep the product up and running without
Analyzer facilitates the Analyst’s to find bugs in the log file any interruption. If any bugs found it has to be fixed as soon as
with less efforts. Few Tools fail to open one complete file into possible to avoid other malfunctions. To find out the bugs easily
the tool to perform search, this tool not just searches the errors in the log files and to fix them at the earliest LOG ANALYZER
in the single file it searches for errors in multiple files and was developed. This tool helps the analyst to find out the bugs
even an entire folder. It displays the result to the user with within the log files in many ways, it also provides 3 main
necessary highlighters and other options. It provides simple functionalities to locate a bug in the Log Files.
and advanced search options to the user.
III. SCOPE
Keywords- Log Analyzer, Simple Search, Keyword Search,
Date and Time Range Search. This project ensures finding bugs within the log files providing
the below features:-
I. INTRODUCTION
Standalone Desktop Based Tool.
Log analysis (or system and network log analysis) is an
Easy, Friendly User interface.
art and science seeking to make sense out of computer-generated
records. The process of creating such records is Ability to perform search on single, multiple files.
called datalogging. Logs are emitted by network devices, Support for Multiple Products: Logs from multiple
operating systems, applications and all manner of intelligent or products like W4N, ViPR, and NCM etc…can be searched
programmable device .A stream of messages in time-sequence within the framework.
often comprise a log. Logs may be directed to files and stored on Supports Multiple File Type
disk, or directed as a network stream to a log collector. Log Search Entire Folder.
messages must usually be Multiple Search Option.
interpreted with respect to the internal state of its source (e.g., Multiple Options to Display Errors.
application) and announce security-relevant or operations- Provide Highlighters for the Errors found.
relevant events (e.g., a user login, or a systems error). Adjustable Panels for readability.
Logs are often created by software developers to aid in IV. LITERATURE SURVEY
the debugging of the operation of application .The syntax and
semantics of data within log messages are usually application or Literature survey is mainly carried out in order to
vendor-specific. Terminology may also vary; for example, analyze the background of the current project which helps to
the authentication of a user to an application may be described as find out flaws in the existing system & guides on which
a login, a logon, and a user connection or authentication event. unsolved problems can be workout. Log Analyzer tools are
Hence, log analysis must interpret messages within the context available in the market most of them are web based application
of an application, vendor, system or configuration in order to with limited features.
make useful comparisons to messages from different log
sources. Log message format or content may not always be fully 1. What is a Search Log?
documented.
A search log is a file (i.e., log) of the communications
II. MOTIVATION (i.e., transactions) between a system and the users of that system.
Rice and Borgman (1983) present transaction logs as a data
Task of the log analyst is to induce the system to emit collection method that automatically captures the type, content,
the full range of messages in order to understand the complete or time of transactions made by a person from a terminal with
that system. Peters (1993) views transaction logs as
Once the server collects and records the data in a file, VI. PROBLEM STATEMENT
one must analyze this data in order to obtain beneficial Note Pad++ is not efficient in all ways.
information.
It only performs keyword search.
Few Tools are given below:- It does not perform other advance searches like
automatic grepping the exceptions from the log file.
PowerGREP:-PowerGREP is a powerful Windows grep Fails to a load a log file more than few MB’s length.
tool. Quickly search through large numbers of files on your PC Using this tool is not that effective. Log Analyst must
or network, including text and binary files, compressed archives, go through the log files line by line in order to fix a bug
MS Word documents, Excel spreadsheets, PDF files, Time consuming
OpenOffice files, etc. Find the information you want with
If a larger log file has to be searched it has to be first
powerful text patterns (regular expressions) specifying the form
split into many chunks, open each chunk and find the errors in it
of what you want, instead of literal text. Search and replace with
manually.
one or many regular expressions to comprehensively maintain
To split the files some other tool should be used first
web sites, source code, reports, etc. Extract statistics and
and then open each split in Note Pad++ every time.
knowledge from logs files and large data sets.
If there are many larger files as such the time
Weblog expert:-WebLog Expert is a fast and powerful
complexity increases.
access log analyzer. It will give you information about your site's
visitors: activity statistics, accessed files, paths through the site, If the issue is critical it has to be escalated within short
information about referring pages, search engines, browsers, amount of time ,if this is the case to split it and them find errors
operating systems, and more. The program produces easy-to- in each split it will affect the client’s environment who is waiting
read reports that include both text information (tables) and for the issue to get fixed.
charts. View the WebLog Expert sample report to get the general
idea of the variety of information about your site's usage it can VII. PROPOSED SYSTEM
provide. The proposed system has a lot of new features which would
help the log analysts perform the log analysis quickly and
Log Parser Lizard:- Log Parser Lizard is a GUI for
accurately. Following are the features of the Log Analyzer.
Microsoft Logparser, definitely the best one available on the
market today. Log Parser is a very powerful and versatile query Standalone Desktop Based Tool: The Tool will be a
software tool that provides universal query access (using SQL) standalone desktop based application which helps the users to
SEVERE –Red, WARNING-Green This project mainly consists of 3 Modules. They are: -
Swing Support. Swing was developed to provide a more 2. Initial Screen after tool as Launched
sophisticated set ofGUI components than the earlier Abstract
Window Toolkit. Swing provides a native look and feel that
emulates the look and feel of several platforms, and also
supports a pluggable look and feel that allows applications to
have a look and feel unrelated to the underlying platform. It
posses these traits Platform-independence, Extensibility, Look
and feel.
2. Development Environment
A platform is a crucial element in software
development. A platform might be simply defined as “a place
to launch software”. In this project, for implementation purpose
3. Simple Search
NetBeans -IDE 8.0.1 is used.
X. INTERPRETATION OF RESULTS
The following snapshots define the results or outputs
that we will get after step by step execution of all the modules of
the system.
5. Date and Time Range Search This project is supported by EMC Software & Services India
Pvt. Ltd. Bangalore. My Sincere thanks to the Escalation
Engineering Team Members for all their support and guidance to
carry out this work.
REFERENCES
End Time
XI. CONCLUSION
Towards Secure and Dependable for Reliable Data Fusion in Wireless Sensor
Networks under Byzantine Attacks
Valmeeki B.R. , Krishna Kumar. P.R., Shreemantha M.C.
Dept. of M.Tech (CSE) ,Cambridge Institute of Technology, B’lore-36
valmeeki1991@gmail.com,, rana.krishnakumar@citech.edu.in,, smchatrabana@gmail.com
Abstract - The Data Storage‘s attack is a severe attack that can data that is 1024 bits (128 bytes) and a state vector that is 512
be easily launched by a pair of external attackers in Wireless bits (64 bytes) in size, and it produces a modified state vector. It
Sensor Networks. In this attack, an attacker sniffs packets or is follow-on to the earlier hash algorithms MD5 and SHA-1, and
data at one point in the network by injecting fake contents or it is becoming increasingly important for secure internet traffic
wrong waiting time for corresponding nodes. In this system, and other authentication problems. As the SHA512 processing
the system proposes novel attackers detection and positioning involves a large amount of computations, it is critical that
scheme based on mobile (Location Based Server) LBS, which applications use the most efficient implementations available.
can not only detect the existence of Network Node attacks, but
also accurately localize the attackers for the system to The algorithm operates on 64-bit QWORDs, so the state is
eliminate them out of the network and enhancing the digital viewed as 8 QWORDs (commonly called A…H) and the input
signature value using Secure Hash Algorithm – 512(SHA-512) data is viewed as 16 QWORDs. The standard for the SHA-2
due to security reason. algorithm specifies a procedure for adding padding to the input
data to make it an integral number of blocks in length. This
Index terms – Wireless Sensor Networks, Location Based happens at a higher level than the code described in this
Server, Digital Signature, Secure Hash SHA-512. document. This paper is only concerned with updating the hash
state values for any integral number of blocks.
1. INTRODUCTION
The SHA512 algorithm is very similar to SHA256, and most of
Wireless Sensor Networks are spatially distributed autonomous the general optimization principles described in this system
sensor to monitor physical or environmental conditions such as apply here as well. The main differences in the algorithm
temperature, pressure and sound. In order to find the attacker specification are that SHA512 uses blocks, digests and data-type
details in the mobile phone, the mobile phone consists of three of computation twice the size of SHA256. In addition, SHA512
logical parts which are involved in the data exchange. The is specified with a larger number of rounds of processing (80
hardware component is the insecure communication unit of the rather than 64).
device responsible for the Bluetooth, Location Base Server
(LBS) or Mobile Device for communication with the external 2. IMPLEMENTATION
machine. The mobile user can connect with the LBS Server via
Bluetooth device to communicate with the mobile. The user will
find the Bluetooth server name and then login into mobile to
view all current attackers in the Storage Node which is Wireless
Sensor Networks.
Rapid growth in the Wireless technology and mobile devices to Entropy based anomaly detection scheme incorporates
deliver new types of location centric applications and services to knowledge and behavior for detecting the varying attacks in
users. Location base Service is also known as LBS, are a general wireless sensor network. Compared with existing mechanisms,
class of computer program-level services that use location data the entropy scheme achieves high filtering probability and high
to control features. As such LBS is an information system and reliability and also optimal utilization of energy. This work has
has a number of uses in social networking today as an been implemented in Java language, the results shows the
entertainment service, which is accessible with mobile devices effective data transmission in wireless sensor networks. Figure 2
through the mobile network and which uses information on the illustrates the total energy of all sensor nodes in the data
geographical position of the mobile device. This has become transmission, which also indicates the balance of energy
more and more important with the expansion of the smart phone consumption in the network and Figure 3 shows the comparison
and tablet markets as well. of time in the data transmission. The results demonstrate that the
Entropy based Anomaly Detection System scheme achieves high
Sender will first browse the file which it wants to send to the en-routing filtering probability and high reliability and also
destination, initially it redistribute the SHA512 standards. Then optimum utilization of energy.
the sender sends browsed file to the router before it delivers the
file to the destination. Once receiving the file from the sender
the router checks the details of end users and attacker details.
3. PERFORMANCE EVALUATION
Fig 3:
Comparisons of Energy Consumption in Data Transmission
4. CONCLUSION
Abstract—Security is one of the top concerns about cloud databases information is based on one master key shared by all
computing and the on demand business model. Worries over users
data privacy and financial exposure from data breaches may
be the cloud service providers greatest road backs to new The enforcement of access control policies through encryption
business. As the cloud infrastructure grows so does the schemes guarantees that data outsourced to the public cloud
presence of unsecured privileged identities that hold elevated databases are always managed in an encrypted way, thus
permission to access data, run program and change guaranteeing confidentiality for data in use and at rest in the
configuration settings. when the data is place in cloud ,the cloud. It minimizes information leakage in the case of user key
cloud provider should ensure the security and availability of loss or a compromised client machine and even in worst scenario
data. Encryption helps in securing the data but still is not where a malicious but legitimate user colluders with a cloud
complete. In this paper we propose an architecture that provider personnel by disclosing his decryption keys. In such a
implements identity and access management in encrypted case a partial data leakage is inevitable but is limited to the data
cloud databases. By enforcing access control and identity set accessible by the additional information about other data that
management, the users are guarantee in their security of data. remain inviolable through standard attack techniques.
this approach minimizes the data leakage problem. the
correctness and feasibility of the proposal is demonstrated Access control is only one subset of identity management
through formal models, while the integration in a cloud base (IM).identity
architecture is left to future work.
management covers a whole range of functions such as access
Index Terms-Cloud Security, Confidentiality, Identity and control ,user provisioning, directory services, account auditing,
Access Control Management role and group management, single sign-on(sso) and privileged
account management.
I.INTRODUCTION
Access control differs from identity management in that access
In a cloud context where confidential information is placed in control is strictly concerned with providing authentication
infrastructure of untrusted 3rd parties ensuring confidentiality credentials. In this approach the point is to provide user access,
and security of data is of main importance [2][5].In order to not prove their identity. This narrow focus according to identity
fulfill these requirements there are few data management management experts, leads to cases of mistaken identity. people
choices. The original data should be accessible only by the who shouldn’t have access to system like malicious users.
trusted parties. if the data is accessed by any untrusted party then masquerade as legitimate users to gain unauthorized access. In
the data needed to be encrypted .To satisfy all these this way identity management revolves around verifying users
requirements has different levels of complexity depending on the ideally with multiple pieces of proof of their identity before
type of cloud service. There are several solutions ensuring issuing of credentials.
confidentiality .confidentiality is a major concern and can be
ensured in several ways in storage as a service(sos).but in data II.LITERATURE SURVEY
base as a service paradigm ensuring confidentiality is still an
open research area. In this context secure DBaas is used, which Security in cloud is one of the major areas of research. The
does not expose the unencrypted data to the cloud provider and survey shows that the researchers are focusing on various
ensures the DBaas qualities, such as availability (readiness of techniques to enhance the data security in cloud.
data),efficiency of data(reliability)and elastic scalability[8].
Ryan K L Ko et.al [4]studied the problems and challenged of the
The confidentiality of data stored in cloud can be achieved trusted cloud, where the unauthorized user can access the entire
through encryption but must guarantee that all decryption keys data without disturbing the actual user. An unauthorized person
are managed by the tenant(client)/end-user and never by the may do two things which is accessing the data and putting
cloud provider. We cannot adopt the transparent data encryption duplicate data because cloud storage provides geographical
feature[7][1] because this approach h makes it possible to build a databases. It is not a trusted one to store the data of the users.
trusted DBMS over untrusted storage. The DBMS is trusted and
For this problem Ryan K L Ko et al proposed a Trust Cloud
decrypts data before their use. therefore this approach is not
framework ,to achieve a trusted cloud to the user, to provide a
applicable to the DBaas, because we consider the cloud provider
service by making use of detective controls in cloud
is untrusted.
environment. Detecting process has accountability access with
Even the proposal of the main authors in[8]has some risks of the cloud. Here user is a responsible person for their data, hence
information leakage because the encryption of the cloud user must tell the accountability with the cloud. Here user is a
*Assert inventory
*Risk assessment
*Architecture review
*implementation
Fig 1:reference model for a multi user accessing encrypted
cloud databases These steps should flow from your information security policy,
which company has already drafted.
First the inventory is divided based upon the risk example high
risk data and low risk data.
V. CONCLUSION
REFERENCES
i. G.Cattaneo,L.catuagno,A.Dsorbo,and P.Persiano,”The Design and
implementation of a Transparent Cryptographic File System For
Unix,”proc.FREENIX track:2001 USENIX Ann.Technical Cong.April.2001
ii. M.Armbrustetal,”A view of cloud computing,”comm..of the
ACM,vol.53,no.4,pp.50-58,2010.
iii. RyanKLKO,Peterjagadpramana,MirandaMowbray,siani Pearson,
Markus Kirchberg,QianhuiLiang,Bu Sung Lee,”Trust Cloud: A framework for
Accountability and trust in cloud computing 2011 IEEE.
iv. Muhammad RizwanAsghar,MihaelaIon,BrunoCrispo,”ESPOON
Enforcing Encrypted Security Policies in Outsourced Environment”,2011 Sixth
International conference on Availability Reliability and Security.
v. W.Jansen and T.Grance,”Guidelines on Security and Privacy in
Public Cloud Computing,” Technical Report Special Publication 800-
144,NIST,2011
vi. Luca Ferreti,MicheleColajanni and Micro Marchetti,Access Control
enforcement on query aware encrypted cloud databases IEEE 2013
vii. “OracleAdvancedsecurity”,Oracle corporation,
viii. http://www.Oracle.com/technetwork/database/options/advanced_secu
rity,April 2013.
ix. L.Ferretti,M.colajanni,and .Marchetti,”Distributed,Concurrent,and
independent Access to Encrypted cloud Databases.”IEEE Transaction on
Parallel and distributed system, 2014
ABSTRACT: A Mobile Ad-hoc Network (MANET) is a One of the typical features of MANET is each node must be able
dynamic wireless network that can be formed without the need to act as a router to find out the optimal path to forward a packet
for any pre-existing infrastructure. It is an autonomous system with low cost. As nodes may be moving continuously, entering
of mobile nodes connected by wireless links. Each node in a and leaving the network, the topology of the network will
MANET operates as a router to forward packets and also as an change automatically. For civilian and military applications
end system. The nodes are free to move into a network in self MANETs provide an emerging technology. One of the important
manner. These nodes often changes location. Proactive, research areas in MANETs is establishing and maintaining the
Reactive and Hybrid are the three main classes of routing ad hoc network through the use of routing protocols.
protocols. A Reactive (on-demand) routing strategy is a
popular routing category for wireless ad hoc routing .The Routing In MANET
design for as Reactive routing follows the idea that each node
tries to reduce routing overhead by sending routing packets Routing based on the straight flow of data from source
whenever a communication is requested. This survey compare to destination in order to maximize the network performance. It
the performance of two on demand reactive routing protocols has two fundamental requirements on the routing protocol such
for MANETs namely Ad hoc On Demand Distance Vector as (i) The protocol should be distributed and (ii) The protocol
(AODV), and Ad-hoc On-demand Multipath Distance Vector should able to compute multiple loop-free routes while keeping
Routing (AOMDV) .AODV are reactive gateway discovery the communication overhead to a minimum.
algorithms where a mobile device of MANET connects by Attacks In MANET
gateway only when it is needed. AOMDV was designed to solve Attacks in MANET can be categorized into Passive
problem in highly dynamic ad hoc networks where link failures attack and Active attack Passive attack This attack does not
and route breaks occur commonly. AOMDV maintains routes actually disrupt the operation of the operation of the network.
for destinations and uses sequence numbers to determine the Example: Snooping is unauthorized access to another person’s
freshness of routing information to prevent routing loops in data. Active attack This attack attempts to alter or destroy the
active communication. AOMDV is a timer-based protocol and data being exchanged in the network.
provides a way for mobile nodes to respond to link breaks and
topology changes. This survey states that Performance of Challenges in MANET
AOMDV is better than AODV by Packet Delivery Ratio, Life One of the main challenges in ad-hoc networking is the
Time of Network, Life Time of System and End-to-End Delay. efficient delivery of data packets to the mobile nodes. Here the
topology is not predetermined because the network does not
Key Words: - MANETS, AODV, DSR, AOMDV, MANET, have centralized control mechanism. Routing in ad-hoc networks
Routing. can be viewed as a challenge due to the frequently shifting
1.INTRODUCTION topology.
[8] The operation of the protocol has two phases: route B. HOP COUNT
discovery and route maintenance. In Ad-hoc routing, when a The advertised hop count should be maintained for each
route is needed to some destination, the protocol starts route destination which is defined as the maximum hop count for all
discovery. Then the source node sends route request (RREQ) the paths .This will be used for sending route advertisements for
message to its neighbors, if those nodes do not have any the destination .An alternate path to the target defines duplicate
information about the destination node, then they send the route advertisement received for each node. If alternate path has
message to all its neighbors and so on, if any neighbor node has a less hop count than the advertised hop count for that
the information about the destination node, the node sends route destination loop freedom is guaranteed for a node by accepting
reply message to the route request message initiator. alternate paths to destination. The advertised hop count therefore
REFERENCES
i. W. Heinemann, A. Chandrakasan, and H. Balakrishnan, “An
Application-Specific Protocol Architecture for Wireless Micro sensor
Networks,” IEEE Trans. Wireless Comm., vol. 1, no. 4, pp. 660-670, Oct.
2002.
ii. L.B. Oliveira et al., “SecLEACH-On the Security of Clustered
Sensor Networks,” Signal Processing, vol. 87, pp. 2882-2895, 2007.
iii. P. Banerjee, D. Jacobson, and S. Lahiri, “Security and
Performance Analysis of a Secure Clustering Protocol for Sensor
Networks,”Proc. IEEE Sixth Int’l Symp. Network Computing and Applications
(NCA), pp. 145-152, 2007.
iv. K. Zhang, C. Wang, and C. Wang, “A Secure Routing Protocol for
Cluster-Based Wireless Sensor Networks Using Group Key Management,”
5. CONCLUSION Proc. Fourth Int’l Conf. Wireless Comm., Networking and Mobile Computing
(WiCOM), pp. 1-5, 2008.
This paper evaluated the performance of AODV, v. Shamir, “Identity-Based Cryptosystems and Signature Schemes,”
Proc. Advances in Cryptology (CRYPTO), pp. 47-53,1985.
AOMDV and DSR using ns-2. Comparison was based on the
vi. J. Liu et al., “Efficient Online/Offline Identity-Based Signature for
packet delivery fraction, throughput and end-to-end delay. We Wireless Sensor Network,” Int’l J. Information Security, vol. 9, no. 4, pp. 287-
concluded that in the static network (pause time 50 sec), 296, 2010.
Light Weight SNMP Based Network Management and Control System for a
Homogeneous Network
Brunda Reddy H K, K Satyanarayan Reddy
Dept. of CSE(M.Tech), Cambridge Institute of Technology, B’lore-36
brundha1991@gmail.com, satyanarayanreddy.cse@citech.edu.in
Abstract—Network information helps in dissects faults and area, and there are a many number of interesting mechanisms
errors in a network. Remedying such faults and errors is a -such as ping, Trace route, DNS, address resolution protocol
major task of an organization network management system. (ARP), and SNMP- available to discover network elements
This paper introduces a mechanism that uses a Light Weight and the connectivity among them.
Simple Network Management Protocol (SNMP) - based R.Siamwalla et al. [ii] did a good work and proposed
solution that addresses discrete kinds of network devices and mechanisms to discover topology by combining ping, trace
discovers Interface-to-Interface connectivity among the route, SNMP, DNS, and ARP. These methods can able to
devices and basic information of those network devices This discover only L3-level devices, and thispaper did not propose
paper proposes a algorithm to discover network, discover any method to discover L2 – or host-level devices, though
device type, and Interface-to-Interface connectivity This paper they proved that SNMP performs better than all other
concentrates on a subnet of an organizations network. mechanisms. Yuri et al. [iii] proposed a mechanism that is
heterogeneousbut this mechanism requires ICMP spoofing in
Keywords—MIB, OID, SNMP, Topology, Subnetwork. order to get complete forwarding table, which is not allowed
in most of today’s networks. Though they did a good work in
explaining the connectivity algorithm they failed to provide
I. I NT RO DU CT I O N details on SNMP MIBs required for collecting network
Network topology is an illustration of nodes and links in a topology information. Lowekamp et al. [iv] proposed a
network and nodes are interconnected with each other. The mechanism by which they would not require complete
network topology can be classified as a physical network forwarding information of bridges; their approach
topology, which is referred as enacting of the physical contradicted of Yuri et al. [iii]. SumanPandey et al. [i]
connectivity relationships that exist among entities or nodes in a extended the work of Lowekamp et al. [iv] and proposes a
network. A physical network corresponds to many logical complete topological discovery mechanism to discover L2-
topologies, in which a network is divided into logical segments L2, L3-L3, L2-L3, and L2 and L3 to end host connectivity.
through subnets. This paper extend the work of SumanPandey et al. [i] and
An organization consists of many departments and an discovers details of each network devices discovered in the
organization’s level network consists of many subnetworks. organisations subnetwork that is supported by the SNMP and
Network topologies can constantly changes as nodes and links for those devices that does not support SNMP, this paper
join a network and network capacity is increased to deal with make use of ICMP echo request to check the device is alive or
added traffic. Keep track of network topology manually is a not and displayed ping information and some basic
frustrating and often impossible job. An inexperienced network information of the device.
administrator joining an organization faces many problems due Organization of this paper is as follows: The network
to the lack of a discovery tool. Even for the experienced person, topology discovery algorithms are explained in section 2, and
keeping track of devices and their connectivity details, without implementation is explained in section 3, and our conclusion
having a proper method of visually presenting them becomes a and future works are explained in section 4.
difficult task. In order to avoid these problems accurate
topology information is necessary for simulation, Network II. NETWORK DISCOVERY ALGORITHM
management and so on. In this section, discovering network nodes and the
Thus, there is a considerable need for automatic discovery of connections between them and details of each discovered
network topology. This paper proposes a Light Weight SNMP devices are explained. Since the approach of this paper is
based solution. The solution using SNMP is simple and mainly depending on the SNMP, it first analyzes the
effective and it is easy to use because even if the host or any Management Information Base (MIB) objects required to
device that does not support SNMP, can still find the discover the network and the devices in the network. Then
connections. This paper generates a solution that performs those MIB’s are used to discover the network, type of device,
better than other systems and generates least amount of traffic details of particular device and connectivity between switch
and network bandwidth will be less. This paper concentrates on and network devices.
subnetwork of the organization-level topological discovery.
A. MIBs used
Related work: Discovering the topology of the Internet is a Discovery mechanism used in this paper is completely based
problem that has attracted the attention of many networking re- on SNMP. Table 1 explains all the SNMP MIB Objects
searchers. Network connectivity discovery is a well known needed.
IV. CONCLUSION AND FUTURE WORK. i. SumanPandey, Mi-Jung Choi, Sung- Joo Lee, James W. Hong “IP
Network Topology Discovery Using SNMP”, POSTECH, Korea 2013.
In this paper, we focus on discovering the devices of ii. R.Siamwalla, R. Sharma, and S. Keshav, “Discovering internet
subnetwork of an organization we also discover the connectivity topology, ” Cornell Univ., Ithaca, NY, Techical Report.
between switch and devices and some details of those devices. iii. Y. Breitbart, M. Garofalakis, B. Jai, C. Martin, R. Rastogi, A.
We discovered different type of devices, including switches, Silberschatz, “Topology Discovery in Heterogeneous IP Networks: The
NetInventory System,” IEEE/ACM Transactions on Networking.
printers and end host and enhanced the already existing iv. B. Lowekamp, D. R. O’Hallaron, T. R. Gross, “Topology
technique of device type discovery. We utilized the SNMP discovery for large Ethernet networks,” ACM SIGCOMM, San Diego, CA,
mechanism, which is the most efficient mechanism and this USA, pp. 237~248.
generates the least amount of traffic in comparison to v. JFreeChart implementations, http://www.jfree.org.
vi. R. Smith, F. Wright, S. Zilles, J Gyllenskog, “Management
mechanisms in other research. Information Base for Printer” RFC 1759, IETF, March 1995
Since our discovery system is applied to a subnetwork of an vii. P. Grillo, S. Waldbusser, “Host Resources MIB”, RFC 1514,
organization, our future goal is to discover the entire September 1993.
organization’s network. For visualizing a network we aim to viii. SNMP, SNMP4J API, http://www.smp4j.org
ix. K. McCloghrie, M. Rose, “Management information Base for
represent a network in a graphical form and include more link Network Management of TCP/IP- based Internets, MIB-II,” RFC 1213,
characteristics such as link capacity and link failure on the IETF, March 1991
graphical representation. To notify the SNMP manager (client) x. E.Decker, P. Langille, A.Rijsinghani, K .McCloghrie, “Bridge
about the problem at SNMP agent (server) like disk crash at MIB,” RFC 1493, July 1993.
systems and so on we plan to use SNMP traps in the future.
Abstract— Digital image processing has become so popular pattern gives special importance to the number of green
over past few decades but increasing level of noise will effect sensors to mimic the human eyes greater sensitive to green
the quality of the image. Noise has to be removed to improve light .The demosaic method based on interpolation to
the image quality. The bayer color filter array (CFA) gives convert two –dimensional- bayer encoded image into the
information about the intensity of light in red, green and true color image, RGB, which is an M-by-N-by-3 array.
blue(RGB) wavelength regions. The CFA image captured by
the image sensor is then demosaick to get full color (RGB) Sensor alignment is one of the following text
image. The present work represent a novel color image strings that specifies the bayer pattern. Each string represents
demosaicking algorithm using a lagrange quadratic the order of the red, green and blue sensors by describing the
interpolation method and directional interpolation method. By four pixels in the upper left corner to the image (left-to-right,
introducing the lagarnge interpolation the interpolation top-to-bottom).
direction of the center missing color component can be
determined with minimum error. Also the center missing color
component is interpolated using the quadratic interpolation G1A R2A G3A R4A
method by exploring the intra channel correlation of the
neighboring pixels. In addition to this the present work B5A G6A B7A G8A
contributes in strengthening the image quality and provides
superior performance in both objectively and subjectively.
G9A R 10A G11A R 12A
Keywords—. Color filter array(CFA) interpolation,
demosaicking, lagrange quadratic interpolation. B13A G14A B15A G16A
I. INTRODUCTION
Fig. 1. Bayer CFA pattern.
Human eyes can perceive a few million colors. Most of these
colors can be produced by mixing just the three primary colors The existing methods were proposed to obtain a full
− red, green and blue − in varying proportions . Image sensor color image by utilizing the color differences between RGB
are used to acquire primary colors. Three separate sensors are planes. Each method has its own both the advantage and a
required for a camera to acquire an image. To reduce the cost disadvantages with respect to the interpolation .In a
and space many cameras are using a single sensor covered with demosaicking technique there are lot of challenges to
a color filter array (CFA). In the CFA-based sensor achieve interpolation with efficient and effective manner to
configuration, commonly 2×2 bayer patterns are used to acquire obtain full 24 bit color image with less degradation.
an color image as shown in fig 1. Color image contains three
RGB planes. The CFA image contains few color pixels of each In this paper proposed interpolation technique is
plane and remaining pixels are missing. Those missing color simplest approach to the demosaicking problem to treat color
components are estimated by considering the existence of planes seperately and fill in missing pixels in each plane
acquired neighboring pixels contained in the CFA image. This using a lagrange based quadratic interpolation. Advantage
process is called interpolation. The interpolation is applied to of this method is more effective in smooth region. However,
each and every missing pixel to obtain a full color image. The existing methods leads to color artifacts and lower resolution
color interpolation process is known as demosaicikng. Although in regions with texture and edge structures. To overcome
many different CFA patterns have been proposed in the camera, from these issues, proposed method reduces color artifacts
the most prevalent is the 2×2 ‘GRBG’ Bayer patter shown in and giving good resolution in the edges. Here introducing an
fig [1]. The color reproduction quality depends on the CFA interpolation in both horizontal and vertical directions and
templates and the demosaicking algorithms that are employed. seperately treated for all the three RGB planes
There are various demosaicking algorithms [1] - [8] have been independently.
proposed in the past few decades based on Bayer pattern.
The rest of the paper is organized as follows. Section II
A bayer filter array or CFA represent the arrangement describes the proposed Lagrange based Quadratic
of color filters that each and every sensor in a single sensor interpolation algorithm. Section III presents experimental
digital camera only acquired red, green and blue pixels. The results, and Section IV the conclusions.
The motivation of this proposed method comes by In the vertical interpolation only two pixels in the coloumn
observing the traditional demosaicking methods [1] - [8]. Due are interpolated because the pattern G6A is already acquired.
to the inaccurate edge information the center missing color The missing pixel G5I in the first column can be
component cannot be interpolated accurately because there is a obtained using equation (3) and the third column missing
inadequate information of irregular edge and texture details pixel G7I using equation (4).
exists. Here the edge directions of the neighboring pixels are
estimated in order to exploit the main direction by lagrange First column,
quadratic interpolation with localy.
Horizontal Interpolation
The missing pixel G2I in the first row can be obtained
using equation 1 and the third row missing pixel G10I using
equation (2). Fig.3. Demosaic Red plane
First row,
Third row,
Vertical Interpolation
Horizontal Interpolation
Second row,
Third row,
Vertical Interpolation
First coloumn,
Second coloumn,
Fig.10.Demosaick image
References
i. Xiangdong Chen,Gwanggil Jeon,Jechange Jeong”voting-based
directional interpolation method and its application to still color image
demosaicking”,vol.24.no.2,February 2014.
ii. Pekkucuksen and Y.Altunbasak,”Edge strength filter based color
filter array interpolation” IEEE Trans. Image Process..vol.21.no.1,pp.393-
397,Jan.2012.
iii. K.H. Chung and Y.H.ChaTrans.Image
Fig8.original image Process.,vol.15,no.10,pp.2944-2945,0ct,2006.
iv. N.X.lian,L.chang,Y.P.Tan,and V. Zagorodnov.”Adaptive filtering
CFA image
for color filter array demosaicking,”IEEE Trans. Image
Process.,vol.16,no.10,pp.2515-2525,oct.2007.
v. R.Lukac,K.N.Plataniotis,and D.Hatzinakos,”color image
zooming on the bayer patterns,”IEEE Trans,Circuits Syst.Video
Technol.,vol,15,no.11.pp,1457-1492,Nov,2005.
vi. D. Paliy, V. Katkovnik, R. Bilcu, S. Alenius, and K. Egiazarian,
“Spa-tially adaptive color filter array interpolation for noiseless and noisy
data,” Int. J. Imag. Syst. Technol., vol. 17, no. 3, pp. 105–122, 2007.
vii. L.Zhang,A. Wu,A.Buades,and X.Li.”color demosaicking by
local directional interpolation and non-localadaptive
thresholding,”J.Electron.Imaging,Vol.20,no.2,p,023016-16,2011.
viii. A.Buades,B.Coll,J.M.Morel, and C.Sbert,”self-similarity driven
color demosaicking,” IEEE Trans.Image Process.,vol.18,no.6,pp.1192-
1202,jun.2009.
ix. P.Simoncelli.”Image quality assessment:from error visibility to
structural similarity.” IEEETrans.Image Process.,vol.13,no.4.pp.600-
612.Apr.2004.
x. R. Lukac and K. N. Plataniotis, “A normalized model for color-
ratio based demosaicking schemes,” in Int. Conf. on Image Process., 2004,
Fig9. CFA image vol. 3, pp. 1657–1660.
Abstract- In order to weave the exponentially increasing billions of records is split into many small tasks each of which is
quantity of biometric data, it must be dealt with a big data running on one or multiple computing nodes [11][13].
perspective using technologies capable of processing massive One solution with respect to the outlined issues is moving
amounts of data efficiently and securely. The main challenge existing biometric technology to the big data platform that
in the biometric industry is to overcome all the threats during ensures appropriate scalability of the technology, sufficient
different phases of the biometric system development life cycle. amount of storage, parallel processing capabilities, new types of
The current biometric models emphasis’s the importance and tools to analyze the data and with the wide spread availability of
significance of big data. This paper capitalizes on the most mobile devices also provides an accessible entry point for
important challenges encountered and critical criteria’s to be various applications and services that rely on mobile clients.
followed in biometric analysis and proposes a general Hence big data biometrics analysis is capable of addressing
approach for the big data biometric analysis. issues related to the next generation of biometric technology, but
at the same time offers new analytical tools possible to be used
I. INTRODUCTION along with the existing biometric systems.
Most people in the internet authenticationuses passwords. The However moving the existing biometric technology to the
Biggest threat with password authentication approaches is the big data environment is a nontrivial task. Biometric architects,
existence of too many password account pairings for each user developers and researchers who attempt to tackle this should be
which leads to forgetting or the same user name and password aware of challenges encountered with big data [10][12].
for multiple sites [1]. One possible solution to this problem can The organization of the paper is structured as follows. In
be the use of biometric systems [2][6][14]. Biometric section 2 we concentrate on the challenges, considerations and
authentication techniques try to validate the identity of a user trends for big data in the field of biometrics. Section 3
based on his/her physiological or the behavioral traits, while concentrates on working strategies like operating territory and
their use on the internet is still relatively modest. The main the focused areas for big data biometric analysis. In section 4 we
reason is accessibility and scalability of existing biometric propose the general approach for the big data biometric analysis.
technology. Finally the paper is concluded with some comments.
Similar issues are also encountered in other deployment domains
of biometric technology such as forensics, law enforcement and II. RELATED WORK
alike. For example according to [3] the biometric databases of
the Federal Bureau of Investigation, the US State Department, Big data biometric analysis is a highly active field, which gained
Department of Defense, the development of Homeland security, popularity only a few years ago. Since the field covers a wide
and Aadhaar project in India are expected to grow significantly range of areas relating to all phases of big data analysis in
over the next few years to accommodate several hundred million biometrics, it is natural that not all possible aspects of the field is
(or even billions) of identities. Such expectations make it appropriately covered in available scientific literature. This is
necessary to devise highly scalable biometric technology, also true for big data biometric analysis [7][8].
capable of operating on enormous amount data which in turn This paper tries to cover challenges faced in the big data
induces the need for sufficient storage capacity and significant environment because big data gives many insights to the
processing power. analysis. Mean while we need to make many considerations and
mark the operating territories in the field of biometrics.
A. Big Data Mining platform
A. Challenges for Big data biometric analysis
In data mining systems, the mining algorithms require Due to improvement in the field of electronic devices and
computational intensive computing units for data analysis and multiple data collection sources produces enormous biometric
comparisons. Computing platform needs two types of resources: data. Processing of this volume, variety, value and velocity data
data and computing processors. For small scale data mining has the following challenges in the field of biometrics.
tasks a single desktop, which contains hard Disk and CPU is Handling large data: Almost every electronic sensing
sufficient. Indeed many data mining algorithms are designed for device generates some kind of digital data. However most of the
this type of problems [5][9]. data is not being used due to challenges such as storage, analysis
Big data mining will rely on cluster computers with a high and closed nature of existing biometric systems. And it also
performance computing platform. A data mining task is applies to the biometric devices and biometric data. Handling
deployed by running parallel programming tools. The role of the such huge biometric data is one real challenge and requires new
software component is to make sure that a single data mining types of storage, analytical skills and open systems.
task, such as finding best match of a query from a database with
V CONCLUSION
Leveraging existing biometric to Big data have an enormous
potential market value and as such attract the interest of research
and development groups from all around the world.This paper
highlights Challenges, considerations, trends, operating
territories and focused areas that need to be considered when
designing big data biometric. A general approach for big data
biometric processing systems is designed as an analysis stack.
BIBLIOGRAPHY
ABSTRACT:Location based services is a part of mobile Location sharing application providesan personal location
multimedia services, where user can find the services and sharing service with known person based on user permission.
products.The services will support people to navigate on daily Location sharing service mainly concerns on user safety service
errands.There arenumerous application extents like mobile by directlysending user location to emergency stations like
works, shopping and sports, tourism delivery, community police stations, fire stations and hospitals. Based on user location
servicesand public transport and safety. Mobile location based emergency service providers will provide the safety service to
services will be based on standard technology like mobile particular user.It also provides guidelines on user required route
devices, wireless networks and maps. Particularly and transport all this services are based on the application called
mobilelocation based services utilize currentposition Google nearby search.
capabilities of a mobile device using GPS technology it extract
the position of the user. Based on user location nearby services
will be determined. One of the major issues in location based
service is privacy controls to users without vastly affecting
user’sservices. Main aspects of this application are providing
service, sharing and safety for users.
I. INTRODUCTION
C. Security in application
Major aspect in mobile applications is maintaining privacy in
user data from unknown authorized. Based on userauthentication
informationuser get response back from application. In mobile
application primary authentication is based on user name and
password.
In these application we are getting information through web
service to maintain URL privacy we are passing secrete key with
base URL .This provides safe information exchange between
application and database through web service.
Following mathematical statements describes the security
privacy in web service.
Universal resource link access security contains following
procedure;
URL=I(info)+key(ss)
Where
Key (ss) =>Server side secrete key,
I (info) =>Respective user information whichincludes
authenticationinformation plus service information.
IV. CONCLUSION
This section briefly presents the result works ofmobile location
based fast services application. This application provides
commerce services to userswith fast delivery services. Main
advantage of mobile location based service is service providers
are filtered based on the respective user location using system
user longitude and latitude coordinates of the mobile device user
locationwill be determined so delivery in service to users will be
comparably fast.Ithelpsin technology support for medium and
small scale business peoples by providing mobile interface to
Services providers to add their items into the application and
alsoit provides web user interface for service providers.
REFERENCES
i. Chandra, A., Jain, S., Qadeer, M.A., “GPS Locator: An Application
for Location Tracking and Sharing Using GPS for Java Enabled Handhelds,”
2011 International Conference on Computational Intelligence and
Communication Networks (CICN), pp.406-410, 7-9 Oct.2011.
ii. http://www.olx.in o
iii. http://www.olx.in
iv. Sandeep Kumar, Mohammed Abdul Qadeer, Archana
Gupta,―Location Based Services using Androidǁ , IEEE 2009.
v. Daniel J. Abadi, Peter A. Boncz, Stavros Harizopoulos, Column-
oriented Database Systems, VLDB ’09, August 24-28, 2009, Lyon, France.
vi. http://en.wikipedia.org/wiki/find
vii. Miguel C. Ferreira, Samuel R. Madden, Compression and Query
Execution within Column Oriented Databases.
viii. Daniel J. Abadi, Query Execution in Column-Oriented Database
Systems,
ix. D. J. Abadi, S. R. Madden, N. Hachem, Column-stores vs. row-stores:
how different are they really?, in: SIGMOD’08, 2008, pp.967–980.
Abstract: The use of mobile phones today has become a smartphone operating system it also supports execution of all
part of our daily life.Recently all the mobile phones or smart local and third party applications.There exist many open source
phones are equipped with Global positioning system (GPS) mobile platforms, butiOSfromapple, androidfrom Google,
sensors to get information about the location. LBS (Location Symbian from Symbian foundation, and windows from
based services) are used to obtain the knowledge about the Microsoft are the most popular. Android provides a platform
geographical position. There exist many applications today where any applications can be downloaded, according to
which are going to share ones location with other in the terms research has done till now more than 68,000 applications are
of location co-ordinates (longitude and latitude)that can be available and number of applications downloaded by android
viewed in the form of Google map also called as map based enabled mobile phone reached more than 1 billion.
location sharing. This paper provides a detail description of Recent year’s android enabled mobile phones going to grab
sharing location with the friends in the form of text (also called the special attention than others that is because of its features,
as text based location sharing) instead in terms of map, since there are many features that android have but one of the
map based location sharing is time consuming compared to interesting feature enabled by android is GPS. GPS is the one
text based location sharing. The Longitude and latitude (Geo- which helps in getting driving directions and proving location
graphical coordinates) properties are used to obtain the informationthe main purpose of using GPS is to provide location
location and that can be converted into text form and shared based information also called as location based service [3].
with the friends. This application is also enriched with the Location based services provide the location of person/device
Near-by services. It provides all the services near to the user and the same location can be shared with others and other
location. Here near-by services are the organizations those technology with respect to location based services is tracking
who provide the services to the user nearer to the organization the location of the other person/device this is also called as “self-
location. reporting position” instead of tracking.
The location of person/device can be obtained by location
Keywords:GPS (global positioning system),Location co-ordinates (longitude and latitude). The GPS sensors inserted
based services (LBS),longitude and latitude. inside the device sense the accurate location and obtain the
longitude and latitude of the position and display the location in
terms of Google map.
I.INTRODUCTION
Today all smartphones have the location sensing capability for the applications which has been downloaded from the
built-in, the successful location based service can be obtained by Google because android is an open source platform for mobile
providing the accurate location co-ordinates. operating system. Anyone can upload the application in to the
Google so few may take advantage and upload application
without any security issues which leads to many computer
II.LITERATURE SURVEY crimes; such problem can be overcome by using the layered
approach to develop android application. This contains the
This section briefly presents the related works on location application sandbox to detect the suspicious application both
based services and its applications. statically and dynamically.
Today, we are dealing with the era of smart phones and The application named Nearest Friends Notification
iPhone’s, which are going to replace the bulky desktops in all Reminder [8] this application feature is to provide notification
manners. We have huge number of applications and usage where when any of friends in the user friend list moved in to the same
person walking on the roadside needs to get relevant data that location. GPS tracker will track the location of the friend only
can be obtained by location based services. GPS is a local when the friends get into the same location as the user. The
positing system becoming popular. It is easier these days to use advantage of using this application is which helps to meet a
the map by connecting to the GPS receiver to devices. GPS friend who is in same area/place.
chips are inserted into the device which obtains the accurate Google has built-in feature Search nearby [9]to search the
location of user by satellite signals and the location can be nearby locations. Which helps the person to find the nearby
viewed in terms of Google map. location together with the location it also provides the option of
The authors Chris Hulls and Alex Haro proposed an navigation and bookmark. Where navigate option provides the
application Life 360:usability of mobile devices for time use direction to location by showing the route and bookmark allows
surveys [4]in 2008. This is a family network location based the place to be marked as interest thereby saving the location and
service application which allows the family members to share retrieving the directions for them.
their location and easily communicate with each other by adding
the family members into the application which makes a family III.PROPOSED SYSTEM
circle. The main features of this application is the person can
Location based services provides the location of the device
instantly see where other family members are located, person
using GPS, location can be obtained by the geographical
can also able to share or not to share their particular location at
coordinates (longitude and latitude). There exist many
particular time and family members can also chat with each
applications today to share the location of person/device, but the
other within a circle, thereby providing the family safety and
purpose of the proposed system is to share the location with
also it gives alert when a new person enter into the circle or
friends in the text form and it also search the nearby services and
when the family member leave the circle.
display the contact of the registered nearby service providers,
Another noteworthy application related to the location where user can able to interact with the registered service
sharing is Find My Friends [5] by Apple in 2011. This providers and make their work done.
application allows the user to track other person location and can
share the location of his own with person of his choice, if a
A. Requirements
person wants to track the location of the other person then the
notification can be send to the person as a request. The location This work is designed for both users as well as for the service
can be turned on or off at any time, location of the person is providers. Both must have the smartphones that support GPS
obtained by GPS, so whenever the GPS is turned off it is and before they use the application both must be registered to the
difficult to track location and to share location. application. This application is used by the users to view the
In the paper GPS and SMS based child tracking system friends location and to view the nearby service providers, where
using smart phone[6] by A. Al-Mazloum, E.omer and M. F. A. service providers can only interact with the user who wish to use
Abdullah presented anapplication based on tracking the child
the services that has been provided by near-by service provider.
using the smart phone. This application specially developed to
provide children safety by tracking the location of the child This application is developed using androidStudio, database
using the smart phone, where once after the application is MySQL, web service using PHP. Android OS has been used for
installed by child and parent the parent can able to track the the implementation the solid reason for why we use this for
child activities. The parent can obtain the child location by implementation is to target more number of users.
sending the request message to the child to obtain the location
and child can response the request of the parent thereby parent B. Application Architecture
can view the location of the child. Here GPS is used to obtain We propose a solution to share the location and to access the
the location of the child and location can viewed in terms of nearby services using the GPS technology, GPS feature exist in
Google map their by providing the child safety. all advanced smartphone. The simple idea of this application is
In the paper titled Android based mobile application and to share and track the location of the person, and to search Near-
development and its security [7] by SuhasHolla and Mahima M by services.
Katti provided a detailed description of how to achieve security
IV.CONCLUSION
Nowadays there exist many applications based on the
location based services. This application provides automatic
updating of location for every 30 seconds and sharing our
location address in text form with the friends. Where the basic
applications used to share location in the map view requires high
speed network. But in the proposed system sharing a location in
text form can be achieved through the low speed network also.
The system uses longitude and latitude properties to share a
location. The additional feature of this application is which
provides user with nearby services option, which shows all
nearby services exist near to the user location. So user can make
use of nearby service by interacting with the service of their
interest.It provides a fully secured location sharing based on
authorization, and privacy can be achieved by providing the data
only to the subscribed user.
REFERENCES
i. Chandra, A., Jain, S., Qadeer, M.A., “GPS Locator: An Application
for Location Tracking and Sharing Using GPS for Java Enabled Handhelds,”
2011 International Conference on Computational Intelligence and
Communication Networks (CICN), pp.406-410, 7-9 Oct.2011.
ii. http://en.wikipedia.org/wiki/android
iii. Sandeep Kumar, Mohammed Abdul Qadeer, Archana
Gupta,―Location Based Services using Androidǁ , IEEE 2009.
iv. Jennie W. Lai, LorelleVanno, Michael W. Link, PhD,Jennie Pearson,
HalaMakowska, Karen Benezra, and Mark Green “Life 360: usability of
mobile devices for time use surveys”AAPOR – May 14-17, 2009
v. http://en.wikipedia.org/wiki/Find_My_Friends
vi. A. Al-Mazloum, E. Omer, M. F. A. Abdullah,”GPS and SMS-Based
Child Tracking System Using Smart Phone”,International Journal of
Electrical, Computer, Electronics and Communication Engineering Vol:7,
No:2, 2013
vii. SuhasHolla, Mahima M Katti, “Android based mobile application
and development and its security” IEEE-2012.
viii. Http://blogs.wsj.com/digits/.../facebook-to-notify-users-when-friends-
are-nearby/.
ix. http://en.wikipedia.org/wiki/Nearby
ABSTRACT confidential file that are stored in the cloud. After the user
specified time expire the files and all the replicas will be self-
Cloud is meant for storing large amount of data
destructed from the cloud.
for long period of time with security. The user may
some of his confidential data in the cloud. To
Every confidential file stored in the cloud may
maintain the good consistency the cloud service
not be required for the long period of time by the user. To
provider replicates the data geographically without the
delete the file after some time vanish methodology was
permission of authorized user. As the data is confidential
proposed. In vanish methodology the secret key will
and the data is replicated and stored in different servers
be divided and stored in the distributed hash table (DHT),
the data can be misused and some malicious activity can
DHT is one of the characteristic of P2P, the node in the
be performed by the unauthorized user or by the cloud
DHT will be refreshed after every 8 hours. So the keys
service provider. In order to overcome the above conflicts
present in the node will be deleted, because of this the user
the SeDaS is proposed. Self- destructing is mainly used for
may not get the enough number of keys to decrypt the
protecting the confidential data. As the data is
file.
confidential to protect the data the user specifies the time
interval for that specific data that is stored in cloud. After
One of the disadvantage of vanish methodology is
the completion of time interval the data and the
the key cannot survive for long period of time. To overcome
replicated copies will be self- destructed without
this challenge SeDaS is proposed which is dependent on
intimating the authorized user. This paper is using active
Active Storage Framework. The SeDaS system mainly
storage and cryptographic techniques to solve above
stipulate two modules, one is self- destruct method object that
challenges.
is fraternized with each and every secret key and another is
for each secret key the survival time parameter.
Keywords: Cloud computing, self-destruction data,
active storage framework, data privacy.
SeDaS is offering:
I. INTRODUCTION
1) Key distribution algorithm is based on Shamir’s
Cloud computing plays a major role for organization algorithm which is mainly used as core algorithm
or the individuals for storing the large amount of data. to store the clients’ distributed key in the object
Cloud provides not only the storage but also provides the storage system.
services like infrastructure-as-a-service (Iaas), platform-as-a- 2) Here the object based storage interface is used to
service (Paas), software-as-a-service (Saas). Because of these store and manage the divided keys that are based on
services the organization and individuals are focusing Active Storage Framework.
towards cloud. 3) Securely deleting files and random encryption keys
that is stored in secondary storage is supported
As cloud is an internet based technology it also by SeDaS.
provides mobility so people are more interested in storing
and retrieving the personal data. The personal data may
contain passwords, passport numbers, account numbers II. RELATED WORK
and some more important documents. In spite of maintaining Levy et al. (2009) proposed “Vanish: Increasing Data
the individual files all the files can be stored in a single Privacy with Self- Destructing Data” [2]
directory in cloud. The user specifies the time for each
Personal data are cached, copied, and This mechanism is not universally applicable to
archived by third parties, often without our knowledge or all users or data types. They focus in particular on sensitive
control. We wish to ensure that all copies of certain data data that a user would prefer to see destroyed early rather
become unreadable after a user-specified time, without any than fall into the wrong hands. Vanish applications may
specific action on the part of a user, and even if an attacker compose VDOs with traditional encryption systems like PGP
obtains both a cached copy of that data and the user’s and GPG. In this case, the user will naturally need to
cryptographic keys and passwords. manipulate the PGP/GPG keys and passphrases. It does not
With the help of novel integration of cryptographic defend against denial of service
techniques, Vanish overcomes the above challenges. The goal attacks that could prevent reading of the data during its
is to self-destruct the data automatically after it is no longer lifetime.
useful. Vanish system leverage the services provided by
Tang et al. (2010) proposed “FADE: A secure overlay
decentralized, global-scale P2P infrastructures and, in
cloud storage system with File Assured Deletion” [3]
particular, Distributed Hash Tables (DHTs). DHTs are
Keeping data permanently is undesirable, as data
designed to implement a robust index-value database on a
may be unexpectedly disclosed in the future due to
collection of P2P nodes. Vanish encrypts a user’s data locally
malicious attacks on the cloud or careless management of
with a random encryption key not known to the user,
cloud operators. The challenge of achieving assured deletion
destroys the local copy of the key, and then sprinkles bits
is that we have to trust cloud storage providers to actually
(Shamir secret shares) of the key across random indices
delete data, but they may be reluctant in doing so. Also,
(thus random nodes) in the DHT.
cloud storage providers typically keep multiple backup copies
Vanish architecture of data for fault-tolerance reasons. It is uncertain,
from cloud clients’ perspectives, whether cloud providers
Data object D is taken by Vanish in order to reliably remove all backup copies upon requests of deletion.
encapsulate it into VDO. To encapsulate the data D, Vanish FADE is a secure overlay cloud storage system that
picks a random data key, K, and encrypts D with K to obtain provides fine-grained access control and assured deletion for
a cipher text C. outsourced data on the cloud, while working seamlessly atop
today’s cloud storage services. In FADE, active data files
that remain on the cloud are associated with a set of user-
defined file access policies (e.g., time expiration, read/write
permissions of authorized users), such that data files are
accessible only to users who satisfy the file access policies.
In addition, FADE generalizes time-based file assured
Figure 1: Vanish System Architecture Figure shows deletion (i.e., data files are assuredly deleted upon time
how to split the data key K into N pieces K1… KN expiration) into a more fine-grained approach called
vanish uses threshold secret sharing. The application or policy- based file assured deletion, in which data files are
the user can set the threshold, which is the parameter assuredly deleted when the associated file access policies
of secret sharing. To reconstruct the original key, N are revoked and become obsolete.
shares are required which is determined by threshold.
Advantages The FADE system
Vanish targets post-facto, retroactive attacks; The FADE system is composed of two main entities:
that is, it defends the user against future attacks on old, • FADE clients. A FADE client (or client for short) is an
forgotten, or unreachable copies of data. The attacker’s job interface that bridges the data source (e.g., file system) and
is very difficult, since he must develop an infrastructure the cloud. It applies encryption (decryption) to the outsourced
capable of attacking all users at all times. Solution utilizes data files uploaded to (downloaded from) the cloud. It also
existing, popular, researched technology - used since 2001. interacts with the key managers to perform the necessary
It does not require special security hardware, or special cryptographic key operations.
operations on the part of the user. Utilizes inherent half-life • Key managers. Minimum group of key managers together
(churn) of nodes in DHT – data definitely destroyed. developed FADE for assured deletion and access control
Disadvantages
Abstract: MANET is an infrastructure less sensor network, Table 1.0 Characteristics of MANETS’s
consisting of collection of mobile devices. Secure Routing is a
s.no Characteristics Description
challenging task in MANETS to overcome this problem 1 Distributed network The control of the network is
anonymous routing protocol is developed. This paper focus on distributed among the nodes i.e. there
comparison of different existing anonymous routing protocol is no background network for the
based on routing category, design, advantages and central control of the network
operations.
disadvantages.
Conclusion
References
i. Dr.S.S.Dhenakaran, A.Parvathavarthini, “An Overview of Routing
Protocols in Mobile Ad-Hoc Network”, “International Journal of Advanced
Research inComputer Science and Software Engineering”, Volume 3, Issue 2,
February 2013
Abstract: Stock market is a widely used investment scheme forecast should be made before the pattern is completed to
promising high returns but it has some risks. An intelligent facilitate the prediction process. The vital idea for successful
stock prediction model would be necessary. Stock market stock market prediction is achieving best results, and to
prediction is a technique to forecast the future value of the minimize the inaccurate forecast of the prices.
stock markets based on the current as well as historical data
available in the market. Stock market prediction is a mainly
based on Technical Analysis and Fundamental Analysis. In
literature it is observed that there are several techniques
available for the predicting the stock market value. This paper
aims at survey on the use of Neural Network (NN), Data
Mining, Hidden Markov Model (HMM), Neuro Fuzzy system,
Rough Set Data Model and Support Vector Machine
techniques for predicting the stock market variation. In this
paper, a methodology is proposed for forecasting to provide
better accuracy when compared to the traditional methods.
Fig 1: Various prediction techniques
Keywords: Data Mining, Hidden Markov Model, Neural
Network, Neuro Fuzzy system, Rough Set. The survey of recent techniques such as NN, Data Mining,
Neuro Fuzzy system, HMM and Rough Set Data Model system
1. Introduction offer useful tools for anticipating the noisy environments of
stock market. This article aims at providing intelligent
techniques to anticipate market prices. A stock market index is
Stock market plays a vital role in the economic
the representation of the movement in the “average of several
performance. Basically, it is used to deduce the economic
individual stocks”. Resistant characteristics are not taken into
situation of a particular nation. However, information regarding
consideration in forecasting process. To overcome these
stock market is typically incomplete, uncertainty and indefinite,
drawbacks, researchers could develop a model to forecast
making it a challenging task to predict the future economic
individual stock prices [3].
performance. More specifically, the stock market's variations are
analyzed and predicted in order to access knowledge that could 2. LITERATURE REVIEW
help to guide the investors, when to buy or when to sell, and to
hold a financial asset. In general, prediction means to know Phichhang Ou and Hengshan Wang applied ten different data
about the future. So, for the investment or trade in the market, mining techniques to anticipate price variation of Hang Seng
prediction of market value is very much essential. The market index of Hong Kong stock market [4]. Among those 10 methods
movement changes frequently, valued ahead, difficult to predict LS-SVM and SVM generate high ranking predictive
and disorganized in nature [1]. Hence by using only technical performance. Mostly, SVM is best as to compare with LS-SVM
analysis methods to anticipate stock market is very difficult, for in sample prediction. Since, in case of hit rate and error rate
similar to that of time series analysis. Essential analysis typically LS-SVM is better than SVM for considering the out sample
works best over longer periods of time, where technical analysis forecast.
is more appropriate for short term trading. Researchers have
made several attempts to predict the performance of financial Suresh et al. use different data mining techniques that are able
market. Many artificial intelligent techniques such as Neural to discover the hidden pattern; forecast the future trends and
Network and Fuzzy Systems have been proposed [2]. Since it is behaviors in financial market [5]. Pattern matching techniques
difficult to interpret their results, they are unable to visualize are found to be descriptive in time-series analysis. In this paper,
clearly the nature of interactions between technical indicators they used an algorithm to accommodate a flexible and dynamic
and stock market variations. The difficulty in case of technical pattern-matching task in time series analysis. Apart from
analysis is that, it requires a complete pattern to make an segment size the instance to sub-time-series size affects the
accurate prediction on the stock movement. Preferably, such a
system performance. In this paper, the ratio was set to 1 and also Md. Rafiul Hassan et al. in [8], deployed a fusion model by
the ratio was reduced to obtain better result. combining Hidden Markov Model (HMM), Artificial Neural
Networks (NN) and Genetic Algorithms (GA) to anticipate
Binoy et al. used hybrid decision tree-neuro-fuzzy system financial market prediction. In the proposed fusion model, an
methodology for forecasting of stock market. Automated stock NN is employed as a black- box to introduce noise to the
market trend anticipation system was proposed using decision observation sequences so that they may be better fitted with the
tree adaptive neuro-fuzzy hybrid system [6]. They used different HMM’s. GA is then applied to find out the optimal initial
techniques like technical analysis and decision tree. First, parameters for the HMM’s, given the transformed observation
technical analysis is used for feature extraction, and then the sequences. By using this fusion model there are wide range of
decision tree for feature selection. The reduced dataset obtained options to find number of alternative data items from historical
by these two is fed as input, to train and test the adaptive neuro- data. This type of data items is responsible for market behavior
fuzzy system for next day stock prediction. They tested their for current day. Then average of the price differences for the
proposed system on four major international stock market data. identified data items is calculated. After this, average is added to
Their experimental results clearly showed that the proposed the current day price. The value obtained is the forecast value of
hybrid system produces much higher accuracy when compared a particular day. This model consist of two phases
to stand-alone decision tree based system and Adaptive Neuro
Fuzzy Inference System (ANFIS). The above proposed neuro- Phase 1: Optimizations of HMM.
fuzzy system is as shown in Fig 2.
Phase 2: Using weighted average method to obtain the forecast.
Aditya Gupta and Bhuvan Dhingra in [7] used Hidden Markov
Model (HMM) for predicting the market prices. With the help of The schematic representation of fusion model is as shown in
historical stock prices they present the Maximum a Posteriori Fig.3.
HMM approach for anticipating stock values for the next day.
For training the continuous HMM they consider the intraday
high and low values and fractional change in stock price. Some
of the already existing methods like HMMs and Artificial Neural
Networks use Mean Absolute Percentage error (MAPE) to
minimize inaccuracy rate. They have tested their approach on
several markets, and compared the performance. Finally, they
present an HMM based Maximum a Posteriori (MAP) estimator
for market predictions. The model uses a latency of days to
predict the stock value for the next day. Using an already trained
continuous HMM MAP, decision is made over all the possible
values of stock. They assume the four underlying hidden states
viz fractional change, fractional high, fractional low emit the
visible observations.
4. Acknowledgement
We express our sincere thank to all the authors, whose papers in
the area of Stock Market Prediction are published in various
conference proceedings and journals.
References
i. S.Arun , Joe Babulo, B. Janaki, C. Jeeva, “Stock Market Indices
Prediction with Various Neural Network Models”, International Journal of
Computer Science and Mobile Applications, Vol. 2, Issue 3, .pp32-35 march
2014.
ii. URL:
http://www.learnartificialneuralnetworks.com/stockmarketprediction.html.
x. Yang, Kongyu, Min Wu, and Jihui Lin. "The application of fuzzy
Step 4: For each generated reduct data set, and its corresponding neural networks in stock price forecasting based On Genetic Algorithm
objects construct the decision rule. discovering fuzzy rules.", In Natural Computation (ICNC), 2012 Eighth
International Conference on, pp. 470-474. IEEE, 2012.
ABSTRACT:Handling internet traffic in these days is not so hardware or network infrastructure to form a single, logical,
easy, but the explosive growth of internet traffic is hard to storage and compute platform, or cluster,that can be shared by
collect, store and analyze internet traffic on a single machine. multiple individuals or groups. Computation in HadoopMap
Hadoop has become a popular framework for massive data Reduce is performed in parallel, automatically, with a simple
analytics. It facilitates scalable data processing and storage abstraction for developers that obviate complex synchronization
services on a distributed computing system consisting of and network programming. Unlike many other distributed data
commodity hardware. In this paper, I present a Hadoop based processing systems, Hadoop runs the user-provided processing
traffic analysis and control system, which accepts input from logic on the machine where the data lives rather than dragging
Wire shark (Log File), and output in form of summary which the data across the network; a huge win for performance.
contains entire internet traffic details and I also implemented
the congestion control algorithm to control the online network The main contribution of my work exists in designing,
traffic in the internet. Implementing and controlling of internet traffic through big-data
analytics. Firstly I have created a virtual network by using the
Mininet tool, instantly makes virtual network on my laptop it
KEYWORDS: Single machine, Hadoop, Commodity contains switches, routers, and hosts. It can be controlled by
Hardware and Wire shark. using open daylight controller. To capture the packets flow from
the virtual network we use a Wire shark like tool. we capture the
1. INTRODUCTION packet log file and save it in a text file and then log file is given
as input to the Hadoop to process the large data of log file and
Internet has made great progress in daily life and brought much we will visualized the summary report that contains the flow
more convenience to much more people daily life in recent analysis details which as sender ip , destination ip and it also
years, fact that it is still provides a kind of BOF (Best of Effort) have the size of byte sent. By using that file we control the
service to application has never been changed since its traffic by using the congestion control algorithm to control the
invention. online traffic.
Mininet is a network emulator, an instant virtual network our
laptop, It runs a collection of end-hosts, Switches, Routers and The main objective of the work includes:-
Links on a single Linux kernel. It uses a lightweight To design and implement a Traffic flow identification
virtualization to make a single system look like a complete system using Hadoop.
network, System, Code, running in the same kernel. The traffic flow identification system will be very
Open daylight is a controller used to control the flows running in useful for network administrator to monitor faults and
the Mininet, Mininet will connect to controller and setup a ‘n’ also to plan for the future.
tree topology.
Wire shark is a tool used to capture packets in the network. It’s a
free open source packet analyzer used for network trouble 2. BACKGROUND WORK
shooting, analysis, software and Communication protocol
development and education. Over the past few years , a lot of tools have been developed and
Hadoop was originally designed for batch oriented processing widely used for monitoring the internet traffic. Mininet is tool
jobs, Such as creating web pages indices or analyzing log data. widely used to setup a virtual network in your laptop. So that we
Hadoop widely used by IBM, Yahoo!! , Face book, Twitter etc., can able to simulate using these Mininet tool to identify the flow
to develop and execute large-scale analytics or applications for of the packets in the virtual network. Wire shark is a popular
huge data sets. Apache Hadoop is a platform that provides traffic analyzer that offer’s user friendly graphics interface.
pragmatic, cost-effective, scalable infrastructure for building Tcpdump is also most popular tool of capturing and analyzing
many of the types of applications described earlier. Made up of a the internet traffic. Open daylight is an controller used to control
distributed file system called the Hadoop Distributed File system of the packets in the Mininet virtual network. It acts like an
(HDFS) and a computation layer that implements a processing controller to control the packets form where to where the packets
paradigm called Map Reduce, Hadoop is an open source, batch needs to be sent that can be controlled by the Open daylight tool.
data processing system for enormous amounts of data. We live
in a flawed world, and Hadoop is designed to survive in it by not Most of the map reduce applications on Hadoop are developed
only tolerating hardware and software failures, but also treating to analyze large text, log files or web, in this we firstly packet
them as first-class conditions that happen regularly. Hadoop uses processing or analysis for Hadoop that analyzes trace file in
a cluster of plain old commodity servers with no specialized blocks it will process the block of file and give the result in
parallel in distributed environment.
For flow control we use the congestion control algorithm for the Today the internet users are growing very rapidly. Each and
control of internet traffic with the Hadoop. By this we able to every person is utilizing the internet through one or the other
control the packets flow control very easily and in the very way. so, possibly internet traffic may also increases. single it is
effective manner. not so easy to handle the very big internet traffic. And storing
these large data and processing these large data is not possible to
3. LITERATURE SURVEY handle by single system.
A lot of research is done to measure the performance of the The problem with this is handling the internet traffic using
internet traffic using Hadoop. S.J Shafer, S. Rixner, and Alan L single server is not scalable to handle bigger network and
[2]. Cox discuss about performance of distributed Hadoop file may be chances of single point of failure.
system. Hadoop is most accepted framework for managing huge
amount of data in distributed environment. Hadoop makes use of 5. PROPOSED SYSTEM
user-level file system in distributedmanner. The HDFS (Hadoop
Distributed File System) is a portable across both hardware and
software platforms. In
this paper a detailed performance analysis of HDFS was done
and it displays a lot of performance issues. Initially the issue was
on architectural bottleneck that exist in the Hadoop
implementation which resulted in the inefficient usage of HDFS.
The second limitation was based on portability limitations which
limited the java implementation from using the features of naive
platform. This paper tries to find solution for the bottleneck and
portability problems in the HDFS.
In the above figure 3 show the entire network how the flows can
be able control. I have implemented the congestion control
algorithm to control the flows from source host to destination
hosts.
If the packet is of byes exceeds above some range then the path
form host-1 to host-2 can be changed or else the packet bytes
doesn’t exceeds the range the old path form host-1 to host-2 can
Figure 2:- Mininet Setup be used for packet transmission.
Here it will check for bytes I wrote the algorithm that will
In the above figure host1 and host2 are the source and handle control on only bytes.
destination of the virtual network with in the laptop. S1 and S2 Case-1: If byte >= specified then the path is changed accordingly
are the switches that are present in between the hosts and and
corresponding paths are controlled by using the Opendaylight Case-2: If bytes<specified then the path is same old one.
controller I can able to control the virtual network setup in the
computer. Flows and operations on the network can be modified
or changed by the Opendaylight control.
6. SCREEN SHOTS
III. Capturing Packet Flow using Wire shark
Wire shark is the tool used to capture and modify the packet
flow with in the network.
After setting up of network through mininet, then next step is
used to capture the packet flows from source host to destination
host in between some switches is used to connect the end hosts.
In this wire shark tool is used to capture the packet flows details
and it is in the form of log Files, and the collected log Files is
stored in text file, and further it is processed to the next step.
7. CONCLUSION
Figure-2:- Controlling of network through open daylight In this paper, I have presented the work on Idenifing and
controller. Monitoring the Internet Traffic with Hadoop.Setting up of the
network and obtaining the trace file and input trace file given as
input to the Hadoop and flow analysis can be done. And we also
implemented the congestion control algorithm to control the
internet traffic analysis. Flow control can be done by using
congestion control algorithm to control the internet traffic in the
Hadoop cluster.
REFERENCES
Abstract— The E-business sector is rapidly evolving and the marketing sites so that fraudulent retailers can be caught
needs for web market places that anticipate the needs of the immediately after the first wave of buyer complaints. In addition,
customers and the trust towards a product which are having proactive moderation systems are built to allow human experts to
more good rating. When most of the people are benefited from manually investigate suspicious retailers or buyers. Even though
the online website trading, culprits are also taking advantage to electronic-commerce sites spend a large budget to fight frauds
conduct fraudulent activities against honest parties to obtain with a moderation system, there are still many outstanding and
more fake profits. Therefore understanding the requirement for challenging cases. Criminals and fraudulent sellers frequently
analysing for user needs and trust providence in order to change their accounts and IP addresses to avoid being caught.
improve the usability and user retention of a website can be Also, it is usually infeasible for human experts to investigate
addressed by personalizing and using a fraud product detection every buyer and seller to determine if they are committing fraud,
system. especially when the e-commerce site attracts a lot of traffic. The
Keywords— online auction, fraud detection, fraud prevention, patterns of fraudulent sellers often change constantly to take
online authentication. advantage of temporal trends. For instance, fraudulent sellers
tend to sell the “hottest” products at the time to attract more
potential victims. Also, whenever they find a loophole in the
1. INTRODUCTION fraud detection system, they will immediately leverage the
Since the emergence of the World Wide Web (WWW), weakness.
electronic commerce, commonly known as e-commerce, has In this paper, we consider the application of an authentication
become more famous. Websites such as eBay and so on likestart- prevention technique for auction fraud detection in a major
up websites allow Internet users to buy and sell products and auction site, where hundreds to thousands of new auctions take
provide services online, which benefits everyone in terms of use- place every day. Therefore, it is necessary to develop an
fullness and profitability. The regular online shopping business automatic prevention system that only directs suspicious cases for
model allows sellers to sell a product or service at a default price, expert inspection, and passes the rest as clean cases. The
where buyers can choose to purchase if they find it to be a good moderation system for this site extracts rule-based features to
deal. Online auction however is a different business structure by make decisions. Where with years of experience human experts
which items are sold through price auction. There is often a have created many set of rules to detect the suspicious fraudulent
starting price and expiration time specified by the retailer. Once culprits and the resulting features are often binary.1 using rank
the auction starts, possible buyers bid against one another, and algorithm that is for instance we can create a binary feature from
the winner gets the item with their highest winning auction. giving the ratings i.e. if the feature value is 1 if the rating of a
Similar to any formats supporting economic transactions, online seller is lower than a threshold; otherwise it is 0. The final
auction attracts criminals to indulge in fraudulent activities. The prevention decision is based on the fraud score of each case.
varying types of auction fraud are: Products purchased by the Which can be done only by preserving the database of a retailers
buyer are not delivered by the retailer. The delivered products do and buyers with all the basic details and investigation is done
not match the descriptions that were posted by the retailer. while keeping their workload at a reasonable level.
Malicious retailers may even post non-existing or fake items with Since the fraudulent sellers change their pattern very fast, it
fake description to cheat buyers, and request payments to be requires the model to also evolve dynamically. However, for
strange directly to them via bank-to-bank wire transfer. offline models it is often non-trivial to address such needs. Based
Furthermore, some culprits apply e-mail techniques to steal high- on the reviews if the case is determined as fraudulent, all the
rated retailer's accounts so that possiblebuyers can beeasily cases from this retailer along with his pending products will be
cheated due to their high rating. Personaffected byfraud removed immediately. Therefore, smart fraudulent sellers tend to
transactions usually lose their amount and in most cases are not change their patterns immediately to avoid being caught. Also,
refundable. As a result, the status of the online auction services is since the training data is from human labelling, the high cost
hurt significantly due to culprit crimes. makes it almost impossible to obtain a very large sample.
To provide some security against fraud, internet marketing Therefore for such systems (i.e. relatively tiny sample size with
sites often provide security to culprit victims to cover their loss many features with temporal pattern), authentication prevention
up to a certain amount. To reduce the amount of such feature selection is often required to provide good performance.
compensations and improve their online fame, internet marketing Human experts are also willing to see the results of authentication
providers often adopt the following approaches to control and prevention feature selection to monitor the effectiveness of the
prevent fraud. The identifies of registered users are validated current set of features.
through email, SMS, or phone verifications. A rating system Our Contribution: In this paper, we study the problem of
where buyers provide feedbacks is commonly used in internet building online models for the authentication prevention
technique system, which essentially evolves dynamically over [7] to train convenient decision treesto select good sets of
time. We propose rank probit authentication model frame work features and make predictions. Developed [23] another
for the binary response we are applying a well known technique convenient approach that uses social networkanalysis and
in statistical literature called search variable selection (SSVS). decision trees. Proposed [38] an offline reversion structuring
The paper is organised as follows. In Section 2 we first framework for the auction fraud detectionmoderation system
summarize several specific features of the application and which incorporates domain knowledgesuch as coefficient bounds
describe authentication prevention frame work with fitting details. and multiple instance learning.
We review the related work in literature in Section 3. In section 4 In this paper we treat the fraud detection problem as a
we show the experimental results that compare all the models binaryclassification problem. The most frequently used
proposed in this paper and several simple baselines. Finally, we structuresfor binary classification include planning reversion
conclude and discuss future work in section5. [26],probit reversion [3], and support vector machine (SVM) [12]
anddecision trees [29]. Feature selection for reversion structures
2. OUR METHODOLOGY isoften done through introducing an act of punishing on the
Our application is to detect online frauds for a major websites coefficients.Typical punishments include vault regression [34]
where hundreds of thousands of new auction cases are posted (L2 penalty)and Lasso [33] (L1 penalty). Compared to vault
every day. Every new case is sent to authentication prevention regression,Lasso shrinks the unnecessary coefficients to zero
system for in advance to assess the risk of being fraud. The instead ofless values, which provides both awareness and good
system is being featured by: performance.Stochastic search variable selection (SSVS) [16]
Rule-based features: Human experts with years of uses“spike and slab” prior [19] so that the posterior of the
experience created many rules to detect whether a user is fraud or coefficientshave some possibility being 0. Another approachis to
not. An example of such rules is “blacklist”, i.e. whether the user consider the variable selection problem as model selection,i.e. put
has been detected or complained as culprit before. Each rule can priors on structures (e.g. a Bernoulli prior oneach coefficient
be concerned as a binary feature that indicates the fraud likeliness. being 0) and compute the marginal posteriorprobably of the
Selective labelling: If the fraud score is above a certain structure given data. People then eitheruse Markov Chain Monte
threshold level, the case will enter a queue for further Carlo to sample structures from themodel space and apply
investigation by human experts. Once it is evaluated, the final Bayesian model averaging [36], ordo a stochastic search in the
result will be labelled as Boolean feature, i.e. genuine or culprit. structure space to find the posteriormode [18]. Among non-linear
Cases with higher scores have higher priorities in the queue to be models, a tree model usually handles the non-linearity and
evaluated. The cases whose fraud score are belowthe threshold variable selection simultaneously.Representative work includes
are determined as clean by the system without any human decision trees[29], random forests [5], gradient boosting [15] and
judgment. Bayesianadditive regression trees (BART) [8].
Authentication prevention considers the scenario thatthe
Fraud churn:Once one case is labelled as fraud by
input is given one piece at a time, and when receiving a batch of
human experts, it is very likely that the retailer is not trust-able
and may be also selling other fraudulent products; hence all the input the structure has to be updated accordingto the data and
items submitted by the same retailer are labelled as fraud too. The make predictions and servings for the nextbatch. The concept of
fraudulent retailer along with his/her cases will be removed from online modelling has been applied tomany areas, such as stock
the website immediately once detected. price forecasting (e.g. [22]), webcontent optimization [1], and
web spam detection (e.g. [9]).Compared to offline models, online
User feedback:Buyers can file complaints to claim
learning usually requiresmuch lighter computation and memory
Loss if they are recently cheated by fraudulent retailers.
load; hence it canbe widely used in real-time systems with
Similarly retailers may also complain if his/her products have
continuous supportof inputs. For online feature selection,
been judged as mistakenly
representativeapplied work include [11] for the problem of object
trackingin computer vision research, and [21] for content-
3. RELATED WORK basedimage retrieval. Both approaches are simple while in this
Online auction fraud is always recognized as an importantissue. paperthe embedding of SSVS to the online structuring is
There are contexts on websites to teach people how toprevent moreprincipled.Multiple instance learning, which handles the
online auction fraud (e.g. [35, 14]). Categorizesauction fraud [10] training datawith bags of instances that are labelled positive or
into several types and proposes strategies tofight them. negative,is originally proposed by [13]. Many papers has been
Statussystems are used extensively by websites to detect auction publishedin the application area of image classification suchas
frauds, although many of them useinnocent approaches. [25, 24]. The logistic regression framework of multiple
Summarized[31] several key propertiesof a good status system instancelearning is presented in [30], and the SVM framework.
and also the challenges for themodern status systems to extract
user feedback. Otherrepresentative work connecting status
4. EXPERIMENTS
systems with onlineauction fraud detection include [32, 17, 28],
We conduct our experiments on a real online auction fraud
where thelast work [28] introduced a Markov random field model
detection data set collected from a major earlier website. We
witha belief propagation algorithm for the user status.Other than
consider the following online structures:
status systems, machine learned models have been applied to
ON-PROB is the online probit reversion structure
moderation systems for monitoring anddetecting fraud. Proposed
Figure 3: The rates of missed customer complaints for work Figure 5: For ON-SSVSBMIL with daily batches, delta= 0.7
load rates equal to 25%, 50%, 75% and 100% for all the and omega = 0.9, the posterior mean of Bjt (j is the feature
offline models and online models with daily batches. index) over time for a selected set of features.
Finally, the most interesting set of features are the ones
For any test group, we regard the number of labeled that have a large variation of pjt day overday. One important
casesas the expected 100% workload N, and for any structure we reasonto use authentication prevention feature selection in our
Abstract—Cloud computing is a new technology where the sharing in the cloud. As a typical cloud application, an
users tend to get the services through internet based on their organization subscribes the cloud services and authorizes its
demand. In this new technology users should receive the staff to share files in the cloud. Each file is described by a set of
services without much delay and the costs also should be keywords, and the staff, as authorized users, can retrieve files of
reduced. The most important aspect in this environment is their interests by querying the cloud with certain keywords. In
maintaining privacy and efficiency. The original key word such an environment, how to protect user privacyfrom the cloud,
based file retrieval scheme proposed by ostrovsky allows the which is a third party outside the security boundary of the
users to retrieve the requested files without leaking any organization, becomes a key problem.
information but it causes heavy querying overhead. In this User privacy can be classified into search privacy and
paper we present an efficient information retrieval query access privacy[2]. Search privacy means that the cloud knows
(EIRQ) scheme to reduce the querying overhead in the cloud. nothing about what the user is searching for, and access privacy
In EIRQ, a user will give the query along with the rank then means that the cloud knows nothing about which files are
the user will retrieve the files based on the rank. The rank returned to the user. When the files are stored in the clear forms,
shows the percentage of files that will be returned to the user. a proper solution to protectuser privacy is for the user to request
allof the files from the cloud; this way, the cloud cannot know
Keywords - Cloud computing, cost efficiency, differential which files the user is really interested in. While this does
query services, privacy. provide the necessary privacy, the communication cost is
high.Private searching was proposed by Ostrovsky et al. [3][4]
I INTRODUCTION which allows a user to retrieve files of interest from an untrusted
server without leaking any information. However, the Ostrovsky
Cloud computing is the delivery of computing scheme has a high computational cost, as it requires the cloud to
resources over the Internet. It has been widely adopted in broad process the query on everyfile in a collection. Otherwise, the
applications and is becoming more pervasive.The main reasons cloud will assume that certain files, without processing, are of no
behind cloud computing sharp growth are increases in interest to the user. It will quickly become a performance
computing power and data storage, exponential growth of social bottleneck when the cloud needs to processthousands of queries
network data, and modern data centres, some of which can suffer over a collection of hundreds of thousands of files.To make
from high maintenance costs and low utilization. There are also private searching applicable in a cloud environment, the
challenges in the development of reliable and cost-effective previous work [7] designed a cooperateprivate searching
cloud-based systems.Cloud computing presents a new way to protocol (COPS), where a proxy server, called the aggregation
supplement the current consumption and delivery model for IT and distribution layer (ADL), is introduced between the users
services based on the Internet, by providing for dynamically and the cloud. The ADL deployed inside an organization has two
scalable and often virtualized resources as a service over the main functionalities: aggregating user queries and distributing
Internet. Cloud computing is the use of computing resources search results. Under the ADL, the computation cost incurred on
(hardware and software) which are available in the cloud can be largely reduced, since the cloud only needs to
execute a combined query once, no matter how many users are
executing queries.
Furthermore, the communication cost incurred on the
remote location and accessible over the network. cloud will also be reduced, since files shared by the users need to
be returned only once.Motivated by this goal, the new scheme,
Users are able to buy these computing resources as a utility, on named Efficient Information retrieval for Ranked Query (EIRQ),
demand.The name comes from the common use of a cloud- in which each user can provide his own percentage along with
shaped symbol as an abstraction for the complex infrastructure it the query to determine the percentage of matched files to be
contains in system diagrams. Cloud computing entrusts remote returned. The basic idea of EIRQ is to construct a privacy
services with a user's data, software and computation. preserving mask matrixthat allows the cloud to filter out a
certain percentage of matched files before returning to the ADL.
Cloud computing as an emerging technology is This is not a trivial work, since the cloud needs to correctly filter
expected to reshape information technology processes in the out files according to the rank of queries without knowing
near future [1]. Due to the overwhelming merits of cloud anything about user privacy.
computing, e.g., cost-effectiveness, flexibility and scalability,
more and more organizations choose to outsource theirdata for
Ranked Queries
To further reduce the communication cost, a differential
query service is provided by allowing each user to retrieve
matched files on demand. Specifically, a user selects a particular
rankfor his query to determine the percentage of matched files to
be returned. This feature is useful when there are a lot of files
that match a user’s query, but the user only needs a small subset
of them.
V CONCLUSION
REFERENCES
i. R. Ostrovsky and W. Skeith, “Private searching on streaming data,”
in Proc. of CRYPTOLOGY, 2005.
ii. R. Curtmola, J. Garay, S. Kamara, andR.Ostrovsky, “Searchable
symmetric encryption: improved definitions and efficient constructions,” in Proc.
ofACM CCS, 2006.
iii. “Private searching on streaming data,” Journal of Cryptology, 2007.
iv. G. Wang, Q. Liu, J. Wu, and M. Guo, “Hierarchicalattribute-based
encryption and scalable user revocation for sharing data in cloud servers,”
Computers & Security, 2011.
v. Q. Liu, C. C. Tan, J. Wu, and G. Wang, “Efficient information
retrieval for ranked queries in cost-effective cloud environments,” in Proc. of
IEEE INFOCOM, 2012.
vi. J. Bethencourt, D. Song, and B. Waters, “New constructions and
practical applications for private stream searching,” in Proc. ofIEEE S&P,
2013.
Abstract : Modern days are fully dependent on internet utilized to transmit the wanted information with a high level of
communication. Through net we can transfer data anywhere security and dependability while is passing through the
in the world to anyplace we want. The Internet was born out of unsecured channels [3]. The difficulties of media framework
academic efforts to share information; it never actually strove like advanced pictures, archives, sound, and feature rely on upon
for the high security process. It plays a key role in becoming two components that media framework data size is here and
people online, it is very easy and effective but dangerous too in there horribly monstrous and wish to be handled inside the
terms of data hacking and eavesdropping by hackers. It is constant [4]. Encryption calculations like DES, plan and RSA
needed that while using Internet data must be secured and aren't fitting for sensible picture encoding, especially underneath
should be personal. Image encryption can be used to protect the condition of undertakings of on-line interchanges [5].
data during transmission. Image encryption is a suitable Militaries, governments, privately owned businesses have
process to protect image data. There are many cryptographic utilized the encryption for quite a while to encourage mystery
algorithms which are being used to secure multimedia data like correspondence. The routine cryptographic frameworks chiefly
images, but they have some definite advantages and have been produced for securing alphanumeric information as
disadvantages. So there is a requirement to develop a strong opposed to the picture and sound signals. The encryption of
image cryptography algorithm for securing the image while sound signs with customary encryption obliged extensive
transferring. In this paper, a new symmetric key cryptography measure of processing power and time. A quick, dependable,
algorithm has been proposed for color 3D images. In this and powerful calculation is needed to scramble both picture and
algorithm a different type of key generation method is being sound with less processing time and high level of precision [6].
introduced. This technique is unique and is used for th e first Mechanical advances in the advanced substance process,
time for key generation. Here two public keys are used in generation and conveyance has offered ascent to a scope of late
cryptography process. Key generation is very important in flag handling applications amid which security dangers won't be
symmetric as well as in the asymmetric key cryptographic taken care of in an exceedingly established style. These
algorithm. Here, we propose a work for developing a new applications fluctuate from the interactive media framework,
symmetric key cryptography algorithm for image data to content creation and dissemination of cutting edge biometric
provide a secure transmission during the network sign procedure for validation, biometric recognizable proof and
communication. All the concepts related to this area are access administration. In a few of those cases, security and
explained. This algorithm is totally lossless, such that image protection dangers may block the reception of late picture and
pixel are preserved during encryption and decryption. feature transforming administrations. Therefore, the vocation of
cryptanalytic systems in picture and feature process applications
Key Words —Encryption, Decryption, Image. is transforming into more regular. The cryptanalytic methods
used in these applications will be utilized as a part of two
I. Introduction measurements (20) grid [7].
The quick improvement of PC system correspondence, there
is so natural to acquire computerized pictures through the system II. Literature Review
and further utilize, imitate and disseminate them. Computerized
innovation brings us much handiness, however it likewise gives As indicated by Dr. Mohammad V Malakooti Mojtaba
a chance to assailant or unlawful client to hack our own Raeisi Nejad [6], they have proposed an calculation for pictures
information. Regularly, there are two noteworthy methodologies taking into account a novel misfortune less advanced encryption
which are utilized to ensure pictures. One is data concealing framework for interactive media utilizing the orthogonal changes
which incorporates obscurity, watermarking, steganography and for the encryption of picture information. This technique is in
spread channel. The other is encryption which incorporates light of the square figure symmetric key cryptography. Creators
traditional cryptography calculation [I]. The field of encryption have an accentuation on the improvement of a novel lossless
and security is getting to be essential in the twenty first century, advanced encryption framework for interactive media. They
when an enormous measure of data is transmitted over the utilized the symmetric properties of the orthogonal changes to
neighborhood and additionally the Internet. The advanced ascertain the opposite of the orthogonal grids amid the execution
information and pictures account more than two-third of data of the decoding methodology. They utilized a few traditional
that is transmitted over the Internet [2].Thus, a very dependable picture encryption methodologies, for example, Discrete Cosine
and hearty encryption calculation is obliged when the data is Change (OCT), Hadamard Transform (HT) and additionally
transmitted over the unsecured channels. Information encryption Malakooti Transform (MT) [6]. As indicated by Sahar Mazloom,
and information inserting are the most vital implies that can be Amir-Masud Eftekhari -Moghadam, a picture encryption is by
Abstract— ATM is the most significant service provided by a new Rank Based Genetic Algorithm for solving the banking
banking sector to the customers. Optimally deploying ATM’s is ATM location problem using Convolution (RGAC) which
very complex. The effective deployment of ATM depends upon outperforms the Heuristic Algorithm based on Convolution
various factors such as, where the customers lives, where they (HAC) algorithm that is inefficient while market size increases.
work, roads they travels and the cost to reach ATM. Genetic (RGAC) increases the search efficiency by improving the
algorithms are used to solve such optimization problems using evolutionary process while meeting a feasible solution.
techniques such as inheritance, mutation, selection, and Moreover, RGAC has proved to be a robust approach for solving
crossover. A banks decision to deploy ATM’s should be logical the ATMs deployment problem and is able to provide high
as well as profitable which provide greater convenience and quality solutions in a reasonable time.
covering larger market area with maximum customers. The The rest of the paper is structured as follows: Section II
objective is to minimize the total number of machines but indicates some important related work RGAC.A detailed
covering all the customer demands in the selected area. This description of the problem encoding and specific operators are
study proposes a Rank Based Genetic Algorithm using explained in Section III. Section IV explains about RGAC and
convolution for solving the Banking ATM’s Location Problem VI section includes concluding remarks.
(RGAC).
RGAC is one of the ATM deployment strategy based II. RELATED WORK
on rank concept which gives high feasible solution in The study [1] investigated placement of minimum number of
reasonable time. RGAC gives cost efficient allocation of ATM machines covering all customer demands in given
ATM’s and computing percentage coverage(PC Covering geographical area. They have developed a heuristic algorithm to
whole area ) which is high as it covers customers demands by efficiently solve the optimal location problem of ATM
maximizing the service utility of each machine . placement by formulating a mathematical model.
In this study, the problem of finding the minimum number of
Key–Words: Genetic Algorithms (GAs), Rank, Automated ATM’s and their locations given arbitrary demand patterns is
Teller Machines (ATM), Percentage coverage (PC), Client considered. They have considered one particular area and
Utility matrix (CU), Service Utility Matrix (SU) ,Rank Based divided the parts of that accordingly such as area with no
Genetic Algorithm using convolution (RGAC). demand, high demand, and normal demand and so on with color
code. Using the variables the placement problem is modeled.
I.INTRODUCTION The study [2] presents the problem of WiFiDP (WiFi
Network Design problem) grouping problem. A hybrid grouping
ATM is an electronic banking outlet, which allows customers to genetic algorithm (HGGA) is proposed as a convenient method
complete basic transactions without the aid of a branch to solve such problems with providing a smaller and low cost
representative or teller. ATMs are scattered throughout cities, connection service. The popularity of WiFi-enabled devices
allowing customers easier access to their accounts. ATMs have represents an enormous market potential for wireless networking
become a competitive weapon to commercial banks whose services and mobile applications, based on this technology. The
objective is to capture the maximum potential customers. The deployment of citywide WiFi access networks is a location
fact is that commercial banks compete not only on the dimension problem as well as its a large assignment. In this case, the
of price but also on the dimension of location. ATM optimal grouping genetic algorithm is combined with a repairing
Deployment Strategies offer the opportunity to provide greater procedure, to ensure feasible solutions, and with a local search to
convenience and to attract more customers by covering the improve its performance for the case of the WiFiDP. The
money market with sufficient ATM facilities. These strategies grouping genetic algorithm (GGA) is a class of evolutionary
also provide greater cost efficiency by finding optimal number algorithm specially modified to tackle grouping problems, i.e.
of ATMs to be installed and greater profitability by increasing scenarios in which a number of items must be assigned to a set
the ATM user base in order to earn much more transactions and of predefined groups. Thus, in the GGA, the encoding,
services fees as well as through the inflow of deposits from the crossover, and mutation operator of traditional GAs are
depositors. modified, obtaining a compact algorithm with very good
The location depends on the transactions demanded by the performance in problems of grouping.
customer of proprietary ATM and non-proprietary ATM. A The study [3] investigate the ATM placement problem
bank’s decision to deploy ATMs should be a rational Economic which is significant service provided by bank to customers.
decision using the best ATM deployment strategy that takes into Many banks utilize ATMs to make cash withdrawal available to
account the high computation complexities. This paper proposes their customers at all times. They have formulated the ATM
Data Oblivious Caching Framework for Hadoop using MapReduce in Big data
Sindhuja.M, Hemalatha.S
Assistant professor- Information Technology, PG Scholar - software engineering
sindhuja.m@rajalakshmi.edu.in,hemzmohan12@gmail.com
RAM Cloud [9] and prior work on databases such as MMDB [2]
stored all data in RAM only. While this is suited for web servers, it
is unlikely to work in data- intensive clusters due to capacity
reasons – FaceBook has more storage on disk that aggregate
memory. Proposed system thus treats memory as a constrained
cache. The study of speculative execution of tasks described the
execution which potentially slows down the entire MapReduce
job in order to accelerate [15] the execution of a MapReduce job
and also does not address the data sharing problem which is
identified. This mechanism is orthogonal to proposed work and
could be integrated straight forwardly.
Figure 2: The Docache infrastructure
Distributed systems such as Zebra [5] and XFS developed for the
Sprite operating system make use of client-side in-memory block Figure 2 shows the overall infrastructure of the system.
caching, also suggesting using the cache only for small system. Docache is a mechanism which is used to access the cached data
However, these systems make use of relatively simple eviction with less time and resources. All the local caches can be
policies and do not coordinate scheduling with locality since they coordinated by the distributed cache called Docache. It uses the
were designed for usage by a network of workstations. data oblivious caching algorithm for processing the data. because
this algorithm is much easier to analyze than a real cache`s
According to PACMan [14] when multiple jobs run in parallel, characteristic such as replacement policies etc. This framework
job’s running time can be decreased only when all the inputs does not depend on variables/hardware parameter such as cache
related to running a job are cached. Caching only part of the inputs size or cache line length. It is efficient usage of processor caches
will not help in improving the performance. These massive and reduction of memory bandwidth requirements.
distributed clustered systems have large memories and job
execution performance can be improved if these memories can be For application data, a distributed cache keeps a copy of a
utilized to the fullest. PACMan is a caching service that subset of the data in the database and it is also temporary in nature.
References
RUN TIME(ms)
1 13750 i. Avita Katal, Mohammad Wazid, R H Goudar “Big Data: Issues,
Challenges, Tools and Good Practices”, 2013 IEEE.
2 25 ii. H.Garcia-Molina and K. Salem., “Main Memory Database Systems: An
Overview”, In IEEE Transactions on Knowledge and Data Engineering, 1992.
iii. Hadoop, http://hadoop.apache.org/, 2013.
3 23 iv. Jing Zhang, Gonqing Wu, XuegangHu, Xindong Wu “A Distributed cache
for Hadoop Distributed File System in time cloud services” on ACM/IEEE 13 th
4 20 international conference, 2012.
v. John H, Hartman and John K. Ousterhout “The Zebra Striped Network
File System”, In ACM SOSP, 1993.
vi. Meenakshi Shrivatava, Dr.Hans-peter Bischof, “Hadoop-Collaborative
Caching in Real Time HDFS”, Google, 2013.
vii. Memcached—A distributed memory object caching system,
http://memcached.org/, 2013.
viii. Nimrod Megidd, Dharmendra S. Modha, “Outperforming LRU with an
Adaptive Replacement Cache Algorithm”, IEEE trans. on distributed system,
2004.
ix. J. Ousterhout.K “The Case for RAMClouds: Scalable High- Performance
Storage Entirely in DRAM” in SIGOPS Operating Systems Review, 2009.
x. Pietro Michiardi, “Map Reduce Theory and Practice of Data Intensive
Applications”, Eurecom, 2011.
xi. K.Senthil Kumar, K. Satheesh Kumar, S. Chandrasekar, “Performance
Enhancement of data Processing using Multiple Intelligent Cache in
Hadoop”,IJIET,Vol.4, Issue 1, June 2014.
xii. Stephen Kaisler, Frank Armour, J. Alberto Espinosa, William Money, “Big
Data: Issues and Challenges Moving Forward”, IEEE, 46th Hawaii
International Conference on System Sciences, 2013.
xiii. Yaxiong Zhao, JIU Wu , “Dache: A Data Aware Caching For Big Data
Applications Using the MapReduce Framework”, Vol.19, No 1, February 2014
xiv. Zhiwei Xaio, Haibochen, Binyu sang “A hierarchical Approach to
maximizing MapReduce Efficiency” on international conference, 2011.
xv. M. Zaharia, A. Konwinski, A. D. Joseph, R. Katz, and I. Stoica, “Improving
mapReduce performance in heterogeneous environments”, in Proc. of
OSDI’2008, Berkeley, CA, USA, 2008.
VII. CONCLUSION
This paper has exposed the major data analyzing problems that
need to be addressed in Big Data processing and storage. We have
described Data oblivious caching, an in-memory coordinated
caching system for data processing in hadoop using MapReduce
framework. During Map-Reduce framework, Mapper nodes
Abstract : Multiple Input Multiple Output-Othogonal multipath fading, high power spectral efficiency,better
Frequency Division Multiplexing MIMO-OFDM is a most performance. However, at the saem time MIMO-OFDM
attractive technology which has been recently proposed in suffers with a problem of PAPR while implementing the
wireless communication. It provides high data rate services system. PAPR is defined as the peak to average power ratio
and better system performances. It improves data throughput which increases the analog to digital and digital to analog
and delivers highest capacity as well. However, MIMO-OFDM converter complexity and as a result it reduces the efficiency
suffers with the drawback of Peak to Average Power Ratio of the radio-frequency(RF) power amplifier.
(PAPR) for the large number of subcarriers which can effect
the system output. Therefore, to overcome the problem of There are several techniques used to reduce PAPR
PAPR reduction, an effective technique PTS (partial transmit performance in MIMO-OFDM system. The techniques are
sequence) is used. In this paper, modified PTS technique categorised into 3 types- Distotion methods, Distortionless
combined with interleaving and pulse shaping method has methods and other methods. These methods includes-
been presented to improve the performance of MIMO-OFDM Clipping,Companding,Selective mapping (SLM), Partial
system in terms of PAPR reduction. The basic idea behind PTS transmit sequence (PTS), Active constellation extension
is to analyses the influence of number of detected peaks on (ACE), Tone reservation (TR). Clipping considers a
PAPR performance and the system complexity by combining predetermined threshold which helps in reducing PAPR to a
signal subblocks and the rotation factors. The simulation loewst value. Interleaving combined with PTS, is also
results are computed by using MATLAB simulation which introduced in implementing MIMO-OFDM to reduce PAPR.
completely improves PAPR performance by using modified Interleaving is basically, transmission of reordering of
PTS combined with interleaving and pulse shaping method for consecutive bytes of data over a large sequence of data to
STBC MIMO-OFDM. reduce the effect of burst error. In this paper, PAPR is
reduced by PTS combined with interleaving and pulse
Key-Words: - MIMO-OFDM, PAPR, STBC, Partial shaping method in MIMO-OFDM. SLM and PTS belong to
Transmit Sequences, Interleaved Subblock Partition the probabilistic class because several different signals are
Scheme, Raised-Cosine pulse shape obtained but only the minimum PAPR signal is taken into
consideration. In SLM, several signals contains same
1. Introduction information data and one OFDM signal of lowest PAPR is
Orthogonal Frequency Division Multiplexing (OFDM) is a high selected. However SLM is a flexible technique but it
speed wireless communication technology which requires high computatinal complexity low bandwidth
has a demanding future in mobile communicatin system. It efficiency.
provides high data rates and high quality multimedia services to
mobile users and also delivers high data throughput and gives Therefore, an effective technique PTS (Partial Transmit
efficient wideband communication system. Due to all these Sequence) has been used in this paper which helps to reduce
advantages of OFDM it has been playing an important role in PAPR to a minimum value. PTS is a distortion type method
various communication system. Multiple antennas are used to and attractive technique used to improve the statistic of a
increase the capacity of wireless lines so have been a great deal multicarrier. In PTS, the data input information is divided
of interest in communication system. Space-time codes with into smaller disjoint subsequences. The input data is carried
OFDM results in wideband communication. By using multiple out and IFFT is performed and each subsequence are then
antennas at the transmitter as well as at the receiver end, spatial multiplied with rotating phase factors. The output combined
diversity can be achieved since it does not increase the transmit with rotating phase factors are then added to obtain 0FDM
power and signal bandwidth. Therefore, many high speed data symbol for transmission. Each and every subsequence
transmission standards have been presented such as determines the PAPR reduction. PAPR is computed for each
WiMAX(IEEE 802.16), WLAN(IEEE 802.11a/g), digital video resulting sequnce and the signal sequence with minimum
broadcasting (ADSL) etc. PAPR is considered and transmitted. The partitioning types
for PAPR reduction can be categorised as- interleaving
MIMO-OFDM is the technology which combines multiple partition, adjacent partition and pseudo-random partition.
input,multiple output, which multiples capacity by transmitting However, PTS in modified form is the better option
different signals over multiple antennas, and orthogognal compared to an ordinary PTS because in an ordinary PTS
frequency division multiplexing (OFDM). MIMO-OFDM has ,all the phase factor combination are considered which
several advantages of high data throughput, robustness against results in the increasing complexity with the several number
5. PROPOSED SCHEME
6. RESULTS
The figures below represents the PAPR reduction ouput and the
CCDF of MIMO-OFDM. Figure 1. Shows the MIMO-OFDM
PAPR reduction of OFDM subsequences which is taken at Fig.7.CCDF of PAPR for different oversampling factor L=
different subcarriers N = 64,128 and 256 For different 2, 4, 8, 16 when N=256, V=4,α=0.6 and Mt=2
subcarriers N= 64,128 and 256, the PAPR performace has been
improved from 8.6dB to 5.6dB. The PAPR reduces as the The results are obtained by using modified PTS technique
number of value of subcarriers used in MIMO-OFDM decreases. which is combined with interleaving and pulse shaping
This performance of PAPR has been improved by using modfied method. The waveform results for different subcarriers N =
PTS technique combined with interleaving and pulse shaping 64,128,256,512 and 1024 are represented in the figures. As
method. the number of subcarriers increases in MIMO-OFDM, the
PAPR also increases with respect to subcarriers.
7. CONCLUSION
The paper revolves around the idea behind the modified
PTS technique which helped to reduce the PAPR
perfromance for MIMO-OFDM. The PTS technique has
been used along with the interleaving and pulse shaping
method. The results presents that the PAPR reduction has
been improved to a greater extent i.e. PAPR has been
reduced from 9.5dB to 5.2dB. It is an effective technique
combined with interleaving and pulse shaping method for
STBC MIMO-OFDM system which is used to achieve a
better trade off between complexity and PAPR
performance. It also provides high data rates and helps to
provide the data thoughput in a very better way. MIMO-
Fig.5.CCDF of PAPR for different subcarriers N = 64, 128, 256 OFDM has several advantage and is very helpful in digital
when V=4, L=4, α=0.6 and Mt=2 multimedia and wirless broadband mobile communication
system.
8. REFERNCES
Abstract-The recent trend and requirement for large storage the live migration efficiency of multiple virtual machines from
in cloud computing has made migration and cloud experimental perspective and investigate different resource
virtualization technology increasingly popular and valuable in reservation methods and migration strategies in the live
cloud computing environment due to the benefits of server migrations. We first describe the live migration frame work of
consolidations, live migration, and resource isolation. Live multiple virtual machines with resource reservation technology.
migration of virtual machines can be used to implement energy Then we perform aseriesof experiments to investigate the
saving and load balancing in cloud data centre. However, to impacts of different resource reservation methods on the
our knowledge ,most of the previous work concentrated on the performance of live migration in both source machine and target
implementation of migration technology itself while didn’t machine .Additionally, we also analyze the efficiency of parallel
consider the impact of resource reservation strategy on migration strategy and workload-aware migration startegy.The
migration efficiency .This paper focuses on the live migration matrix such as down time ,total migration time, and work load
strategy of multiple virtual machines with different resource performance over heads are measured
reservation methods. We first describe the live migration
framework of multiple virtual machines with resource 2. Related Work
reservation technology .As soon as the virtual machine size
increases then the data which is in migration list is transferred Resource allocation is one of the major aspects of cloud
to the corresponding virtual machine. computing, dynamic resource allocation has its own challenges
that has to be while implementing it, there are many techniques
Keywords : virtualization technology, Amazon, which have come up in order to deal with it. The cloud
Google,Yahoo!,Microsoft, IBM and Sun comprises of data center hardware and software [1].The resource
allocation concept is been analyzed in many computing areas
1.Introduction such as grid computing and operating systems. The prediction
plays a very important role during the process of resource
Cloud computing has recently received considerable attention in allocation. The prediction of CPU utilization for the upcoming
both academics and industrial community as a new computing demand has been studied in the literature. A prediction method
paradigm to provide dynamically scalable and virtualized was proposed based on Radical Basis Function (RBF) network
resource as a service over the internet Currently, several large by Y.Lu et.al for the purpose of predicting the user access
companies, such as Amazon, Google,Yahoo!,Microsoft, IBM demand. Also came up with the concept of multi –scaling. The
and Sun are developing their own cloud platforms for consumers statistical expected value can be obtained by service provider by
and enterprises to access the cloud resources through services . using methods [13][14].Content Delivery Network [CDN] and
Recently, with the rapid development of virtualization central resource by data center should be used to represent and
technology, more and more data canters use this technology to include multi-scaling for Video On Demand (VOD) with
build new generation data center to support cloud computing due guarantee Quality Of Service (QOS) [1].There exists the
it the benefits such as server consolidation, live migration and necessity for the consolidation of cost involved in the process of
resource isolation. Live migration of virtual machines means the streaming along with optimized experience of the user. There are
virtual machine seems to be responsive all the time during the different types of resource provisioning plan that can be chosen
migration of the clients’ perspective. Compared with traditional as per the requirement which are provided by cloud
suspend/resume migration ,live migrate holds many benefits providers.[15] The two common types include on-demand plan
such as energy saving, load balancing, and online maintenance and resource reservation by analyzing both types, the resource
.Many live migration methods are proposed to improve the reservation is said to be more inexpensive compared to on-
migration efficiency As the live migration technology widely demand plan.
used in modern cloud computing data center, live migration of
multiple virtual machines becomes more and more frequent. 3. System Architecture
Different from the single virtual machine migration, the live
migration of multiple virtual machines faces many new The components of the system architecture includes
problems, such as migration failures due to the insufficient Service Provider
resource in target machine, migration conflicts due to the Cloud server
concurrent migrations, and migration trashing due to the Android User
dynamic changes of virtual machine workloads.All the above
issues should be overcome to maximize the migration efficiency Service Provider: The login operation is the main operation
in virtual cloud data center environments In this paper, we study carried out at the service provider module. The login into the
Fig2:Data Flow
4.Conclusion
Live migration of virtual machines is an efficient technology
used to implement energy saving and load balancing in
Fig1: System Architecture virtualized cloud computing data center. This paper, we study
the live migration efficiency of multiple virtual machines from
experimental perspective and investigate different resource
is been utilized and also the remaining space. The cloud server reservation methods in the live migration process as well as
also contains the prediction details, which provides predictive other complex migration strategies such as parallel migration
information about the future requirements of space required for and workload-aware migration. Experimental results show that:
the files based on the previous statistical data. While the media (1) Live migration of virtual machine brings some performance
files are viewed the service provider /cloud administrator will overheads. (2)The performance overheads of live migration are
have to check for the current amount of space that is available affected by memory size,CPU resource, and the workload types.
for further addition of the new file. If there exists a condition, (3)Resource reservation in target machine is necessary to avoid
where the available space is less compared to the space available the migration failures.(3) The adequate system resources in the
on the cloud, then comes the concept of migration list. The data source machine can make more parallel number of migrations
which does not fit into the available space is placed on the and can obtain better migration efficiency.(5) The workload-
migration list which can be added on to the cloud server, the file aware migration strategy can efficiently improve the
will be stored in different place, all the data which is ready to be performance of migrated workload. Based on the experimental
added but its temporarily not possible to be added due to discoveries, three optimization methods, optimization in the
insufficient space .The migration list can be defined as the source machine, parallel migration of multiple virtual machines
temporary space where all the data is stored when the memory is and work-aware migration strategy, are proposed to improve the
not sufficient for the completion of the addition process.After migration efficiency. Future work will include designing and
the memory is expanded with sufficient space for the addition of implementing intelligent live migration mechanism to improve
the new file the re – migration process takes place i.e. the data the line migration efficiency in the multiple virtual machines
that is placed on the migration list is added to the server. scenario and studying the migration strategies as an optimization
problem using mathematical modelling methods.
Android user: The android user should first register with the
cloud by providing the necessary details. Once the registration References
process is completed successfully, the user can login with the i. M. Armbrust, A. Fox, R. Griffith, A. Joseph, R. Katz, A. Konwinski, G.
corresponding username and password. Lee, D. Patterson, A. Rabkin, I. Stoica et al., “A view of cloud computing,”
Communications of the ACM, vol. 53, no. 4, pp. 50–58, 2010.
The available files can be viewed and also be carried out .On the ii. C. Waldspurger, “Memory resource management in VMware ESX
server,” ACM SIGOPS Operating Systems Review, vol. 36, no. SI, p. 194, 2002.
selection of particular file, the user can view the rank of that iii. P. Barham, B. Dragovic, K. Fraser, S. Hand, T. Harris, A. Ho, R.
particular file along with the comments provided by other users. Neugebauer, I. Pratt, and A. Warfield, “Xen and the art of virtualization,” in
In this way the user can also provide the comments and rank for Proceedings of the nineteenth ACM symposium on Operating systems principles,
the particular multimedia file. 2003, p. 177.
Abstract : Many researches on real-time vehicle tracking is satellite with the resolution (frequency) based on the user
conducted, like wise tracking a school bus is important, requirement like 10 readings/minute in NMEA format will
sending a child to school by bus can be wracking for parents. It contain raw data with huge information.
is important to know if their child has boarded to the right bus
,safe on it and reached a correct destination(i.e school) on The microcontroller processes the raw data information
time. According to the statistics conducted by world health according to the algorithm present. The location co-ordination
organization (WHO).In India about 41% of children die due to instructs the GSM modem to provide serial communication to
lack of road transportation safety. This paper presents a server(database).
reliable real time tracking system using global positioning
system (GPS),global system for mobile communication (GSM)
services and RFID or smartcard ,which keeps real time
tracking of child at all time. Parents can log in into their
mobile or web to track the bus to know whether the bus is
running late and minimizes the time children to wait at bus
stop by which less time to be exposed to be exposed to criminal
predators, bad weather or any other dangerous condition
avoided to the child.
2. Introduction
There is a huge demand for tracking devices, which is
actually considered a life saving devices. These devices keep
track of children and update about the real time tracking to their Figure 1:Block diagram of School bus tracking
parents. During the time of disasters, these system helps the
parents to track their children location. According to hind[1] Figure 1 shows the overview of the tracking system installed in
tracking provides several services like stolen of assets, to keep the school bus.
track of the behavior of the employee at workplace environment.
RDID is used find the child login and logout from the
Parents must know about child safety in school bus, bus. There are two types of RFID namely active and passive
sending a child to school by bus is a wracking for their parents. RFID. In which active RFID reader will be centralized in the
The parents should know whether the child has boarded the bus, bus. In Passive RFID reader will be installed at door.
safely reached the school the school, found the right bus to reach
home on time. To keep the real time tracking of children by
using the GPS installed in school bus make parents to be bit
relaxed on their safety while travelling in school bus by
installing such safety components make bus and child tracking
easier and safer- includes accountability, increases the
convenience and savings.
The GPS receiver will receive the location co- The raw GPS data collected from the device installed in
ordination(longitude ,latitude, speed, device data) from the the school bus will send it to the NMEA server at the
REFERENCES
Abstract— Coming along with the recent development of Iot for healthcare by accurately monitoring, measuring and
(Internet of Things), wireless devices have invaded the medical analyzing a variety of health status indicators
science with a broad spectrum of possibilities. Along with
improving the quality of life of patients, wireless technology II. RELATED WORKS
enables patients to be monitored remotely during emergencies
and provides them health information, reminders, and support 1. The Sure CALL Labor Monitor technology is used for calling
potentially extending the reach of health care by making it the onset of labor. Uterine Electromyography Labor Monitoring
available anytime. The wireless sensor networks are inserted detects uterine muscle contractions from abdominal recordings
into the vaginal canal and can detect electrical signals of electrical signals generated in the uterus, uterine EMG
associated with uterine contractions, sensing that labor has activity can be measured by abdominal surface electrodes.
begun. These sensors detect signals directly from the specific
points in the body where they originate responsible for sensing Tocodynamometers are external pressure measurement devices
uterine contractions, even during the preterm labor. These which are being used to measure the contractions of the uterus
sensors are responsible for transmitting information to the and are the primary type of external monitor. The
cellphones which in turn alerts the maternity centers so that patient wears the device on a tightly attached belt, which must
patient receives apt treatment. maintain a constant pressure on a pressure-sensing device. As
the uterus contracts, a strain gauge measures the pressure exerted
Keywords : Internet of Things, wireless technology by the abdomen on the device.
I. INTRODUCTION
The Internet of Things refers to a wireless connectivity medium
between objects. Internet of things is not only a global network
for people to communicate with one another, but it is also a
platform where devices communicate electronically with each
other and the world around them. From any time and place,
connectivity for everyone, we will now have connectivity for
anything.
The wireless sensor nodes containing the sensors for sensing the
muscle contaractions are inserted into the vaginal canal and can
detect electrical signals associated with the uterine wall
contractions, a sign that labor has begun. Wireless sensor nodes
REFERENCES
CONCLUSION
ABSTRACT:Cloud Computing became an optimal solution for researches were introduced dynamic resource management [7 ,8]
business customers to maintain and promote their business with virtual systems. These systems will consider the available
needs to clients via internet. Now a days the cloud computing resources at server and allocates them to applications based on
allows the business costumer to scale up and down their application workload requirements. To achievethis dynamic
resource usage based on needs. In order to achieve resource managements systems follows unevenness algorithms and on
multiplexing in cloudcomputing, recent researches were demand resource allocation strategies. This approach will
introduced dynamic resource allocation through virtual manage the resources dynamically in an efficient manner with
machines. Existing dynamic approaches followed unevenness virtualization of cloud systems. Dynamic mapping of virtual
procedures to allocate the available resources based on current requirements with physical resources will also helpto avoid SLA
workload of systems. Unexpected demand for huge amount of violations[9] in cloud environment. Sometimes unexpected
resources in future may cause allocation failure or system demand for huge amount of resources in future may cause
hang problem. In this paper we present a new systematic allocation failure or system hang problem.
approach to predict the future resource demands of cloud from In order to mitigate these problems, in this paper we
past usage. This approach uses the resource prediction present a new systematic approach to predict the future resource
algorithm to estimate future needs to avoid allocation failure demands of cloud from past usage. This approach analyzes the
problem in cloud resource management. And skewness resource allocation logs of virtual server, SLA agreements and
algorithm to determine the unevenness in the multi- follows the demand prediction algorithm to estimate future
dimensional resource utilization of a server. needs to avoid allocation failure problem in cloud resource
management. Our approach uses the present and past statistics to
Keywords: Dynamic resource allocation, Cloud computing, predict the future requirements in an efficient manner. To do this
Resource Prediction Algorithm, Virtual machine migration, we proposed two different methodologies in this paper are (i)
Load balancing, Skewness, Green computing,Hotspot hours-bounded (ii) days-bounded resource prediction
migration and Cold spot migration. techniques.By integrating the results of these methodologies our
approach assess the reliable resource requirements in future.
INTRODUCTION Experimental results aresupporting our strategy is more scalable
and reliable than existing approaches.
Cloud computing is a fast growing technology that currently The rest of the paper is organized as follows:
being studied in[1].It has moved computing and data away from Section 2 discusses about related work, followed by proposed
desktop and portable PCs, into large data centers [2]. It has the system design which consist of load balancing Cloud
capability to harness the power of Internet and wide area architecture, Load prediction algorithm, Skewness algorithm and
network (WAN) to use resources that are available remotely, finally results and future scope.
thereby providing cost effective solution to most of the real life
requirements[3][4]. Majority of Business customers interested RELATED WORK
towards cloud computing and they started their app migration Computing is an emerging computing technology thatis
with cloud environment to promotetheir business operations to rapidly consolidating itself as the next big step in
end client with low investments and high availability. Due to this thedevelopment and deployment of an increasing number
increased adoption, Resource Management in Cloud (RMC) ofdistributedapplications.Cloud computing nowadays becomes
becomesanimportant research aspect in this area. quite popular amonga community of cloud users by offering a
Earlierapproaches [5][6] were used evenness procedure in variety of resources.Cloud computing platforms, such as those
resource distribution to allocate the available resources among provided byMicrosoft, Amazon, Google, IBM, and Hewlett-
the running applications. This approach may leads to Packard, letdevelopers deploy applications across computers
resourceover flow due to high amount of resourceallocation than hosted by acentralorganization.These applications can access a
required and resource underflow due to less amount of resource largenetwork of computing resources that are deployed
allocation than required. Always resource needs for a running andmanaged by a cloud computing provider.
application changes from time to time depends on number of In cloud platforms, resource allocation (or load
live clients. balancing) takes place at two levels. First, when an application is
uploaded to the cloud, the load balancer assigns the requested
In order to overcome resource overflow and resource
instances to physical computers, attempting to balance the
underflow problems from evenness distribution recent
computational load of multiple applications across physical
PRUC ← predictFutureRequirements(SRUC)
if(PRUC <= THRESOLD) setOverFlowFlag()
elsesetUnderFlowFlag()
wishList.addToWishlist(PRUC)
endforeach;
FRPC = doMigration(wishList)
return FRPC
End
Abstract--In this project pixel-based approach for urban land information about the data is available. The problem of
covers classification from high resolution satellite image using clustering points in multidimensional space can be posed
K means clustering and ISO data clustering. Pixel based image formally as one of a number of well-known optimization
analysis of image segmentation that is, clustering of pixels into problems, such as the Euclidean k-median problem, in which the
homogenous objects, and subsequent classification or labeling objective is to minimize the sum of distances to the nearest
of the pixels, and modeling based on the characteristics of center, the Euclidean kcenterproblem, in which the objective is
pixels is done using MATLAB GUI model. When applied to a to minimize themaximum distance, and the k-means problem, in
satellite image, the clustering algorithm approach involves two which the objective is to minimize the sum of squared distances.
steps. First, each group or cluster is homogeneous; i.e. Efficient solutions are known to exist only in special cases such
examples that belong to the same group are similar to each as the planar 2-center problem. There are no efficient exact
other. Second, each group or cluster should be different from solutions known to any of these problems for general k, and
other cluster, i.e. examples that belong to one cluster should be some formulations are known to be NP-hard. Efficient
different from the examples of other clusters. The algorithm approximation algorithms have been developed in some cases.
was implemented in MATLAB GUI model and was tested on These include constant factor approximations for the k-center
remotely sensed images of different sensors, resolutions and problem, the k-median problem and the k-means problem. There
complexity levels. are also approximation algorithms for the k-median and k-means
problems, including improvements based on coresets. Work on
Keywords: pixel based approach, high resolution satellite the k-center algorithm for moving data points, as well as a linear
image. time implementation of a 2-factor approximation of the k-center
problem have also been introduced. In spite of progress on
INTRODUCTION theoretical bounds, approximation algorithms for these
clustering problems are still not suitable for practical
Clustering algorithms for remote sensing images are used to implementation in multidimensional spaces, when k is not a
being divided into two categories: pixel-base and object-based small constant. This is due to very fast growing dependencies in
approaches. Using pixel based clustering algorithms for high the asymptotic running times on the dimension and/or on k. In
resolution remote sensing images, one could often find the practice, it is common to use heuristic approaches, which seek to
―pepper and salt‖ effect in the results because of the lack of and a reasonably good clustering, but do not provide guarantees
spatial information among pixels.Using pixel based on the quality of the results. This includes randomized
clusteringalgorithms for high-resolution remote sensing images, approaches, such as clara and clarans, and methods based on
one could often find the ―pepper and salt‖ effect in the results neural networks. One of the most popular and widely used
because of the lack of spatial information among pixels. In clustering heuristics inremote sensing is isodata. A set of n data
contrast, object-based clustering algorithms are not based on points in ddimensionalspace is given along with an integer k
spectral features of individual pixels but on image objects, i.e., indicating the initial number of clusters and a number of
segments. Consequently, in terms of semantics, the quality of additional parameters. The general goal is to compute a set of
image objects is heavily dependent on segmentation algorithms. cluster centers in d-space. Although there is no
In this letter, a novel clustering algorithm is proposed to detect specifyoptimization criterion, the algorithm is similar in spirit
geoobjectsfrom high-spatial-resolution remote sensing images tothe well-known k-means clustering method, in which the
using both neighborhood spatial information and probabilistic objective is to minimize the average squared distance of each
latent semantic analysis model (NSPLSA). The proposed point to its nearest center, called the average distortion. One
algorithm is not based on either pixels or segments but on significant advantage of isodataover kmeansis that the user need
densely overlapped sub images, i.e., rectangular regions, with only provide an initial estimate of the number of clusters, and
prefixed size. The probabilistic latent semantic analysis model based on various heuristics the algorithm may alter the number
(PLSA), which is also called aspect model, is employed to of clusters by either deleting small clusters, merging nearby
model all sub images. Every pixel in each sub image has been clusters, or splitting large clusters. The algorithm will be
allocated a topic label. The cluster label of every pixel in the described in the next section. As currently implemented, isodata
large satellite image is derived from the topic labels of multiple can run very slowly, particularly on large data sets. Given its
sub images which cover the pixel. Unsupervised clustering is a wide use in remote sensing, its efficient computation is an
fundamental tool in image processing for geosciences and important goal. Our objective in this paper is not to provide a
remote sensing applications. For example, unsupervised new or better clustering algorithm, but rather, to show how
clustering is often used to obtain vegetation maps of an area of computational geometry methods. Can be applied to produce a
interest. This approach is useful when reliable training data are faster implementation of isodataclustering. There are a number
either scarce or expensive, and when relatively little a priori of minor variations of isodata that appear in the literature .These
K MEANS ALGORITHM
CONCLUSION
REFERENCES