Secured Hadoop as A Service Based on Infrastructure Cloud Computing Environment

Copy­right Notice & Dis­claimer

© Atul Patil, 2016. All rights reserved. This arti­cle, titled “Secured Hadoop as A Ser­vice Based on Infra­struc­ture Cloud Com­put­ing Envi­ron­ment”, was authored and pub­lished by Atul Patil. It was orig­i­nal­ly fea­tured in the Inter­na­tion­al Jour­nal of Advanced Research in Com­put­er and Com­mu­ni­ca­tion Engi­neer­ing (IJARCCE), ISSN (Online): 2278–1021, ISSN (Print): 2319–5940, Vol­ume 5, Issue 3 (March 2016). The orig­i­nal pub­li­ca­tion can be accessed at https://www.ijarcce.com/upload/2016/march-16/IJARCCE%20261.pdf.

Dis­claimer: This arti­cle is repub­lished here by the orig­i­nal author, Atul Patil, in accor­dance with the copy­right poli­cies of the Inter­na­tion­al Jour­nal of Advanced Research in Com­put­er and Com­mu­ni­ca­tion Engi­neer­ing (IJARCCE). The con­tent remains unchanged to pre­serve its orig­i­nal­i­ty. For any inquiries or copy­right-relat­ed mat­ters, please con­tact the author direct­ly.

Abstract: This paper pro­pose sys­tem that gives out require­ment based allo­ca­tion of Hadoop as a ser­vice with infra­struc­ture cloud envi­ron­ments. Hadoop is effec­tive big data analysing and pro­cess­ing plat­form these days. But in instead of its promis­ing nature, researchers or pro­fes­sion­al orga­ni­za­tions are not tech­ni­cal­ly sound or hav­ing capac­i­ty to imple­ment and main­tain a work­ing Hadoop envi­ron­ment. That‟s, we are pro­vid­ing the secured Hadoop as a ser­vice. For cloud ser­vices to use com­put­ing ser­vices or ana­lyt­ics ser­vices by the cloud users is tru­ly prob­lem­at­ic. It’s a big issue to com­plete user‟s needs. Hence. On-Demand Hadoop ser­vice through cloud infra­struc­ture pro­vides a way to han­dle big data on the go. Poten­tial­ly strength­en­ing in secu­ri­ty prob­lems and achieves equal Job sched­ul­ing and quick process of huge infor­ma­tion in less quan­ti­ty of time and resources by com­put­ing the sci­en­tif­ic or any high per­for­mance com­put­ing jobs. Hadoop and Cloud tak­en the apps and soft­ware sys­tems and also dif­fer­ent data­bas­es to cloud data cen­tres, wher­ev­er the han­dling of the sen­si­tive data and process­es is not safe. Its secu­ri­ty loop­hole. Solu­tion is giv­en by pro­cess­ing and secur­ing the infor­ma­tion using cipher­ing deci­pher­ing and stor­ing them into the cloud servers.

Key­words: Hadoop as a Ser­vice, Encryption/decryption algo­rithm, Stor­age uti­liza­tion.

I.  INTRODUCTION

Cloud con­sid­ered as a quick­ly ris­ing new tech­nol­o­gy for deliv­er­ing com­put­ing as a util­i­ty. In cloud com­put­ing var­ied cloud cus­tomers demand type of ser­vices as per their dynam­i­cal­ly ever-chang­ing needs. Thus it’s the work of cloud to avail all the ser­vices to the con­sumers. But as a result of the sup­ply of lim­it­ed num­ber of resources it’s very trou­ble­some for cloud CSP to pro­duce all ser­vices. From the cloud providers’ per­spec­tive cloud resources should be allot­ted in a very hon­est man­ner. So, its impor­tant thing to com­plete cloud cus­tomers’ Qual­i­ty require­ment regard­ing ser­vices. So for acces­si­bil­i­ty a provider has to extra pro­vide keep an out­sized pro­por­tion of nodes idle so CSP unable to com­plete needs of cus­tomers. The neces­si­ty to stay of these nodes idle results in low uti­liza­tion. Way improve to keep least sec­ondary servers ide­al. But this occurs to most­ly reject­ing a tremen­dous amount of requests to some extent at that a provider now not pro­vides on-demand com­put­ing [2].

Oth­er con­sid­er­a­tion when eval­u­at­ing Hadoop provider of ser­vice with which the can accom­plish elas­tic demand. Any­one can think how open­ly Hadoop as ser­vice can accom­plish chang­ing demands for com­pute and stor­age resources with ease with­out wor­ry­ing about hec­tic of man­ag­ing and imple­ment­ing Hadoop. For exam­ple, Hadoop Big data jobs pro­duce lots of inter­me­di­ate results that may be tem­porar­i­ly stored and on cloud sec­ondary servers. Hadoop as a ser­vice trans­par­ent­ly expand and con­tract stor­age with­out sys­tem admin­is­tra­tor inter­ven­tion. Hadoop admin­is­tra­tors not required to fix para­me­ters or risk delay­ing jobs. Also show how well the Hadoop as a ser­vice from cloud han­dles work­loads. Orga­ni­za­tions that process all high per­for­mance jobs and sci­en­tif­ic analy­sis

by sci­ence indus­try will face a large traf­fic of mix­ture of work­loads. In the past decades Infra­struc­ture Ser­vice in cloud has become an attrac­tive ser­vice to the pro­vi­sion and man­age­ment of com­put­ing resources. An impor­tant thing of Infra­struc­ture ser­vice cloud is pro­vid­ing cus­tomers on- demand con­trol to com­put­ing infra. So to pro­vide use based resources, CSP should either keep extra resources on (or pay a high val­ue for oper­a­tive resources under­uti­lized) or remove an over­sized pro­por­tion of user requests (in that case the access isn’t any longer on- demand). At the same time, not all users need real­ly need basis access to infra­struc­ture [3].

Sev­er­al Jobs and sci­en­tif­ic work­ings are based on oppor­tunis­tic sys­tems wher­ev­er paus­es in ser­vice are pos­si­ble. Here a sys­tem is pro­posed, Hadoop as a ser­vice with the help of infra­struc­ture ser­vice that gives out on- need based pro­vid­ing the com­put­ing infra­struc­ture.

The tar­get is to han­dles and process big data in less CPU clock cycles and keep users away from has­sles of con­fig­ur­ing Hadoop at remote servers and strength­en­ing the use of resources by exe­cut­ing jobs through fair4s algo­rithm for equal dis­tri­b­u­tion of jobs, addi­tion­al­ly increase the uti­liza­tion of CPU by intro­duc­ing upload/download activ­i­ty. Data secu­ri­ty kept intact through RSA algo­rithm also used DES for com­par­i­son

.And used one which not only increase secu­ri­ty but also gives bet­ter uti­liza­tion.

They give var­i­ous tech­niques for analysing, pro­cess­ing and mod­el­ling work­loads. How­ev­er, the job prop­er­ties exe­cu­tion poli­cies are many in those sys­tems from the ones in a Hadoop sys­tem.

II.  THE PROPOSED SYSTEM

Fig.1 IAAS Cloud Archi­tec­ture

Hadoop Cloud pro­vides solu­tion for CPU, stor­age, time para­me­ters of improve­ment how­ev­er mov­ing mas­sive amounts of knowl­edge in asso­ci­at­ed in cloud pre­sent­ed an insur­mount­able chal­lenge [4].Cloud com­put­ing is a very unde­feat­ed par­a­digm of ser­vice des­tined com­put­ing and has rev­o­lu­tion­ized the means com­put­ing infra­struc­ture is abstract­ed and used. Fol­low­ing most well-liked cloud Solu­tions include:

  1. Infra­struc­ture Cloud Ser­vices
  2. Plat­form Cloud Ser­vices
  3. Soft­ware as Cloud Ser­vices

Thought can even be extend­ed to Stor­age as a Ser­vice. Var­i­ous data­base ser­vices also pro­vid­ed through cloud on air. Changes in infor­ma­tion access pat­terns of appli­ca­tion and there­fore the have to be com­pelled to scale intent on thou­sands of com­mod­i­ty Hard­ware led to birth of a replace­ment cat­e­go­ry of sys­tems referred to as Key-Val­ue stores [11].Area of big data ana­lyt­ics, we pro­pose the Map Reduce par­a­digm as Hadoop as a ser­vice through cloud with open-source imple­men­ta­tion Hadoop, in terms of usabil­i­ty and per­for­mance.

The Sys­tem imple­ments these mod­ules:

  1. Hadoop Cloud Con­fig­u­ra­tion
  2. Hadoop As Ser­vice Login por­tal
  3. Hadoop Cloud Admin­is­tra­tor por­tal
  4. Job Sched­ul­ing Algo­rithm
  5. Encryption/decryption mod­ule
  6. Third Par­ty Audit­ing.
    1. Hadoop Con­fig­u­ra­tion (Hadoop as a Ser­vice)

The Hadoop is a frame­work that per­mits for the decen­tral­ized process of big data across clus­ters of com­put­ers using straight­for­ward pro­gram­ming mod­els which is map reduce mod­el. it’s designed to pro­por­tion from sin­gle to then thou­sand servers, pro­vid­ing mas­sive com­pu­ta­tion and stor­age capac­i­ty, instead of think about

under­ly­ing hard­ware to give large avail­abil­i­ty, the infra­struc­ture itself is intend­ed to han­dle prob­lems, thus deliv­er­ing a most avail­able ser­vice on prime of a clus­ter of nodes, every of which can be vul­ner­a­ble to fail­ures [6]. Hadoop imple­ments Map reduce, using HDFS. The Hadoop Dis­trib­uted File Sys­tem gives clients to pos­sess name­space, unfold across sev­er­al lots of or thou­sands of clus­ters, mak­ing one big file sys­tem.

Frame­work allows to process large data with ease. Any of those splits (also told frag­ments or blocks) may be exe­cute on sec­ondary servers with­in the cloud infra­struc­ture. The present Hadoop sys­tem con­sists of the Hadoop archi­tec­ture, Map-Reduce, the Hadoop dis­trib­uted file sys­tem.

Job­Track­er is rou­tine for allo­cat­ing and run­ning MapRe­duce jobs in Hadoop on mas­ter serv­er node. There‟s one Job track­er exe­cutes on hadoop clus­ter. Job track­er works JVM. And slave node is assigned with task track­er node loca­tion. Job­Track­er in Hadoop per­forms; sched­ul­ing of assign­ments to task track­ers [9].

A Task­Track­er is slave node ser­vice with­in the Hadoop sec­ondary nodes that takes and process tasks like Map, reduce oper­a­tions from a Job­Track­er. Sin­gle task track­er per node. Task track­er runs on its JVM. Each Task­Track­er is hav­ing vacant slots, these tells the amount of jobs that it will set­tle for. The Task­Track­er has JVM to work­out tasks this is often to con­firm that process fail­ure does­n’t take down the task track­er [10].

The Hadoop Dis­trib­uted File Sys­tem (HDFS)

HDFS is a fault tol­er­ant and self-heal­ing dis­trib­uted fil­ing sys­tem designed to point out a clus­ter of busi­ness nor­mal servers into a mas­sive­ly scal­able pool of stor­age. Devel­oped specif­i­cal­ly for large-scale process work­loads where qual­i­ty, flex­i­bil­i­ty and turnout square mea­sure nec­es­sary, HDFS accepts data in any for­mat despite schema, opti­mizes for prime sys­tem of mea­sure­ment stream­ing, and scales to tried deploy­ments of 100PB and on the way side [8].

  • Hadoop as Ser­vice Login and Reg­is­tra­tion

It offered Inter­face to Login. Client will login to the Hadoop as a ser­vice through user inter­face and then upload the file and down­load file from web appli­ca­tion to hadoop cloud where we have already con­fig­ured the Hadoop pro­cess­ing which will process the data through effi­cient sched­ul­ing like fair4s algo­rithm and all obtain the detailed sum­mery of his account. Dur­ing this means secu­ri­ty is pro­vid­ed to the con­sumer by authen­ti­ca­tion infor­ma­tion for log­ging in to por­tal and stores it in info at the most serv­er that ensures the safe­ty. Client activ­i­ties records kept and used for audit trails. With this facil­i­ty, it ensures enough secu­ri­ty to con­sumer and infor­ma­tion hold on at the cloud servers sole­ly may be changed by the con­sumer.

  • Hadoop Cloud Admin­is­tra­tor

It is admin­is­tra­tion of Hadoop Ser­vices and Cloud infra­struc­ture. Cloud ser­vice sup­pli­er has an author­i­ty to fea­ture and take away clients and con­fig­ure Hadoop

ser­vices. It ensures enough secu­ri­ty on client‟s infor­ma­tion hold on at the cloud servers. Con­joint­ly the log records of every reg­is­tered and autho­rize con­sumer on cloud sole­ly will access the ser­vices. This spe­cif­ic con­sumer log record is helps in improve secu­ri­ty.

  • Fair4s Algo­rithm

Algo­rithm pro­cess­ing any kind of work­load small, large jobs Users spec­i­fy the work­load in terms of a map, reduce oper­a­tions Pro­grams writ­ten dur­ing this pur­pose­ful style area unit Auto­mat­i­cal­ly par­al­lelized and exe­cut­ed on an over­sized clus­ter of com­mod­i­ty machines. [7].

Our imple­men­ta­tion of Fair4s algo­rithm runs on an over­sized clus­ter of com­mod­i­ty machines and is very scal­able. Map-Reduce is Pop­u­lar­ized by open-source Hadoop project. Our fair4s algo­rithm works on process of enor­mous files by divid­ing them on vari­ety of chunks and assign­ment the tasks to the clus­ter nodes in hadoop mul­ti­mode con­fig­u­ra­tion. In these ways in which our planned Fair4s Job sched­ul­ing algo­rithm improves the uti­liza­tion of the Cloud sec­ondary servers with para­me­ters like time, CPU, and stor­age. Var­i­ous fea­tures of the Job sched­ul­ing algo­rithm are enlist­ed below.

  • Fea­tures of Fair4s:

Extra func­tion­al­i­ties avail­able in Fair4s algo­rithm cre­ate it work­load effi­cient than effi­cient mea­sure list­ed out below these func­tion­al­i­ties per­mits algo­rithm to pro­vides out effi­cient per­for­mance in process huge work load from total­ly dif­fer­ent clients.

  1. Set­ting Slots Quo­ta for Pools- All jobs are divid­ed into many pools. Every job belongs to at least one of those pools. Where­as in Fair4S, every pool is designed with a max­i­mum slot occu­pan­cy. All jobs belong­ing to a uni­form pool share the slots quo­ta, and also the range of slots employed by these jobs at a time is restrict­ed to the utmost slots occu­pan­cy of their pool. The slot occu­pan­cy high­er lim­it of user teams makes the slots assign­ment a lot of ver­sa­tile and adjustable, and ensures the slots occu­pan­cy iso­la­tion across total­ly dif­fer­ent user teams. Though some slots are occu­pied by some giant jobs, the influ­ence is bare­ly restrict­ed to the native pool with­in.
  2. Set­ting Slot Quo­ta for Indi­vid­ual Users-In Fair4S, every user is designed with a most slots occu­pance. Giv­en a user, regard­less of what num­ber jobs he/she sub­mits, the entire range of occu­pied slots won’t exceed the quo­ta. This con­straint on indi­vid­ual user avoids that a user sub­mit too many roles and these jobs occu­py too sev­er­al slots.
  3. Assign­ing Slots based on Pool Weight- Fair4S, every pool is designed with a weight. All pools that look ahead to a lot of slots type a queue of pools. Giv­en a pool, the preva­lence times with­in the queue is lin­ear to the bur­den of the pool. There­fore, high wait­ed pool are allot­ted with a lot of slots. Because pool weight can be changes and so small job fair­ness comes into pic­ture.
  4. Extend­ing Job Pri­or­i­ties- Fair4S intro­duces an in depth and quan­ti­fied pri­or­i­ty for every job. The task pri­or­i­ty is described by asso­ciate degree inte­gral range ranged from zero to a thou­sand. Gen­er­al­ly, at inter­vals a pool, a job

with a bet­ter pri­or­i­ty will pre­empt the slots used by anoth­er job with a low­er pri­or­i­ty. A quan­ti­fied job pri­or­i­ty con­tributes to dif­fer­en­ti­ate the pri­or­i­ties of small jobs in numer­ous user-groups. Pro­gram­ming Mod­el

  • Fair4s Algo­rithm

Fair4S, which is mod­eled to be fair for small jobs. In vari­ety of work­ing sit­u­a­tions tiny jobs are huge and lots of them require instant respons­es, which is an impor­tant fac­tor at pro­duc­tion Hadoop sys­tems. The inef­fec­tive nature of hadoop sched­ulers and GFS read write algo­rithm for work­ing with tiny sized jobs moti­vates us to use and ana­lyze Fair4S, which intro­duces pool weights and extends job pri­or­i­ties to guar­an­tee the rapid respons­es for small jobs [1] In this sce­nario clients is going to sub­mit jobs through client login on mas­ter serv­er where the Fair4s exe­cutes. On mas­ter serv­er the Audit­ing func­tions and equal dis­tri­b­u­tion of job is done through our pro­posed algo­rithm in effi­cient man­ner.

  • Pro­ce­dure of Slots Allo­ca­tion
  • The pri­ma­ry step is to allot slots to job pools. Every job pool is orga­nized with two para­me­ters of max­i­mum slots quo­ta and pool weight. In any case, the count of slots allot­ted to a job pool would­n’t exceed its most slots quo­ta. If slots demand for one job pool varies, the utmost slots quo­ta is man­u­al­ly adjust­ed by Hadoop oper­a­tors. If a task pool request for extra slots, and deci­sion tak­en by check­ing quo­ta and wait for slot allo­ca­tion. The sched­uler allo­cates the slots by round-robin algo­rithm. Prob­a­bilis­ti­cal­ly, a pool with high allo­ca­tion weight are addi­tion­al like­ly to be allot­ted with slots.
  • The sec­ond step is to allot slots to indi­vid­ual jobs. Every job is orga­nized with a para­me­ter of job pri­or­i­ty that may be a worth between zero and a thou­sand. The duty pri­or­i­ty and deficit are removed and mixed into a weight of the duty. Inside employ­ment pool, idle slots are allot­ted to the roles with the high­est weight.
    • Encryption/decryption

In this, data get encrypted/decrypted by exploita­tion the RSA                          encryption/decryption                  algo­rithm encryption/decryption algo­rithm uses pub­lic key & map, pri­vate key for the encryp­tion and deci­pher­ment of data. Here we have test­ed dif­fer­ent Encryption/Decryption algo­rithm and the per­for­mance in terms of CPU Uti­liza­tion, Stor­age , and Time is very good of RSA Algo­rithm as com­pared to oth­er algo­rithms and also pro­vides the in depth secu­ri­ty Con­sumer trans­fer the file in con­junc­tion with some secrete/public key so pri­vate key’s gen­er­at­ed & file get encrypt­ed. At the down­load time using the pub­lic key/private key pair expect­ed job decrypt­ed and down­loaded. Like client upload the file with the pub­lic key and also the file name that is used to come up with the dis­tinc­tive pri­vate key’s used for encrypt­ing the file. Dur­ing this approach uploaded file get encrypt­ed and store at main servers and so this file get split­ted by using the Fair4s Sched­ul­ing algo­rithm that pro­vides dis­tinc­tive secu­ri­ty fea­ture for cloud data. In an exceed­ing­ly reverse method of down­load­ing the data from

cloud servers, file name and pub­lic key wont to gen­er­ate secrete and com­bines

The all parts of file so data get decrypt­ed and down­loaded that ensures the tremen­dous quan­ti­ty of secu­ri­ty to cloud infor­ma­tion.

Fig.2 RSA encryption/decryption

  • Admin­is­tra­tion of client files(Third Par­ty Audi­tor)

This mod­ule pro­vides facil­i­ty for audit­ing all client files, as numer­ous activ­i­ties are done by client. Files Log records and got cre­at­ed and hold on Main Serv­er. for every reg­is­tered client Log record is get cre­at­ed that records the var­ied activ­i­ties like that oper­a­tions (upload/download) per­formed by client. Addi­tion­al­ly Log records keep track of your time and date at that var­ied activ­i­ties car­ried out by client.

For the secu­ri­ty and secu­ri­ty of the client data and con­joint­ly for the audit­ing func­tions the Log records helps. Addi­tion­al­ly for the Admin­is­tra­tor Log record facil­i­ty is pro­vid­ed that records the Log info of all the reg­is­tered clients. In order that Admin­is­tra­tor will con­trol over the all the info hold on Cloud servers. Admin­is­tra­tor will see client wise Log records that helps us to notice the fraud infor­ma­tion access if any fake user attempt to access the info hold on Cloud servers.

III.  RESULTS

Results of this under­ly­ing project will be explained well with the help of project work done on num­ber of clients and one Mas­ter serv­er and then five to ten Slave servers so then tak­en results bases on fol­low­ing para­me­ters tak­en into con­sid­er­a­tion like

  1. Time
  2. CPU Uti­liza­tion
  3. Stor­age Uti­liza­tion.

Our eval­u­a­tion exam­ines Pro­vid­ed RSA Encryption/Decryption algo­rithm pro­vides bet­ter per­for­mance on cloud infra­struc­ture as com­pared to the DES algo­rithm in Stor­age, Time, and CPU also get improved broad­ly. In this ways we have not only pro­vid­ed the Hadoop as a ser­vice through the cloud but we also tak­en into con­sid­er­a­tion the Secu­ri­ty aspect and pro­vid­ed secured hadoop as a ser­vice on infra clouds.

Upload­ing and Encrypt­ing Data Using RSA and DES Algo­rithm. Result are as below

Fig 3 Results Encryption/Decryption Algo­rithms.

Decrypt­ing and Down­load­ing Data Using RSA and DES Algo­rithm. Result are as below

Fig 4 Results Encryption/Decryption Algo­rithms.

CPU Uti­liza­tion Result are as below

Fig 5 CPU Uti­liza­tion RSA Vs DES

IV.  CONCLUSION

We have pro­posed secured Hadoop as a ser­vice cloud ser­vice that pro­vides need based of Hadoop ser­vice through infra­struc­ture cloud with opti­mized uti­liza­tion, oppor­tunis­tic pro­vi­sion­ing of cycles from idle nodes to dif­fer­ent process­es with resolv­ing the data secu­ri­ty issues through use of effi­cient encryption/decryption algo­rithm. But in instead of its promis­ing nature, not all com­pa­nies or pro­fes­sion­al orga­ni­za­tions are tech­ni­cal­ly sound or capa­ble of imple­ment­ing and main­tain­ing a suc­cess­ful Hadoop envi­ron­ment. Because of its dif­fi­cult nature to

man­age. As a result, we are pro­vid­ing the secured Hadoop ser­vice. Which keeps end user away from all these has­sles and keep pro­cess­ing and analysing big data through­out the indus­try Hence all unuti­lized nodes that remains idle are all get utilised because of hadoop map reduce nature and most­ly improve­ment in secu­ri­ty prob­lems and achieves load bal­anc­ing and quick process of huge data in less amount of your time. For file upload­ing and file down­load­ing; and opti­mizes the proces­sor uti­liza­tion and stor­age space use. Dur­ing this paper, we tend to addi­tion­al­ly plan a num­ber of the tech­niques that area unit imple­ment­ed to guard data and pro­pose design to pro­tect data in cloud. This mod­el was pro­posed to store data in cloud in encrypt­ed infor­ma­tion using RSA tech­nique that relies on encryp­tion and decryp­tion of data. Till cur­rent­ly in sev­er­al planned works, there’s Hadoop con­fig­u­ra­tion for cloud infra­struc­ture. How­ev­er still the cloud nodes remains idle.

Thus Secured Hadoop as a ser­vice on the infra­struc­ture clouds will pro­vide the way to process big data on cloud with pro­vid­ed secu­ri­ty to the data and along with the has­sle free plat­form which keeps away users from Hadoop con­fig­u­ra­tions and gives out Hadoop ser­vice as web appli­ca­tion ser­vice.

Con­fer­ence on High Per­for­mance Com­put­ing and Com­mu­ni­ca­tions & 2013 IEEE Inter­na­tion­al Con­fer­ence on Embed­ded and Ubiq­ui­tous Com­put­ing, 2013.

  • E. Deel­man, G. Singh, M. Livny, B. Ber­ri­man, and J. Good, “The cost of doing sci­ence on the cloud: the mon­tage exam­ple,” in Pro­ceed­ings of the 2008 ACM/IEEE con­fer­ence on Super­com­put­ing. IEEE Press,2008, pp. 1–12.
  • J. Szti­panovits, J. A. Stankovic, and D. E. Cor­man, “Indus­try Acad­e­my Col­lab­o­ra­tion in Cyber Phys­i­cal Sys­tems (CPS) Research,” CRA, Tech. Rep., 2009..

REFERENCES

  • I. Raicu, I. Fos­ter, and Y. Zhao. Many-Task Com­put­ing for Grids and Super­com­put­ers. In Many-Task Com­put­ing on Grids and Super­com­put­ers, 2008. MTAGS 2008. Work­shop on, pages 1–11, Nov.2008.
  • Open Crowd. Cloud Tax­on­o­my, Land­scape, Evo­lu­tion. http:/ / www. opencrowd.com/assets/images/views/views_cloud-taxlrg. png. June 8, 2010.
  • T. Zhang, W. Lin, Y. Wang, S. Deng, C. Shi, and L. Chen,“ The design of infor­ma­tion secu­ri­ty pro­tec­tion frame­work to sup­port smart grid,” in Con­fer­ence on Pow­er Sys­tem Tech­nol­o­gy, 2010, pp. 1 –
  • Nivyakant Sgraw­al et al., “Big Data and Cloud: Cur­rent State and Future Oppor­tu­ni­ties”, EDBT, pp 22–24, March 2011.
  • B. KRISHNA KISHORE UPPE NANAJI AND Y.SWAPNA

„„Improv­ing Uti­liza­tion of Infra­struc­ture Clouds, ISSN: 2278–7844

  • Rofrey Cean et al., “Sim­pli­fied data pro­cess­ing on Big clus­ters”, com­mu­ni­ca­tions of the acm, Vol S1, No. 1, pp.107–113, 2008 Jan­u­ary.
  • ENISA. Ben­e­fits, risks and rec­om­men­da­tions for infor­ma­tion secu­ri­ty. Euro­pean Net­work and Infor­ma­tion Secu­ri­ty Agency. Novem­ber 20, 2009.
  • D. Mavulya, J. Jan, R. Gand­hi, and P. Narasimhan, „an Analy­sis of Traces from a Mapre­duce Clus­ter,‟‟ in Proc. CGGRID, 2010, pp. 94–103.
  • K. Mean et al.,“ a flex­i­ble data pro­cess­ing tool”, In CDCM, Jan 2010.
  • T. Rtone­frak­er et al., “MapRe­duce and par­al­lel DBMSs: friends or foes?” In CACM. Jan 2010.
  • S. Das et al., “Ricar­do: Inte­grat­ing R and Hadoop”, In SIGMOD 2010.
  • J. Cohen et al.,“MAD Skills: New Analy­sis Prac­tices for Big Data”, In VLDB, 2009.
  • Srizhen Rang et al., “ SaaS-Based Cloud Com­put­ing in the Uni­ver­si­ty Research and Teach­ing Plat­form”, ISIEE, pp. 230–243, 2013.
  • Ama­zon­Web Ser­vices LLC. Ama­zon Sim­ple Stor­age Ser­vice. http:

//aws.amazon.com/s3/, 2009.

  • H. Tang, R. Zin, S.A. Brandt, E.L. Miller, D.D.E. Long, and T.T. Mclar­ty, „„File Sys­tem Ana­lyt­ics for­Large Scale Sci­en­tif­ic Com­put­ing Appli­ca­tions,‟‟ in Proc. MSST, 2004,
  • Yu, Jie, Guang­ming Liu, Wei Hu, Wen­rui Dong, and Wei­wei Zhang. “Mech­a­nisms of Opti­miz­ing MapRe­duce Frame­work on High Per­for­mance Com­put­er” 2013 IEEE 10th Inter­na­tion­al

Leave a Comment

error

Enjoy this blog? Please spread the word :)