Mostrar el registro sencillo del ítem

dc.contributor.authorSouza Junior, Paulo Ricardo Rodrigues dept_BR
dc.contributor.authorMatteussi, Kassiano Josépt_BR
dc.contributor.authorVeith, Alexandre da Silvapt_BR
dc.contributor.authorZanchetta, Breno Fanchiottipt_BR
dc.contributor.authorLeithardt, Valderi Reis Quietinhopt_BR
dc.contributor.authorMurciego, Álvaro Lozanopt_BR
dc.contributor.authorFreitas, Edison Pignaton dept_BR
dc.contributor.authorAnjos, Julio Cesar Santos dospt_BR
dc.contributor.authorGeyer, Claudio Fernando Resinpt_BR
dc.date.accessioned2023-04-07T03:26:39Zpt_BR
dc.date.issued2020pt_BR
dc.identifier.issn2169-3536pt_BR
dc.identifier.urihttp://hdl.handle.net/10183/256835pt_BR
dc.description.abstractThe rapid growth of stream applications in financial markets, health care, education, social media, and sensor networks represents a remarkable milestone for data processing and analytic in recent years, leading to new challenges to handle Big Data in real-time. Traditionally, a single cloud infrastructure often holds the deployment of Stream Processing applications because it has extensive and adaptative virtual computing resources. Hence, data sources send data from distant and different locations of the cloud infrastructure, increasing the application latency. The cloud infrastructure may be geographically distributed and it requires to run a set of frameworks to handle communication. These frameworks often comprise a Message Queue System and a Stream Processing Framework. The frameworks explore Multi-Cloud deploying each service in a different cloud and communication via high latency network links. This creates challenges to meet real-time application requirements because the data streams have different and unpredictable latencies forcing cloud providers' communication systems to adjust to the environment changes continually. Previous works explore static micro-batch demonstrating its potential to overcome communication issues. This paper introduces BurstFlow, a tool for enhancing communication across data sources located at the edges of the Internet and Big Data Stream Processing applications located in cloud infrastructures. BurstFlow introduces a strategy for adjusting the micro-batch sizes dynamically according to the time required for communication and computation. BurstFlow also presents an adaptive data partition policy for distributing incoming streams across available machines by considering memory and CPU capacities. The experiments use a real-world multi-cloud deployment showing that BurstFlow can reduce the execution time up to 77% when compared to the state-of-the-art solutions, improving CPU efficiency by up to 49%.en
dc.format.mimetypeapplication/pdfpt_BR
dc.language.isoengpt_BR
dc.relation.ispartofIEEE Access. [Piscataway, NJ]. Vol. 8 (2020), p. 219124 - 219136pt_BR
dc.rightsOpen Accessen
dc.subjectProcessamento de dadospt_BR
dc.subjectStream processing applicationsen
dc.subjectBig datapt_BR
dc.subjectMulti clouden
dc.subjectMicro-batchesen
dc.subjectComputação em nuvempt_BR
dc.subjectData partitionen
dc.titleBoosting big data streaming applications in clouds with burstFlowpt_BR
dc.typeArtigo de periódicopt_BR
dc.identifier.nrb001135243pt_BR
dc.type.originEstrangeiropt_BR


Ficheros en el ítem

Thumbnail
   

Este ítem está licenciado en la Creative Commons License

Mostrar el registro sencillo del ítem