Distribution, Data, Deployment: Software Architecture Convergence in Big Data Systems

Gorton, Ian; Klein, John
May 2015
IEEE Software;May2015, Vol. 32 Issue 3, p78
Academic Journal
Exponential data growth from the Internet, low-cost sensors, and high-fidelity instruments have fueled the development of advanced analytics operating on vast data repositories. These analytics bring business benefits ranging from Web content personalization to predictive maintenance of aircraft components. To construct the data repositories underpinning these systems, rapid innovation has occurred in distributed-data-management technologies, employing schemaless data models and relaxing consistency guarantees to satisfy scalability and availability requirements. These big data systems present many challenges to software architects. Distributed-software architecture quality attributes are tightly linked to both the data and deployment architectures. This causes a consolidation of concerns, and designs must be closely harmonized across these three architectures to satisfy quality requirements.


Related Articles

  • MOSAICKING MEXICO - THE BIG PICTURE OF BIG DATA. Hruby, F.; Melamed, S.; Ressl, R.; Stanley, D. // International Archives of the Photogrammetry, Remote Sensing & S;2016, Vol. 41 Issue B2, p407 

    The project presented in this article is to create a completely seamless and cloud-free mosaic of Mexico at a resolution of 5m, using approximately 4,500 RapidEye images. To complete this project in a timely manner and with limited operators, a number of processing architectures were required to...

  • Analysis and improvement of map-reduce data distribution in read mapping applications. Espinosa, A.; Hernandez, P.; Moure, J.; Protasio, J.; Ripoll, A. // Journal of Supercomputing;Dec2012, Vol. 62 Issue 3, p1305 

    The map-reduce paradigm has shown to be a simple and feasible way of filtering and analyzing large data sets in cloud and cluster systems. Algorithms designed for the paradigm must implement regular data distribution patterns so that appropriate use of resources is ensured. Good scalability and...

  • A Big Data Analysis Paradox. Nugent, Alan // Information-management.com;8/6/2014, p4 

    The author asserts that big data analysis is really about small data because small data is the product of big data analysis. He cites the limitation of systems, networks and software to address the scale and describes how the industry deals with the shortcomings by creating smaller, contextually...

  • What Next? Advances in Software-Driven Industries. Ebert, Christof; Hoefner, Gerd; V.S., Mani // IEEE Software;Jan2015, Vol. 32 Issue 1, p22 

    Software-driven industries are advancing in five dimensions: collaboration, comprehension, connectivity, cloud, and convergence. However, companies often can get stuck in an overly narrow technology focus. To avoid this, they should connect architecture and functionality, master the entire...

  • Research on Software Reliability Assessment with Optimum Reserved Strategy Genetic Programming. SUN Sheng-Juan; ZHAO Jing; CHEN Hua-Shan // Journal of Convergence Information Technology;Dec2012, Vol. 7 Issue 23, p317 

    The software failure data can be analyzed by Genetic Programming (GP for short). Meanwhile, the Particle Swarm Optimization (PSO for short) algorithm is taken to find the rational parameters during the dynamic modeling andfinally the optimized model structure of software reliability can be...

  • Semiparametric stochastic metafrontier efficiency of European manufacturing firms. Verschelde, Marijn; Dumont, Michel; Rayp, Glenn; Merlevede, Bruno // Journal of Productivity Analysis;Feb2016, Vol. 45 Issue 1, p53 

    In this paper a semiparametric stochastic metafrontier approach is used to obtain insight into the performance of manufacturing firms in Europe. We differ from standard TFP studies at the firm level as we simultaneously allow for inefficiency , noise and do not impose a functional form on the...

  • Bug Bash Kills Hadoop Bugs, Fixes Big Data Issues. Panettieri, Joe // Information-management.com;5/8/2015, p1 

    Seeking to improve the reliability and scalability of big data systems, more than 100 developers worldwide are finding and stamping out Hadoop bugs during today's Apache Hadoop Global Bug Bash.

  • Demystifying Big Data. Pavolotsky, John // Business Law Today;Nov2012, Vol. 21 Issue 21, p1 

    The article discusses legal issues related to handling "Big Data." It is informed that issues related to data handling are the sufficiency of intellectual property rights in the data and whether data invades a third party's privacy rights or violate a third party's publicity rights. The author...

  • An Enhanced k-means algorithm to improve the Efficiency Using Normal Distribution Data Points. Napoleon, D.; Lakshmi, P.Ganga // International Journal on Computer Science & Engineering;2010, p2409 

    Clustering is one of the unsupervised learning method in which a set of essentials is separated into uniform groups. The k-means method is one of the most widely used clustering techniques for various applications. This paper proposes a method for making the K-means algorithm more effective and...


Read the Article


Sorry, but this item is not currently available from your library.

Try another library?
Sign out of this library

Other Topics