Splunk can now store archived indexes on Hadoop. At the cost of performance, this offers a 75% reduction in storage costs without losing the ability to search the data. And with the new adapters, Hadoop tools such as Hive and Pig can process the Splunk-formatted data.
RebelLabs published their Developer Productivity Report, the result of a survey started in March 2015, where they polled the Java development community on Java performance and performance testing methods. To see how these numbers line up with a real world experience, InfoQ spoke with Kirk Pepperdine, CTO at JClarity and well-known performance expert.
In order to improve scalability, Parse moved part of their services, including their API, from Ruby on Rails to Go, Charity Majors, Engineer at Parse, recounts. In doing so, both their reliability and deployment times benefited greatly.
Owen Garrett, heads of products at Nginx, Inc., has described on Nginx’s blog which design decisions allow NGINX to provide top-in-class performance and scalability.
Teams rarely consider success as a mode of failure, but not preparing for exceeding their goals can be just as dangerous as ignoring basic software and infrastructure needs. Mark Simms and Mark Souza discuss anti-patterns they've seen and some of the best ways to architect to win in spite of your own success.
Google is making available to customers Cloud Bigtable, their own database used for more than a decade for services such as Search, GMail, Maps or YouTube. While they are not open sourcing Bigtable as they did with other products, the new cloud service is accessible through an open source interface, the Apache HBase 1.0.1 API.
It’s not unusual in financial service systems to have problems that requires significant vertical, as opposed to horizontal, scaling. During his talk at QCon London Peter Lawrey described the particular problems that occur when you scale a Java application beyond 32GB.
Network performance, virtualization and testing are some of the considerations to address performance and scalability issues with NoSQL databases. Alex Bordei wrote about scaling NoSQL databases and tips for increasing performance when using these data stores.
When a database administrators think of high performance data loading, they are thinking of bulk operations, a feature noticeably lacking in Entity Framework. But that doesn’t have to be the case. We recently spoke with Jonathan Magnan of ZZZ Projects about their new offerings.
High performance systems is about clean and representative models, the code doesn't have to be ugly, obscure and hard to read, Martin Thompson stated at the recent DDD Exchange conference in London.
Six years after Mono, Microsoft’s implementation of the CLR has finally gained support for SIMD via RyuJIT. Still in community preview, RyuJIT is the next generation JIT compiler for .NET.
Facebook, Google, LinkedIn, and Twitter have decided to make sure that a relational databases is “web-scale”, so they have put their efforts behind WebScaleSQL, a branch of MySQL 5.6 Community Edition.
More than anything else, architectural choices matter when designing a system with high scalability and availability. Using Azure customers as an example, Microsoft talks about the patterns and anti-patterns they see with their Azure customers and how it affects the four facets of system architecture.
New database developments indicate a return to SQL, but not by running the traditional relational stores on bigger and better hardware, not even on sharded architectures, but through NewSQL solutions.
Augmenting the roster of tribulations haranguing Obamacare and the healthcare.gov website comes a technical deep-dive performed by leading performance monitoring organization AppDynamics that paints a picture of a sophomoric development initiative used to build that site.