SQL Server has been released to manufacturing with general availability set for April 1st. Through this week we’ll be covering various aspects of the new release starting with data warehousing.
Indexes in SQL Server’s In-Memory OLTP don’t work exactly like normal indexes. This isn’t necessarily a bad thing, but the differences need to be kept in mind to avoid performance problems.
Opserver is an open source monitoring solution, released by StackExchange, of StackOverflow's fame. Opserver provides a quick overall view of each monitored system's health, while allowing the user to deep dive using a drill-down approach. InfoQ talked with Nick Crave, one of Opserver’s creators, for additional insight.
One of the biggest challenges when researching a new technology is determining where to start. A typical SQL Server installation could easily have hundreds of tables. Examining each one by hand to determine which would benefit from conversion, is a daunting challenge. This is where the AMR Tool comes into play.
In this report we look at the internals of SQL Server’s In-Memory OLTP to see how it uses timestamp-like transaction ids in lieu of locks.
SQL Server 2014 will offer Clustered Columnstore Indexes. These will offer the performance and compression benefits of column-oriented storage without the need to restrict the underlying table to read-only access.
Originally this report was titled “Natively Compiled Queries”, but that doesn’t do justice to how deep this runs. When a memory optimized table is created, SQL Server will create a DLL specifically for that table. All data access for the table, including indexes, occurs through this DLL.
SQL Server 2014’s Memory Optimized Tables handle indexes very differently than traditional tables. First and foremost, you must have at least one index and cannot have more than eight indexes. Only the primary key can be marked as unique and don’t even think about foreign keys or filtered indexes.
In SQL Server 2014 Microsoft will be unveiling its lock-free technology known as Memory Optimized Tables. Using a new storage and query subsystem, these represent a radical departure from traditional database design.
Microsoft recently updated database lifecycle management guidance page with SQL Server DLM diagram along with curated resources for SQL Server Data Tools, SQL Server Management Studio and Windows Azure SQL Database.
Microsoft updated several tools for developers targeting SQL Server, SharePoint, and Office. The Business Intelligence tools support analysis, integration, and reporting for SQL Server. The Office/SharePoint tools suite bring support for the 2013 versions of each and .NET Framework 4.5.
Brent Ozar, renowned SQL Server consultant recently shared few techniques for improving productivity with SQL Server.
Few months back, Microsoft announced HDInsight, Microsoft’s Hadoop distribution for managing, analysing and making sense out of large volumes of data. InfoQ connected with Val Fontama, Senior Product Marketing Manager for SQL Server, to know more about how the Enterprise Big Data @ Microsoft story is panning out.
Microsoft recently released several enhancements to Windows Azure Management Portal with support for 6 additional languages, database improvements and enhancements for virtual machines and networking including support for operation logs.
During the PASS Summit 2012, a technical conference for SQL Server professionals, Microsoft announced Hekaton, an in-memory row-based data management system targeted at transaction processing (TP) workloads. Besides the advertised increase in TP speeds of up to 10x for old applications and up to 50x for new optimized ones, Microsoft touts Hekaton as being fully integrated into SQL Server.