I recently needed to use PowerShell to call a T-SQL script using Invoke-Sqlcmd with the -Variable parameter. In my case, I wanted to specify the locations for backup files and where to restore the data and log files. Rather than add the conditions into the SQL script, I wanted to make the SQL script as generic as possible, and use SQLCMD variables for the different scenarios. In this blog, I’ll show how to use the array sub-expression operator and “splatting” […]
Talks on Managing Data and Applications Anywhere
In my blog post Using Power BI to analyze Windows Perfmon data for SQL Server, I introduced a concept that I call “business-critical transactions” (BCT) to understand the overall throughput of your SQL Server applications. When I worked on the SQL Server team, the SQL Performance Engineering team used a set of customer applications with server-side replay scripts that would get executed before development teams checked in their code. If a development team’s build showed significant regression in performance, they could […]
In this second blog post for my series, Power BI for the SQL Server DBA, I’m going to show how you can take Windows Perfmon (Performance Monitor) data for SQL Server and analyze it with Power BI. You can find the source code, Power BI reports, and other related files on my GitHub site at https://github.com/db-best-technologies/power-bi-for-sql-dbas. Here is a summary of what’s coming up in the series. It is subject to change based on feedback and reader demand: Capturing performance […]
Over the last few months of benchmarking SQL Server performance on AWS, I discovered a way to optimize SQL Server performance running on EC2 instances by paying close attention to several key Windows performance counters. This blog post is the first in my new series called Power BI for the SQL Server DBA. I’m going to show how you can easily capture performance counters to optimize SQL Server performance and cost on AWS EC2 instances. I’m going to start with […]
Managers for a well-known accounting and tax-preparation company contacted DB Best looking for an alternative solution to a big problem. They were long term customers of SAS but were now facing two key challenges: high SAS license costs and the fact that SAS developers could not keep up with business demands and provide for effective support. The DB Best team analyzed our customers’ initial SAS solution and determined that the best possible option was to migrate their workloads to Microsoft […]
Our fully automated Migration Platform confidently guides you through an upgrade Microsoft SQL Server, Oracle Database, and PostgreSQL database platforms. Read the blog post and watch the related video to discover the seamless database upgrade powered by DB Best Migration Platform.
Last year, we published the blog post on Validating AWS EC2 SQL Server Deployments Using Benchmark Tools. Since then, AWS released enhancements like Optimizing CPU Option that allow you to choose a larger EC2 instance for your SQL Server 2017 instances for greater Memory and IOPS / vCPU to keep your licensed vCPU count down.
Read this bog post to discover how benchmarking your SQL Server workloads on AWS in a pre-production environment can help you pick the best candidate EC2 instance with an optimized storage configuration based on real data.
In Oracle, you can specify virtual columns in a table definition. However, when migrating your Oracle database to Microsoft SQL Server, you need to utilize computed columns. You may want to use SSMA to automate the conversion. The problem is that SSMA does not correctly convert source Oracle virtual columns to SQL Server. So, you will need to update the converted code manually.
Read this blog post to discover our proven solution to Oracle virtual columns conversion to Microsoft SQL Server.
One of the customers contacted DB Best to help address performance issues with their SQL Server application. They started experiencing problems after upgrading from SQL Server 2008 R2 up to SQL Server 2016. However, after we analyzed their database system, we discovered a complex issue with updating SQL Server statistics.
We utilized industry best practices and internal expertise to correct issues. Read our blog post to learn how we approached this complex task.
What do you do when your data science team needs to perform read-only data warehouse queries against your mission-critical online transaction and batch processing database? Our customer Getty Images had exactly this problem with their production AWS deployment of SQL Server 2014 AlwaysOn on EC2 instances. Their data science team was performing read-only queries over 4 terabytes of historical data. These queries are ad-hoc and look at multiple months of historical data for analysis. The queries would consume all available […]