In 1948 Claude Shannon published his Mathematical theory of communication. This paper is cited as the beginning of what we now call information theory, the scientific study of the quantification, storage, and communication of information.
Dr. Shannon's research leads to a handful of axioms, two of which are as follows:
- An event with probability 100% is perfectly unsurprising and yields no information.
- The less probable an event is, the more surprising it is and the more information it yields.
Put those together and we have the concept "information is surprise". This topic came up during a recent podcast and has been on my mind ever since. The world today is an endless stream of data transmitted and received, most of which is what we call "noise". And we often talk about how to get a signal through the noise.
To me, the signal has meant action. For example, I don't want to receive an alert unless some action is required on my part. If no action is needed, don't send an alert, just log the data for review later.
When I start thinking about information, and surprise, it alters my views on data collection, curation, and consumption.
I see the current world, where every piece of data, every possible metric, is collected and analyzed. Everyone hoping to find gold with data mining, or find a signal through the noise using machine learning.
I also see a new world, where only information is collected and analyzed.
Both worlds are looking for insights, or answers to questions regarding root cause and such. But the current world results in the common scenario of providing an answer we already knew. And if we already know the answer, then there's no surprise, and thus no information exchanged. Just noise. And useless pie charts.
The new world, built only upon information, and surprise, leads to better answers. The new world leads to information being exchanged.
That's the end goal we should all work towards.
Today's Sponsor

Someone is complaining about the database, again. Whatever the bottleneck may be, you know it is not your fault, but it is your responsibility.
Here's a secret for you. The majority of database issues are not edge case scenarios requiring hiring a team of experts to solve.
Common root causes include locking and blocking, poor indexing, or a badly written query. SolarWinds provides a comprehensive suite of cross-platform database performance monitoring and tuning tools for both Earthed and Cloud workloads. And with the recent purchase of Sentry One, makers of SQL Sentry, we can also help with the edge cases, too.
If you haven't tried our products, there's no better time to get started.
Community Links

Azure analytics end-to-end
Good resource for understanding the Azure analytics architecture and all the services available complete with links. I often find myself having to remind others how an entire world exists outside of the database engine, and this image is a good start.
Raw Data Podcast
Arun Ulag
Who better to talk about the past, present, and future of BI than Microsoft Corporate Vice President of the Business Intelligence Platform Arun Ulag. I was surprised to learn how Arun and I never crossed paths before, considering we interacted with the same people for a few years.
Events

We are moving forward with plans to host Live! 360 this November in Orlando.
Live! 360 brings the IT, Developer, and Data communities together for six days of training, knowledge sharing, and networking. With unlimited access to Live! 360’s five co-located events, you and your team will get the training you need to keep you and your business competitive and future-ready.
The call for speakers is open through May. As co-chair for SQL Server Live! I encourage everyone to submit a session to speak.
Send any questions about the event to me at SQLRockstar@thomaslarock.com
Azure FastTrack – Stories from the Field: HammerDB Vs. Azure SQL
The Western New England Azure Data Group has it's next meeting on 13-May. Join Azure FastTrack CXP Product Group team members Bradley Ball and Josh Luedeman review how they use HammerDB to help customers set a baseline for their Azure environments.
Data Janitor Roundup
Event source start options in Azure Time Series Insights Gen2
Wonderful feature, allowing you to specify how much pre-existing data you want ingested when connecting to the event source for the first time. Most of the example scenarios I have touched do not use existing event source. In retrospect, that doesn't seem real-world, unless you are building something from scratch.
Microsoft Data Loss Prevention Alerts Dashboard Now Generally Available
Microsoft has commercially released its Data Loss Prevention Alerts Dashboard, which is used for viewing data loss prevention (DLP) policy violations in organizations. This sounds great, but it isn't magic. You still need to go through the process of identifying and labeling your data as sensitive in some way. I'm hopeful this service get's extended to check inside of a database, too.
AQUA (Advanced Query Accelerator) – A Speed Boost for Your Amazon Redshift Queries
With advances in storage performance outpacing those in CPU performance, your performance bottleneck gradually drifts to network and CPU bandwidth. AWS has decided the best approach to this problem is to create the elusive "turbo" button by putting a cache layer next to the storage layer.
Azure Database Migration Guides
Database migrations often involve multiple phases, steps, and tools based on various scenarios and workload requirements. This page helps break down 28 different migration scenarios. So the next time someone asks you to migrate from Access to SQL Server, you will find the necessary steps here.
Sponsor an Issue
Reach thousands of data professionals who care about data, databases, and helping others make their data the best version possible. Get in touch right here.
Tweet of the Week