data visualization

← Back to Blog

Navigating the Data Seas: A Guide to Choosing the Right Data Visualization Tools for Your Business

In the ever-expanding realm of data analytics, the ability to effectively communicate insights is as crucial as the data itself. Choosing the right data visualization tools for your business can make the difference between drowning in information and riding the wave of actionable insights. In this guide, we will explore the key considerations to ensure you select tools that align with your business objectives and empower your team to unlock the true potential of your data.

Understand Your Data Needs and Objectives

Before diving into the world of data visualization tools, take a step back and clearly define your data needs and objectives. What kind of data does your business generate and analyze? Are you looking to track key performance indicators (KPIs), identify trends, or communicate complex relationships within your data? Understanding your needs will guide you in selecting tools that cater specifically to your business goals.

Consider User Skill Levels and Accessibility

 

Not all data visualization tools are created equal when it comes to user-friendliness. Consider the skill levels of your team members and choose tools that offer a balance between sophistication and ease of use. Opt for platforms that facilitate collaboration, allowing both data scientists and non-technical users to create and interpret visualizations effortlessly. Accessibility is key to ensuring that insights are democratized across your organization.

Scalability and Performance

As your business grows, so will your data. Ensure the chosen visualization tools can scale alongside your organization. Consider factors such as data volume, processing speed, and the ability to handle real-time data. Scalable tools not only accommodate your current needs but also future-proof your data infrastructure, saving you from the headache of migrating to a new platform as your business expands.

Integration Capabilities

Efficient data analysis rarely happens in isolation. Your data visualization tools should seamlessly integrate with other components of your tech stack, including databases, data warehouses, and third-party applications. This integration streamlines workflows, reduces manual data handling, and ensures a cohesive ecosystem where data can flow seamlessly across different tools.

Customization and Flexibility

No two businesses are identical, and your data visualization tools should reflect your unique requirements. Look for platforms that offer a high degree of customization and flexibility. This includes the ability to tailor visualizations to your branding, create custom dashboards, and adapt the tool to specific industry needs. A customizable solution ensures that your visualizations align with your company’s identity and workflow.

Security and Compliance

Data security is paramount, especially when dealing with sensitive business information. Choose data visualization tools that prioritize security features, such as encryption, access controls, and compliance with industry regulations. Whether you operate in education, healthcare, insurance, or any other sector, ensure that the tools you choose adhere to the necessary compliance standards to safeguard your data.

User Support and Training Resources

Even with the most intuitive tools, users may encounter challenges or need guidance on maximizing the platform’s capabilities. Look for data visualization tools that provide robust user support, including documentation, tutorials, and responsive customer service. Accessible training resources can empower your team to make the most of the tools at their disposal, accelerating the adoption and proficiency of data visualization within your organization.

 

Conclusion

Choosing the right data visualization tools for your business is a strategic decision that can significantly impact your ability to extract actionable insights from your data. By understanding your unique needs, considering user accessibility, ensuring scalability, prioritizing integration capabilities, allowing for customization, focusing on security and compliance, and embracing ongoing user support, you will be well-equipped to create data-driven success.

Read More
← Back to Blog

Functionality That Differentiates Informer 5

Entrinsik’s data analytics and BI platform, Informer, radically simplifies the process of accessing, cleansing, blending, and analyzing disparate data by using a single platform to create a cohesive, curated, governed data hub for self-service data analysis across organizations. Informer facilitates self-service reporting using Informer Datasets and Filtering. Informer minimizes self-service reporting errors because you are starting from a common baseline of curated data – a Dataset within Informer. The platform also facilitates collaboration which enables people to comment on reports, share insights, and more. There are several areas of functionality that our customers find highly beneficial and that differentiate Informer from other self-service reporting solutions.

  • Saved Filters. If you can look at a Dataset and write a filter – which you can do easily with Informer – you can save that filter to use when creating a report. The filter you create for the Dataset represents important criteria you’ve established for pulling in the data you want within your Dataset. Maybe you need “current transactions for today” or “transactions of employees that live in a certain state.” Such filters are available to all users that have access to the dataset. You can define very complicated filters for users and they simply add them with a couple of clicks.
  • Informer Discover. Informer’s Discover feature is ideal for self-service reporting. Often, users have a specific question to ask of their data. They log on to the reporting tool, get the answer, and they are done. However much of the time, users don’t have a finite question to ask; they want to see significant trends or groupings in the data but may not know where to start looking. Discover solves that by enabling you to click on specific columns like “state” and “order amount,” and Informer immediately displays charts relevant to those aggregations. Users then have an immediate visualization to interact with to help tell if those aggregations are significant.
     
    For example, let’s say you click on a state and get a distribution for all the orders for the last three days, and the numbers are the same per state. Maybe the geographic distribution for those orders is not significant, but maybe in your business plan you’re spending marketing dollars evenly across certain states per capita. You see sales are a lot better in one of those states and click on that state and see the numbers are skewed a lot more there. Well, you’ve just piqued your curiosity and start thinking this is something to investigate. You may even find it worthy of a report, all by simply clicking the button and immediately viewing the trend. People may not necessarily know what questions to ask when conducting data analysis. That’s not unusual with self-service reporting. However, the answers they start seeing with Discover may stimulate additional questions to ask.
     
    If you click on sales people in Discover, you may notice their sales are spiking up. From there you can further aggregate, or drill down into the data, even map the orders. Discover makes it easy for you because you can simply check a box and instantly see those aggregations.
  • Informer Collaboration. Another differentiator is Informer’s collaboration functionality. Maybe you want people to view your data, but you don’t want everyone looking at it because the information might not be ready for everyone to review. With Informer’s sharing capability, you can do that. For example, I may be a member within an Informer Team but I may not know anything about a particular Dataset. Maybe I started clicking some of the check boxes in Discover and think, “Wow, Bob’s sales are through the roof so I’m going to create an Informer report and keep it within my Team. I’m also going to send a comment to the Team and say, ‘this needs further exploring, but my database expertise isn’t as good as some of you. Can you look at this with me and see if there’s a way to tweak this to provide an answer for why Bob is doing so good?’”.
  • Single Source of Truth. Given Entrinsik’s many years of designing and supporting BI and reporting software, we found that self-service reporting is both a blessing and a curse. For example, someone might start thinking they’re getting solid answers on a database query they’ve written to answer a certain question. Other people might be running different queries to try to get the answer to the same question. Ironically, quite often the answers differ. Who is right? Maybe the reason the results differ is because in one person’s query, practice orders and refunds are eliminated from the set of current orders; one user knows about those nuances, whereas another might not. Having these two sets of numbers out there, and carrying equal weight, can lead to inconsistencies in the overall analysis of company data.
     
    Informer solves this problem by putting the power in your hands to curate data, so you can make sure people see what they need to see. When your Informer Dataset is curated by someone who understands its value of providing a single source of truth, you eliminate a lot of situations like, “My numbers didn’t come out the same as Bob’s numbers; why is that? I just spent countless hours trying to figure out why that is.”  Unproductive time and money wasted on people spending time trying to reconcile different answers is eliminated with Informer’s curated Datasets. With Informer Datasets, you’ve now got a single source of truth where you can say, “I’m going to make a Dataset of current orders and this is it, and anyone who wants to report on orders uses this Dataset”. That delivers ROI to your organization.
     
    Another reason a curated Dataset is preferred is that no database table that I’ve ever seen has perfect data in it.  There are always transactions that are entered wrong and there are always rows that don’t belong — they still exist because in the past, we ignored them. People will also tell you they take data from a reporting tool, dump it into Excel, and within Excel, they eliminate the rows that are wrong. There is a better way!  The Informer Dataset can be “that Excel place” where you toss out the junk. One of the common uses of Informer is for people to find bad data in a database and cleaning it up.
     
    Here’s a hypothetical example:  I’m the one who knows how to create Informer Datasets in our organization and I’m going to create a Dataset based on how people are using Excel spreadsheets. I talk to people and we all agree this is what the current orders and corresponding Dataset needs to be. So, I write my query for current orders and I’ve also been told by Joan in Accounting “You have to eliminate all the orders with ‘Type X’ because those aren’t real orders”. So, I add within my Dataset the following criteria: ‘take out type X’. And she also says take out all the orders that have negative values because those are also junk. So, I add that to my criteria.  And maybe there are a combination of fields that can’t be done with a query, however it can be done with an Informer Data Flow step.
  • Data Flow. A Data Flow is simply the stream of data from an initial source, often a database query, into the final Informer Dataset. A flow step is an operation that a designer can add to the Data Flow. A flow step can restructure data, add to the data, calculate on the data, and it can also clean the data. Using flow steps, you can go through each row and say, ‘if this is true for this row, toss it out of the Dataset’. So, you can toss out certain records that just don’t belong. You can run calculations across columns that may span across multiple data sources. You can run multiple pass calculations. You can create scores based on multiple factors in a row. The possibilities are limitless.

 
In conclusion, with Informer, you have oversight and governance of the data, because the curator constructed the Dataset and the rules (Filtering and Data Flows) which were agreed upon. The result is, you have the “golden standard” of that data for everyone conducting self-service reporting. This maybe an ongoing process but people don’t have to worry about what’s good and what’s bad because the curation process takes care of that.

For more please contact sales@entrinsik.com or call 888-703-0016.

This article was written by Andrew Morovati
Informer Chief Solutions Architect

Read More
← Back to Blog

True Self-Service Reporting

Self-service reporting has long been a goal of the IT departments I consult with. They long for the ability to pass along tasks involving query writing and data retrieval to the end-user. Basically, they want their user base to “write their own reports.” There are several solutions on the market today that attempt to provide such relief.

Most BI solutions or reporting tools within enterprises are connected to a database which users can query. To make this work for the user base, the IT department creates a modeling layer which represents certain subsets of tables and subsets of columns within those tables. In the most basic use cases, IT sets up security credentials for users and then the end-user logs in, runs an SQL query, and gets back a grid of data to work on.

This process seems straightforward: ask the tool for the data you want, and you get it back. However, the data modeling has to be clear enough to the users so they know which columns in the model represent which data. The modeling scheme can be simplified by making field names more understandable. That’s fine for a simple “Person” table containing trivial columns (first name, last name, city, state) and the query is “People that live in PA.” The user just clicks on “state”, selects equals, types in “PA”, selects the output columns, and runs the report. Some tools handle self-service reporting in this simplistic way. Even more sophisticated tools that offer visualization and other reporting options still rely on the end-user for knowing where the data is and how it’s structured.

Running More Complex Queries

Once the query gets more complex, the user is going to require more knowledge, or “domain knowledge”. Suppose the query is for specific types of transactions run by specific types of persons – let’s say sales transactions run by sales reps who also have person records. This query hops among three tables, and the hops themselves may be indirect. We may not be hopping from “Sales Transactions” to “Sales Reps” to “Person.” The query may go from “Transactions” to “Sales Rep Transaction Middle Table” to “Sales Rep” to “Sales Rep Middle Table” to “Person.” For their report to be correct, users will need to know that they’ve got to make these hops. Although this is a rather trivial situation for a database query, the user will need a certain amount of domain knowledge to create the report.

Accessing a broader set of data is going to be intimidating for a lot of users. Fortunately, the ideal self-service reporting solutions remove the requirement of knowing how tables and models interact. These solutions simply enable you to say, ‘I want to find out everyone from PA’. So, the question is, how can you empower your users with a truly self-service solution that makes it easy for them to do all sorts of analysis and reporting?

Entrinsik offers an answer. With Entrinsik’s Informer you can create Informer Datasets — a curated, single-source of truth that enables you to bring in data from each of your different Data Sources. A data domain expert would have authorization to create a very large Dataset that might include transactions, sales rep names, sales rep addresses, last year’s sales, person information, etc. The Dataset could easily have fifty columns and millions of rows.

Informer provides for very easy filtering on this data by showing you what columns are filtered to support your reporting and analytics need. For example, you can easily look at a column heading and say, ‘this state column is what I want to filter on to create a report’. The Dataset will quickly provide a list of states and you simply select the states you want. You don’t even know that the process involved hopping over several tables, or perhaps gathering data from an entirely different database, or even a calculation derived from other columns. All you know is the data is there.

The Open Query Challenge

Among the challenges associated with traditional self-service reporting is the open query problem. When people are permitted to run self-service database queries, they are often inclined to choose the columns they want first. Then they run a wide open, unfiltered database query – maybe during the day when production and sales transactions are going on – and the system pulls out an enormous amount of data. They scan the first page and then iterate over this until they figure out how to get the query to shrink the set of data to the scope they want. Often these queries are huge, and your database gets hit repeatedly, which results in the system slowing down due to a lot of unnecessary CPU activity on the database side. That doesn’t need to happen if the data can be offloaded into a staging solution.

As an IT strategy, data warehousing solved this issue by redirecting people away from hitting the production database. With this strategy, either a manual or automatic infrastructure extracts data from the production database and copies it onto another database which individuals can access for reporting purposes. However, the skillsets necessary to design the ETL processes to populate the data warehouse are typically rare. In many cases the users with the domain knowledge of the production database need to collaborate with IT staff in order to create these warehouses. In addition, giving individual network users access to several data marts or warehouses creates a distributed security problem. For example, controlling “who can access what” that’s stored in a distributed way across the data landscape.

Informer’s approach uses an Informer Dataset as a barrier between the user and the production database. With Datasets, designers don’t need to know sophisticated ETL techniques, and users don’t need to know a lot to produce sophisticated reporting. You won’t run the risk of users interfering with the production database because Informer handles that load for them. Since these users are acting on the Datasets only, Informer acts as a buffer between the user and data source. And Informer’s security model ensures “who can access what” is located centrally.

Reporting Consistency

Another issue with self-service reporting is consistency among analytics. Suppose Leif runs his report for “Sales figures for PA sales reps,” and Isabel runs her “PA sales by salesperson” report. Often, their row counts, and subsequent aggregations (like “total sales”), will differ. This can be a problem, no matter whether they are given access to report directly against the database or against the data warehouse. Informer solves this with its Dataset. Once the concept of “sales transactions” can be defined, and potentially complicated filtering is agreed upon (e.g. “Exclude intra-division transactions”), Leif and Isabel can use this Dataset to compile their reports.

We have said that a data domain expert can author an Informer Dataset. Where does the data domain expert come from? Typically, larger organizations employ data domain experts — people who understand the data source schema inside and out. These individuals can construct broad Datasets (for example “Sales Transactions”) that maybe used for self-service reporting. However, even at smaller organizations, lots of people utilize database queries or Excel spreadsheets in their work. And typically, there are users who compile the data for others to use. In a scenario like “Bob always uses this spreadsheet that he reviews and has final say on, and then he distributes it for everyone to use” — clearly the person to be creating the Informer Dataset is Bob. Instead of Bob cutting, pasting, positioning and duplicating values from one data source into the spreadsheets, he will have a much easier time designing an Informer Dataset. Moreover, that Dataset can be secured in Informer so only the team of users that need it can see it. This avoids security concerns of passing around unsecured spreadsheets.

One interesting Entrinsik Informer use case comes from a higher education customer. Management wanted to compare, per cohort, the number of students that were present for their second fall semester, second spring semester, their third, etc. This was handled by a manager working with multiple Excel spreadsheets. At a certain time during each semester she saved these spreadsheets as a snapshot in time for reporting purposes. Then, she looked at each of these spreadsheets for each student to find their status at that time and added the information into certain cells in a different spreadsheet. It took her weeks to do this, but she needed to accomplish this because school management wanted to analyze the information.

Informer was able to help her by uploading the spreadsheets into separate Informer Datasets. Then, a summary Dataset was created that linked to each one of the individual Datasets. All she needs to do is simply upload the new census data and other data she gathers into the separate Datasets and then Informer creates a summary Dataset for her. She doesn’t have to do any manual work. She just runs the Dataset and it all shows up for her. Because she had a lot of experience working with this information, she knew what data she needed, and she knew how to get to the data. However, she didn’t know any way to automate it. Now she owns that Dataset for everyone to use, and they do a lot more analysis now than just the one thing they were doing before.

Data Latency Concerns

What about data latency? What’s an acceptable amount of data latency between the production database and the reporting database for your users? How do you solve for when the sales folks say, for example, they need up-to-date information now and they can’t wait for the reporting database to be updated overnight? How much transpired time is acceptable for them to pose questions on the current data and still have valid answers based on when the query ran — one hour, six hours, one time a day? The answer depends on a case-by-case basis but fortunately Informer’s Dataset can refresh on a schedule that you establish based on what is convenient for your operations. So, if users need to have the data refreshed each hour, Informer can do that.

Keep in mind that when you’re working with trends, you don’t need 100% up-to-the-minute transactions. That’s not what trends are about. They are about broad strokes. However, in the case where someone does need to know the latest information “right now.” Informer has an ad hoc query facility. That doesn’t solve the problem regarding whether the user knows how to run a query, so only certain security roles can run these.

A Solution That Works for Everyone

An ideal solution provides for two types of updates: refresh only the information that has changed since the last scheduled update and update all records. Informer supports both. Datasets can also be maintained via Informer’s REST API. Rows can be pushed into the Dataset, or the Dataset may be refreshed upon a certain event occurring.

With Entrinsik Informer Datasets, you can provide true self-service reporting to all of your employees and enable them to find new and more productive ways to improve operations and achieve your goals.

For more please contact sales@entrinsik.com or call 888-703-0016.

This article was written by Andrew Morovati
Informer Chief Solutions Architect

Read More