Our Insights.

Covering strategic and technical aspects of analytics.

By subscribing, you agree to receive a monthly newsletter from 173Tech.

Maximise Return On Data, The SCALE Principle.

February 23, 2021 | By Candice Ren.

By applying modern analytics, we successfully help companies grow efficiently at all stages, from MVP all the way to over $100 million run rate and unicorn status.

In this article, we would like to share with you the common data pains we witnessed in many companies and how our SCALE principle helps overcome these and maximise return from your data.

Common Data Pain Points

Messy and inconsistent data. To answer a simple question, you have to dig through many sources and each source reports numbers in a different style. After numerous copy-pasting and manipulating with Excel formulas, you realise that things do not add up.

A typical scenario, Facebook claims its campaigns generated £5,000 sales last month and Google another £5,000. However, your total monthly sales was £9,000. You then check internal databases for channel attribution. A completely different story again. 30% of sales came from organic sources.

What can you trust and how do you make decisions on next month's marketing budget?

Data silos. Numbers from different aspects of your business are scattered around in various sources, gathered in different frequencies, and in some instances, being held hostage by third party tools.

If you are an eCommerce business, you might receive fulfilment costs from your 3PL on a monthly basis, product sales from Shopify daily, and digital marketing spend weekly from your agency, while cost of goods sold is stored in an Excel spreadsheet and updated as required.

How do you keep track of profit margins on SKU level and decide on which one(s) to scale? How would you evaluate it across demographics?

Error-prone & time-consuming manual processes. To answer the profitability question above, you need to log into each source, query the data for the relevant time period, copy and paste into a spreadsheet, then create a master sheet with a range of manually entered formulas applied to various chunks of data. This process is lengthy and difficult to debug with human mistakes possible at every step of the way.

As you scale, this setup will soon become impossible to maintain. We have seen spreadsheets that take minutes to open and even longer to respond to a single change. Various versions of the original spreadsheet are then created to cater for slightly different scenarios, introducing more errors. Additional errors reduce the reliability of your data and result in more time spent on rechecks.

If you cannot trust your numbers, you cannot build a data-driven culture and make data-informed decisions.

Unable to respond to urgent requests. Your app is experiencing a steep drop in user activities since last week. This could be due to a wide range of possible factors: new campaigns bringing in low quality users, bugs in the latest release, new features users do not understand, or server error preventing push notifications. It could be specific to a country or platform or other dimensions. It requires data interrogations from multiple angles.

Your team starts frantically pulling data from various places to pinpoint the issue, while precious time is being lost and your customers continue to churn.

Disconnection between data and business. Often when checking a dashboard from one of your service providers, you are left with the feeling that it was not very helpful. They contain a number of nice-looking charts and total figures but fail to tell you anything you did not already know.

This is expected due to a number of reasons:

  1. These charts typically visualise simple aggregations (e.g. sum, average or count) over generic metrics (e.g. sales or number of customers), without deep reflections on the true health of your growth.
  2. Data is not processed and summarised into meaningful insights based on your unique business model. It is designed as a one-size-fits-all solution and easy to switch on for everyone.

The main issue with generic and out-of-the-box solutions is that sooner rather than later you will grow beyond the box.

The SCALE Principle

We formulated our best practice data strategies into the five components of the SCALE principle. It ensures maximum insight generation from your data while removing all your data pains.

SCALE Principle Analytics


The first step towards a world-class analytics stack is to standardise business metrics and apply best practices over your entire analytics processes.

Data Dictionary

To standardise business metrics, start with a list of KPIs core to your success and their definitions. Take monthly revenue for example, it is an important metric with many caveats. Does it include taxes, refunds, or deferred revenue?

Create a data dictionary for all your KPIs. Ensure that everyone in your company is well-versed in its contents, and that it is constantly updated as your business and analytics stack evolves. A consistent data vocabulary allows for clearer communications and goal setting between different teams.

Add technical definitions to each item in your data dictionary. This bridges the gap between business and data. It also provides the technical translation for the automation step later.

Analytics Processes

First, select the right tool stack tailored to your usage needs and existing tech ecosystem. Your tool stack should cover automated data delivery, data warehousing, data modelling, reporting visualisation, and version control. Next, formalise your implementation processes covering requirements gathering, prioritisation, development, peer review, and production release.


This is where all your data sources come together. Data from all customer touch points (marketing, sales, customer support, CRM, fulfilment, costs etc.) is gathered in one centralised location whilst retaining all the information you care about.

The key here is to select the right tools for your data infrastructure for optimum performance and scalability. For your data warehouse, some options include Snowflake, Redshift or BigQuery, depending on your existing tech ecosystem and intended usage. If you have large volumes of data, you might also want to have a data lake to store raw data, using tools like Hadoop or Amazon S3 buckets.

Another important aspect of centralisation is your code base and reporting. Ideally all your analytics scripts (e.g. SQL, Python tasks) should be easily accessible from one location. We also recommend a single visualisation platform for all your reporting dashboards to avoid disjointed insights.


Once you have defined the data sources to collate and their destination, you should consider automating the following aspects of your pipeline:

  • Data extraction
  • Data modelling
  • Data visualisation

Data extraction ensures that information is delivered consistently from all sources at your desired time interval. Some tool options include Stitch or Fivetran. Consider it your logistic guy who picks up the data package every morning and deposits it into your data warehouse.

For all the data packages arriving in your data warehouse, perform data modelling for two key purposes:

  1. Ensure all sources are linked via unique identifiers to create your own single customer view.
  2. Apply the business logic defined in your data dictionary and transform raw data into meaningful KPIs.

We are the creators and maintainers of the open source data processing and modelling framework SAYN. It covers many task types including Python and SQL transformations and helps analytics teams improve data engineering efficiency by easily orchestrating and automating data processes.

Once created, you can utilise the data models to create dashboards that clearly visualise your KPIs. Best practice is to have a top level dashboard that summarises key trends and quickly unveils opportunities or issues in your business. Then design a dashboard per business vertical (e.g. marketing performance dashboard, finance P&L dashboard etc). Metabase, a free tool, is a good option as you start your data journey. As your team and data capabilities scale, consider moving to a more robust solution such as Looker.


Now with your own centralised data gold mine, you can start learning from it.

With an efficient data pipeline, you can train data science models to segment, predict and influence customer behaviours. Customer lifetime value (CLTV) predictions, churn propensity scores, recommender systems, automated consumer sentiment with natural language processing (NLP), your power here is unlimited. These models can be integrated into other parts of your products and services to create unique competitive advantages. It creates a constant and dynamic learning loop from observing user patterns, creating algorithms, feeding it into your product development and observing new feature usage.

Another element of the learning loop is user testing. To encourage a desired behaviour on your app, your team came up with a number of ideas to achieve it. How do you know which one will be the most effective? Run an experiment and test these options against a control group. Test results should be modelled, automated, and visualised in a dedicated testing dashboard. This will allow you to capitalise on the winning variant early and stop any poor performing test promptly.

Make sure you have a process where all learnings feed into your internal knowledge base and are shared across all teams.


In our experience, we have seen many companies succeed through leveraging their data and integrating continuous learning and testing into their agile product development. An efficient and democratised data stack empowers all teams and individuals. It is a game changer for companies embracing it.

For your data science and analytics team, data efficiency and reliability ensure little time is wasted on digging and cleansing data. Instead, they can focus on mining deep product and behavioural insights, and building state-of-the-art algorithms.

For product and marketing teams, a well-structured data warehouse and user-friendly visualisation tool enable everyone to create ad-hoc reporting tailored to their changing needs. It provides timely feedback on current projects and efforts, so one can pivot and adapt quickly as new data insights stream in.

So make sure to train all teams on your chosen visualisation tool. One efficient way is to appoint a data champion per team, who will act as the data power user.

Whether you are looking to set up analytics from scratch or upgrade your existing infrastructure, I hope this article provides you with a structured and easy to follow plan to build your world-class data stack. If you have any questions or feedback, please do reach out to us. We are always happy to chat!

Subscribe to our monthly newsletter for more insights.

By subscribing, you agree to receive a monthly newsletter from 173Tech.

Get In Touch

Send us a quick message about your data challenges, we are always happy to chat!