Skip to content

GuerrillaAnalytics/proj001_lfb

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

45 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

proj001_lfb

Example training project for Guerrilla Analytics ways of working.

Getting started

Please refer to the project wiki to install and configure the necessary software to complete this training.

Import first data file 002_import_data

  • see the setup folder

    • script for creation of input and wp workproduct schemas. These are the basic structure of the data environment
  • see L0 scripts folder for data import code.

    • target table is created
    • load script pushes the file from the inputs folder into the target table.
    • all data is loaded into the inputs namespace in the data environment with the same namespace as the inputs folder structure.
    • input file names are maintained in database table names as much as is practical

Create first workproduct 003_first_workproduct

  • see the first workproduct under the wp folder
    • the workproduct has an ID and this ID is used in workproduct filenames. This helps with traceability
    • the workproduct code is pointing at the raw input data loaded in the previous step. This means the workproduct code potentially needs to do significant cleaning work before using the data.

Checkpoint

At this point, we have already come quite far in terms of data provenance and better management of our project.

  1. All input data is stored in one place with a simple convention that identifies the data source and the version of the source.
  2. All input data is loaded into a well defined place in a database (the inputs schema). 3. There is not need to document this because the same namespace and files names are used in the inputs folder and in the database.
  3. Workproducts point at a single source of the truth that is never altered so we have reduced the risk of inconsistency in interpreting the data.

However, some challenges remain.

  1. Every workproduct needs to clean its own data. This is inefficient and it risks inconsistency in the application of cleaning rules and other business rules.
  2. If new data arrives, references to raw datasets have to be changed in many places in the code.

Clean data 004_clean_data

  • See the L1 folder in the pipeline.
    • We are now building the pipeline into a versioned namespace proj001_lfb_0_0_1. This allows us to keep backwards compatibility with older pipelines and make sure that older work products continue to work.
    • a teardown script clears out this namespace at the start of the pipeline build. This makes sure old orphaned datasets are not persisted and accidentally used by workproducts.
    • the L1 script points at the latest raw data. This is useful when data is broken or gets refreshed and we want to be clear that we are pointing at the latest data.
  • See the L2 code folder
    • The cleaning code gives the dataset a sensible cleaned name used by the rest of the pipeline. This is important because it allows us to decouple the downstream pipeline code from the raw data being processed. Raw data will be refreshed and we want to minimised the downstream code changes this causes.
  • See the script build.sh in the root of the pipeline code folder. This automates running all .sh and .sql scripts that are found in the pipeline code.
    • a build script allows you to quickly and easily tear down and rebuild a pipeline. This encourages frequent rebuilds, making you more agile and helping you uncover integration bugs between the various pipeline data steps.
  • See the second workproduct folder
    • This new workproduct is now using the pipeline clean dataset in L2.
    • This simplifies the workproduct code.
    • This makes the code more efficient
    • There is only one place where agreed cleaning rules are applied - the pipeline.

Checkpoint

This is another sensible checkpoint for a team's maturity.

  1. Cleaning rules are now consistently applied in one place instead of scattered through workproducts
  2. Potentially large datasets are in a database where they can be accessed by multiple team members concurrently and in a performant way.
  3. If rules change, a new pipeline version can be built into a new pipeline namespace (proj001_lfb_0_0_1, proj001_lfb_0_0_2 etc.). This means that older workproducts do not break. The impact of changes can be measured by diffing datasets between different pipeline versions.
  4. A build script allows the building of the pipeline to be easily automated.

Create a data model 005_model_data

Now start examining the entities in the data so they can be pulled into a normalised data model at L3. wp_003_normal_model explores relationships in the incidents data and draws conclusions on how the data should be modeled.

  • open wp_003_normal_model. See how we explore the uniqueness of entities and whether they have one-to-many relationships by testing various joins.
  • Examine the L3 pipeline code where the normalised tables are created.
    • the pipeline semantic version number has been increased from proj001_lfb_0_0_1 to proj001_lfb_0_1_0 to reflect the backwards compatible changes of the new L3 datasets created and the additional column in L2 incidents.
    • Execute the pipeline to build all your tables into a new namespace proj001_lfb_0_1_0.
    • note that an issue with postcode_full and postcode_district led us to decide that postcode_district is needed in L3. L2 code was therefore modified to pull through this field when it had previously been dropped.

Doing ad-hoc analytics 006_analytics

Here we do some analytics by creating an analytics dataset from our normal model in L3.

  • open wp_004_event_analytics and read through the code.
    • two datasets are extracted from L3 and joined together in pandas.
    • We do not put this analytics dataset into the pipeline until we know it is worth the investment. That is, until we know the same analytics dataset will be required multiple times in other workproducts.

Data refresh 007_data_refresh

We have been informed that new data on incidents has arrived. The extraction of this data was delayed until now. We need to incorporate this data into our pipeline and refresh some of the analytics we have done so our customers have the latest numbers.

  • the pipeline version has been increased to proj001_lfb_1_0_0. This major version change is because we are including new data that will potentially completely change work product conclusions.
  • note that pipeline code changes only occured in L0 through to L2. This is by design. Because datasets get mapped into a general name from L2 onwards, further downstream pipeline code is not broken by new data arriving.
  • a second version of the work product wp_004_event_analytics has been created.
    • We create a second version because we would like to quantify any changes in the analytics due to the new data that was added to the pipeline.
    • This second version points at the new pipeline version proj001_lfb_1_0_0. No other code in the workproduct needs to change.
    • The results from this new version of the workproduct are saved into a new workproduct namespace wp_004_event_analytics_02 so the previous version is not overwritten.
    • This allows us to do an analysis on how the new data may have changed the borough rankings.

About

Example training project for Guerrilla Analytics ways of working

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published