who we are

Data Indicators

Since Data Indicators inception in 1997, we have set out to offer innovative, highly specialized data-centric services to address our clients’ challenges in harnessing time-sensitive, valuable, and actionable insights from their data. Today, we steadfastly remain on the same path. Our services offerings have grown with the Cloud revolution and convergence of new technologies. We are heavily invested in helping our clients navigate their digital transformation journey and unlock real time insight, smart predictions and intelligent action from their data that can uncover new business values and opportunities.

01
Radical Integrity

Our people truly care for our work and for each other.

02
People First

We believe that a culture will build a thriving company.

03
Process Perfection

We’re driven to becoming the best version of ourselves.

Why Choose Us

Work with a Dedicated
Cloud Company

Data Indicators engineers are experts in development and support services both on premises as well as Cloud-based big data on Azure, AWS, and Google Cloud

ACTIVE CLIENTS
+
PROJECTS DONE
+
TEAM ADVISORS
+
GLORIOUS YEARS
+

Major Healthcare Supply Chain Management Company Reduces Cloud Storage Cost and Optimizes Spend Analytics

Multi-Cloud Optimization and Analytics Use Case

Optimized cloud spend analytics, resulting in a significant reduction in expenses, automated cost-center chargebacks, and granular insights for IT leaders

Project Scope

Data Indicators was engaged to help the client reduce cloud storage cost and optimize cloud spend analytics. The development project’s goal was to understand the costs of GCP and AWS, report accurate cost-center chargebacks, understand GKE namespace usage and costs, and create combined cloud spend views for IT leaders. The team also aimed to create optimization/tagging alerts to help the client manage their costs effectively.

Results delivered by Data Indicators

  • The Client was able to establish an optimized cloud spend analytics data platform with actionable insights across IT business partners.
  • Finance was able to charge each cost center the appropriate amount without the help of a full-time data analyst, as all the logic was automated in LookML.
  • IT leaders could analyze their spend down to the project, environment, resource, SKU, or day to identify opportunities and make changes.
  • The Client’s IT team also has the ability to change the data model as the business changed.
  • By retiring the need for CloudAbility, the client saved the growing expense of 0.5% of their annual cloud spend.

Technology stack used for the project

Google BigQuery, Google Cloud SQL for PostgreSQL and Microsoft SQL Server, Google Virtual Machines, Google Cloud Storage, and Apache Airflow. To support data cataloging across the data mesh, we are implementing Secoda. For data governance, we have chosen Immuta. In addition, we used Docker where appropriate and Terraform for infrastructure management.

Enabling data-driven decision making at a major healthcare company

Enabling data-driven decision-making Use Case

Standardized tooling and centralized data management supports fast onboarding and regulatory compliance for data teams

Summary

A large healthcare client needed a solution to help them make data-centric decisions, and a project was created to deliver this. The solution involved the collection and curation of large amounts of data and providing The Client with critical insights to inform decision-making. With the help of the Data Indicators team, a scalable, secure, and easily accessible solution was developed and deployed using GCP hosting and tools.

Solution

Data Indicators developed and delivered a core suite of processes and tooling that allowed The Client’s teams to retain ownership and responsibility over their product data while providing governance and data cataloging capabilities. This enabled The Client to ensure that data usage agreements and regulatory requirements are adhered to. The solution has resulted in a 400% faster time-to-onboarding for data teams and provided a standardized suite of tooling for ETL and data quality, as well as a centralized data catalog and centralized data access governance, including rights management for data usage agreements and support for regulatory compliance and auditability.

Key benefits

  • Faster time-to-onboarding for data teams.
  • Standardized suite of tooling for ETL and data quality.
  • Centralized data catalog.
  • Centralized data access governance.
  • Supports regulatory compliance and auditability.

Technology stack used for the project

Google BigQuery, Google Cloud SQL, Google Cloud Storage, Google App Engine, Java, Spring Boot, Typescript, RESTful, Python and Google Apigee.

Architectural Guidance Use Case

Architectural Guidance Use Case

Comprehensive architectural guidance on foundational, structural, semantic, and organizational levels of interoperability, covering interconnectivity, data format and models, governance, and best practices.

Summary

Client needed a partner to provide architectural guidance on the four levels of interoperability, which include:
Foundational: Interconnectivity requirements between systems.
Structural: Defining the format, syntax, and organization of data.
Semantic: Underlying data models and use of data elements.
Organizational: Governance, legal policies, standards, and best practices.

Project Scope: Data Indicators was tasked with the following deliverables

  • Provide an overview of the current state and architecture
  • Conduct discovery workshops with Client’s business and IT units to understand their short-term and long-term needs.
  • Prepare a gap analysis of the current versus future state based on Client’s strategic outlook.
  • Advise on the best architectural practices and technology recommendations.
    Document best practices and common languages in the application integration and interoperability architecture space.
  • Furnish a high-level depiction of the future state.
  • Assist in the level of effort required to execute the architectural recommendations to the Client.
  • Support the final state interoperability architecture, including API management, event-based architecture (IoT and other events), and HL7 events

In-scope services delivered by Data Indicators

  • Worked with the Client team to build on GCP’s Apigee (API) pipeline.
  • Developed modernized API services using GCP’s components and libraries.
  • Conducted unit and integration testing of the developed code and functionality.
  • Deployed to various environments and provided guidance on production environments.
  • Provided training and handoff support.

Technology stack used for the project

•Java, Spring Boot, Typescript, RESTful, Python, MongoDB, Apigee

MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case

MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case

Unlock company’s potential with centralized customer profile store, real-time engagement, data segmentation, governance, and event-based data streaming pipelines.

Summary

The client was seeking a reliable partner to help them establish an integrated MarTech infrastructure that could support the creation of a centralized API-driven customer profile store, real-time customer engagement, content personalization, segmentation, and data science. They also required support in enhancing and maintaining their Hadoop-based data lake, implementing event-based data streaming pipelines, and ensuring data governance, privacy, and regulatory compliance. The client was looking for a trusted advisor to guide them through this process and help them achieve their marketing goals.

Project Scope

Data Indicators was tasked with the following deliverables:

  • Supporting enhancements to current state and architecture
  • Conducting discovery workshops with Client’s business/marketing and IT units to understand their short-term and long-term needs
  • Preparing a gap analysis of the current versus future state based on Client’s strategic objectives.
  • Advising on the best architectural practices, methodology and technology recommendations.
  • Documenting best practices and common languages in the application integration and interoperability architecture space.
  • Furnishing a high-level depiction of the future state.
  • Assisting in the level of effort required to execute the architectural recommendations to Client.
  • Data and software engineering, within teams across the organization
 

In-scope services delivered by Data Indicators

  • Developed data ingestion, streaming pipelines and modernized API services
  • Conducted unit and integration testing of the developed code and functionality.
  • Deployed to various environments and provided guidance on production environments.
  • Provided training and handoff support.

Technology stack used for the project

Apache Spark ( Python and Scala ), Apache NiFi for real-time orchestration and transformation, Apache Airflow for batch orchestration and transformation, Snowflake Data Warehouse,  Snowflake Snowpipe real time ingestion, Snowflake Snowpark , AWS S3, AWS EC2, Docker, Kubernetes and Adobe Experience Cloud