We get results

What is Staff Augmentation?

At Data Indicators, we acknowledge the crucial role of having a proficient and adept IT and software development team for the growth and success of any business. Our staff augmentation services are tailored to help businesses like yours in supplementing their current IT and software development team with elite professionals in the field.

Our services:
  • We collaborate with our clients to identify and recruit skilled and experienced IT and software development professionals who possess the right talents and abilities to match their specific requirements.
  • We offer flexible resources to aid you in adjusting your IT and software development workforce as per your needs.
  • Technical Expertise: Our team of experts possess specialized knowledge in various technical languages such as Java, .Net, and cloud platforms & tools such as AWS, Azure, GCP, Salesforce, Snowflake, as well as cybersecurity etc.

Major Healthcare Supply Chain Management Company Reduces Cloud Storage Cost and Optimizes Spend Analytics

Multi-Cloud Optimization and Analytics Use Case

Optimized cloud spend analytics, resulting in a significant reduction in expenses, automated cost-center chargebacks, and granular insights for IT leaders

Project Scope

Data Indicators was engaged to help the client reduce cloud storage cost and optimize cloud spend analytics. The development project’s goal was to understand the costs of GCP and AWS, report accurate cost-center chargebacks, understand GKE namespace usage and costs, and create combined cloud spend views for IT leaders. The team also aimed to create optimization/tagging alerts to help the client manage their costs effectively.

Results delivered by Data Indicators

  • The Client was able to establish an optimized cloud spend analytics data platform with actionable insights across IT business partners.
  • Finance was able to charge each cost center the appropriate amount without the help of a full-time data analyst, as all the logic was automated in LookML.
  • IT leaders could analyze their spend down to the project, environment, resource, SKU, or day to identify opportunities and make changes.
  • The Client’s IT team also has the ability to change the data model as the business changed.
  • By retiring the need for CloudAbility, the client saved the growing expense of 0.5% of their annual cloud spend.

Technology stack used for the project

Google BigQuery, Google Cloud SQL for PostgreSQL and Microsoft SQL Server, Google Virtual Machines, Google Cloud Storage, and Apache Airflow. To support data cataloging across the data mesh, we are implementing Secoda. For data governance, we have chosen Immuta. In addition, we used Docker where appropriate and Terraform for infrastructure management.

Enabling data-driven decision making at a major healthcare company

Enabling data-driven decision-making Use Case

Standardized tooling and centralized data management supports fast onboarding and regulatory compliance for data teams

Summary

A large healthcare client needed a solution to help them make data-centric decisions, and a project was created to deliver this. The solution involved the collection and curation of large amounts of data and providing The Client with critical insights to inform decision-making. With the help of the Data Indicators team, a scalable, secure, and easily accessible solution was developed and deployed using GCP hosting and tools.

Solution

Data Indicators developed and delivered a core suite of processes and tooling that allowed The Client’s teams to retain ownership and responsibility over their product data while providing governance and data cataloging capabilities. This enabled The Client to ensure that data usage agreements and regulatory requirements are adhered to. The solution has resulted in a 400% faster time-to-onboarding for data teams and provided a standardized suite of tooling for ETL and data quality, as well as a centralized data catalog and centralized data access governance, including rights management for data usage agreements and support for regulatory compliance and auditability.

Key benefits

  • Faster time-to-onboarding for data teams.
  • Standardized suite of tooling for ETL and data quality.
  • Centralized data catalog.
  • Centralized data access governance.
  • Supports regulatory compliance and auditability.

Technology stack used for the project

Google BigQuery, Google Cloud SQL, Google Cloud Storage, Google App Engine, Java, Spring Boot, Typescript, RESTful, Python and Google Apigee.

Architectural Guidance Use Case

Architectural Guidance Use Case

Comprehensive architectural guidance on foundational, structural, semantic, and organizational levels of interoperability, covering interconnectivity, data format and models, governance, and best practices.

Summary

Client needed a partner to provide architectural guidance on the four levels of interoperability, which include:
Foundational: Interconnectivity requirements between systems.
Structural: Defining the format, syntax, and organization of data.
Semantic: Underlying data models and use of data elements.
Organizational: Governance, legal policies, standards, and best practices.

Project Scope: Data Indicators was tasked with the following deliverables

  • Provide an overview of the current state and architecture
  • Conduct discovery workshops with Client’s business and IT units to understand their short-term and long-term needs.
  • Prepare a gap analysis of the current versus future state based on Client’s strategic outlook.
  • Advise on the best architectural practices and technology recommendations.
    Document best practices and common languages in the application integration and interoperability architecture space.
  • Furnish a high-level depiction of the future state.
  • Assist in the level of effort required to execute the architectural recommendations to the Client.
  • Support the final state interoperability architecture, including API management, event-based architecture (IoT and other events), and HL7 events

In-scope services delivered by Data Indicators

  • Worked with the Client team to build on GCP’s Apigee (API) pipeline.
  • Developed modernized API services using GCP’s components and libraries.
  • Conducted unit and integration testing of the developed code and functionality.
  • Deployed to various environments and provided guidance on production environments.
  • Provided training and handoff support.

Technology stack used for the project

•Java, Spring Boot, Typescript, RESTful, Python, MongoDB, Apigee

MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case

MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case

Unlock company’s potential with centralized customer profile store, real-time engagement, data segmentation, governance, and event-based data streaming pipelines.

Summary

The client was seeking a reliable partner to help them establish an integrated MarTech infrastructure that could support the creation of a centralized API-driven customer profile store, real-time customer engagement, content personalization, segmentation, and data science. They also required support in enhancing and maintaining their Hadoop-based data lake, implementing event-based data streaming pipelines, and ensuring data governance, privacy, and regulatory compliance. The client was looking for a trusted advisor to guide them through this process and help them achieve their marketing goals.

Project Scope

Data Indicators was tasked with the following deliverables:

  • Supporting enhancements to current state and architecture
  • Conducting discovery workshops with Client’s business/marketing and IT units to understand their short-term and long-term needs
  • Preparing a gap analysis of the current versus future state based on Client’s strategic objectives.
  • Advising on the best architectural practices, methodology and technology recommendations.
  • Documenting best practices and common languages in the application integration and interoperability architecture space.
  • Furnishing a high-level depiction of the future state.
  • Assisting in the level of effort required to execute the architectural recommendations to Client.
  • Data and software engineering, within teams across the organization
 

In-scope services delivered by Data Indicators

  • Developed data ingestion, streaming pipelines and modernized API services
  • Conducted unit and integration testing of the developed code and functionality.
  • Deployed to various environments and provided guidance on production environments.
  • Provided training and handoff support.

Technology stack used for the project

Apache Spark ( Python and Scala ), Apache NiFi for real-time orchestration and transformation, Apache Airflow for batch orchestration and transformation, Snowflake Data Warehouse,  Snowflake Snowpipe real time ingestion, Snowflake Snowpark , AWS S3, AWS EC2, Docker, Kubernetes and Adobe Experience Cloud