The basic premise of search engine reputation management is to use the following three strategies to accomplish the goal of creating a completely positive first page of search engine...
The basic premise of search engine reputation management is to use the following three strategies to accomplish the goal of creating a completely positive first page of search engine...
The basic premise of search engine reputation management is to use the following three strategies to accomplish the goal of creating a completely positive first page of search engine...
Major Healthcare Supply Chain Management Company Reduces Cloud Storage Cost and Optimizes Spend Analytics
Multi-Cloud Optimization and Analytics Use Case
Optimized cloud spend analytics, resulting in a significant reduction in expenses, automated cost-center chargebacks, and granular insights for IT leaders
Project Scope
Data Indicators was engaged to help the client reduce cloud storage cost and optimize cloud spend analytics. The development project’s goal was to understand the costs of GCP and AWS, report accurate cost-center chargebacks, understand GKE namespace usage and costs, and create combined cloud spend views for IT leaders. The team also aimed to create optimization/tagging alerts to help the client manage their costs effectively.
Results delivered by Data Indicators
The Client was able to establish an optimized cloud spend analytics data platform with actionable insights across IT business partners.
Finance was able to charge each cost center the appropriate amount without the help of a full-time data analyst, as all the logic was automated in LookML.
IT leaders could analyze their spend down to the project, environment, resource, SKU, or day to identify opportunities and make changes.
The Client’s IT team also has the ability to change the data model as the business changed.
By retiring the need for CloudAbility, the client saved the growing expense of 0.5% of their annual cloud spend.
Technology stack used for the project
Google BigQuery, Google Cloud SQL for PostgreSQL and Microsoft SQL Server, Google Virtual Machines, Google Cloud Storage, and Apache Airflow. To support data cataloging across the data mesh, we are implementing Secoda. For data governance, we have chosen Immuta. In addition, we used Docker where appropriate and Terraform for infrastructure management.
Enabling data-driven decision making at a major healthcare company
Enabling data-driven decision-making Use Case
Standardized tooling and centralized data management supports fast onboarding and regulatory compliance for data teams
Summary
A large healthcare client needed a solution to help them make data-centric decisions, and a project was created to deliver this. The solution involved the collection and curation of large amounts of data and providing The Client with critical insights to inform decision-making. With the help of the Data Indicators team, a scalable, secure, and easily accessible solution was developed and deployed using GCP hosting and tools.
Solution
Data Indicators developed and delivered a core suite of processes and tooling that allowed The Client’s teams to retain ownership and responsibility over their product data while providing governance and data cataloging capabilities. This enabled The Client to ensure that data usage agreements and regulatory requirements are adhered to. The solution has resulted in a 400% faster time-to-onboarding for data teams and provided a standardized suite of tooling for ETL and data quality, as well as a centralized data catalog and centralized data access governance, including rights management for data usage agreements and support for regulatory compliance and auditability.
Key benefits
Faster time-to-onboarding for data teams.
Standardized suite of tooling for ETL and data quality.
Centralized data catalog.
Centralized data access governance.
Supports regulatory compliance and auditability.
Technology stack used for the project
Google BigQuery, Google Cloud SQL, Google Cloud Storage, Google App Engine, Java, Spring Boot, Typescript, RESTful, Python and Google Apigee.
Architectural Guidance Use Case
Architectural Guidance Use Case
Comprehensive architectural guidance on foundational, structural, semantic, and organizational levels of interoperability, covering interconnectivity, data format and models, governance, and best practices.
Summary
Client needed a partner to provide architectural guidance on the four levels of interoperability, which include:
Foundational: Interconnectivity requirements between systems.
Structural: Defining the format, syntax, and organization of data.
Semantic: Underlying data models and use of data elements.
Organizational: Governance, legal policies, standards, and best practices.
Project Scope: Data Indicators was tasked with the following deliverables
Provide an overview of the current state and architecture
Conduct discovery workshops with Client’s business and IT units to understand their short-term and long-term needs.
Prepare a gap analysis of the current versus future state based on Client’s strategic outlook.
Advise on the best architectural practices and technology recommendations. Document best practices and common languages in the application integration and interoperability architecture space.
Furnish a high-level depiction of the future state.
Assist in the level of effort required to execute the architectural recommendations to the Client.
Support the final state interoperability architecture, including API management, event-based architecture (IoT and other events), and HL7 events
In-scope services delivered by Data Indicators
Worked with the Client team to build on GCP’s Apigee (API) pipeline.
Developed modernized API services using GCP’s components and libraries.
Conducted unit and integration testing of the developed code and functionality.
Deployed to various environments and provided guidance on production environments.
Provided training and handoff support.
Technology stack used for the project
•Java, Spring Boot, Typescript, RESTful, Python, MongoDB, Apigee
MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case
MarTech Infrastructure, Real-Time Engagement, and Enhanced Data Lake Maintenance Use Case
Unlock company’s potential with centralized customer profile store, real-time engagement, data segmentation, governance, and event-based data streaming pipelines.
Summary
The client was seeking a reliable partner to help them establish an integrated MarTech infrastructure that could support the creation of a centralized API-driven customer profile store, real-time customer engagement, content personalization, segmentation, and data science. They also required support in enhancing and maintaining their Hadoop-based data lake, implementing event-based data streaming pipelines, and ensuring data governance, privacy, and regulatory compliance. The client was looking for a trusted advisor to guide them through this process and help them achieve their marketing goals.
Project Scope
Data Indicators was tasked with the following deliverables:
Supporting enhancements to current state and architecture
Conducting discovery workshops with Client’s business/marketing and IT units to understand their short-term and long-term needs
Preparing a gap analysis of the current versus future state based on Client’s strategic objectives.
Advising on the best architectural practices, methodology and technology recommendations.
Documenting best practices and common languages in the application integration and interoperability architecture space.
Furnishing a high-level depiction of the future state.
Assisting in the level of effort required to execute the architectural recommendations to Client.
Data and software engineering, within teams across the organization
In-scope services delivered by Data Indicators
Developed data ingestion, streaming pipelines and modernized API services
Conducted unit and integration testing of the developed code and functionality.
Deployed to various environments and provided guidance on production environments.
Provided training and handoff support.
Technology stack used for the project
Apache Spark ( Python and Scala ), Apache NiFi for real-time orchestration and transformation, Apache Airflow for batch orchestration and transformation, Snowflake Data Warehouse, Snowflake Snowpipe real time ingestion, Snowflake Snowpark , AWS S3, AWS EC2, Docker, Kubernetes and Adobe Experience Cloud