Database to minimize batch processing time for asset manager

Client: US-based investment management and research firm serving institutional clients including pension funds, foundations, central banks, and endowments

OPPORTUNITY: The client wanted to build a modular, scalable architecture for an efficient database to minimize batch processing time. The client wanted to build a database that could be used to create exhaustive dashboards to monitor the overall performance of the business.

SOLUTION: After understanding the complexities and challenges, our data analytics team applied the following three-step approach to transforming the client’s legacy system into an intelligent, automated database:

  • Our team developed a data warehouse combining all the master data tables and transaction tables from multiple databases and sources
  • The team then implemented data standardization and historical record maintenance practices by identifying data entry points, choosing data standards, and defining the normalization matrix
  • Finally, SGA’s team implemented a data cleansing process by detecting and correcting corrupt or inaccurate records. The team also removed inconsistencies based on the client’s standard taxonomy.

VALUE DELIVERED:

  • Provided a single, unified view of market, client, and public data through real-time data integration
  • Helped the client develop data governance capabilities to check data quality and validate business rules
  • Automated extraction logic to access data across various systems

VALUE DELIVERED

►
1
Provided a single, unified view of market, client, and public data through real time data integration.
►
2
Helped the client develop data governance capabilities to check data quality and validate business rules.
►
3
Automated extraction logic to access data across various systems.