Search our sites

Search past winners/finalists


  • MESA logo
  • SATE logo

Apexon, San Antonio, Texas, United States: Krishnamurty Raju Mudunuru, Data Architect and Lead Data Engineer

Company: Apexon
Company Description: Apexon is a digital-first technology services firm focused on accelerating business transformation and delivering human-centric digital experiences. For over 27 years, Apexon has been meeting customers wherever they are in the digital lifecycle and helping them outperform their competition through speed and innovation.
Nomination Category: Thought Leadership Categories
Nomination Sub Category: Thought Leader of The Year - Business Products
2024 Stevie Winner Nomination Title: Krishnamurty Raju Mudunuru
  1. Which will you submit for this nomination, a video of up to five (5) minutes in length about the nominated individual, OR written answers to the questions for this category? CHOOSE ONE:
    Written answers to the questions
  2. If you are submitting a video of up to five (5) minutes in length, provide the URL of the nominated video here, OR attach it to your entry via the "Add Attachments, Videos, or Links to This Entry" link above, through which you may also upload a copy of your video.

    In many organizations, onboarding new data assets into a data lake is challenging due to data quality issues and the complexity of maintaining accurate catalog information. This process often results in unreliable data and inefficient resource use, hindering decision-making and operational efficiency. A streamlined solution is needed to ensure high data quality and proper catalog management, facilitating better data integration and usability.

    Despite the availability of various ETL tools in the industry, handling different file formats, ensuring data quality, and maintaining metadata often require extensive coding. Additionally, no tool specifically addresses the challenge of onboarding data assets efficiently.

    I seized this opportunity to develop a data architecture featuring a user interface and cloud-native microservices. This solution addresses the major challenges of onboarding assets into a data lake, ensuring data quality and capturing business, technical, and operational metadata.

    UI and Metadata Management
    To address the challenges of onboarding data assets into a data lake, a comprehensive data architecture that begins with an intuitive user interface (UI) integrated with a robust database is developed. This database is designed with a sophisticated data model where all tables interact cohesively, facilitating the maintenance of business, technical, and operational metadata. This metadata model is crucial for capturing and organizing essential information, ensuring that the data remains accurate and accessible. Connectors are employed to interface with diverse data sources, including flat files, databases, XML files, and JSON formats. For scenarios where connectors are not used, a predefined Excel template is available for users to input business metadata. 

    Configurable Ingestion Services
    Ingestion services are built based on the metadata, allowing for highly configurable and parameter-driven data ingestion processes. These services can extract full datasets or incremental data updates, depending on the specified configuration. The ingestion services are constructed using a microservices-based architecture, leveraging a stack of open-source technologies to ensure scalability, flexibility, and cost-effectiveness. This architecture enables seamless integration with various data sources, making the onboarding process efficient and reliable. The parameterized nature of the ingestion services ensures that they can be easily adapted to different data requirements and organizational needs.

    Data Quality and Standardization
    A critical component of the solution is the data quality dashboard, which provides users with insights into the quality issues they might face. This dashboard not only identifies data quality problems but also facilitates the standardization of data into specific formats based on configuration settings defined during the registration process. By standardizing the data, the system ensures consistency and reliability, making the data more usable for various analytical and operational purposes. This innovative approach to data quality and standardization, combined with the comprehensive metadata management and configurable ingestion services, provides a robust and scalable framework.

    Results and Advantages
    This data architecture offers significant benefits across various industries. For example, a major financial client onboarded over 2,500 data assets within 3-4 months, saving approximately $2.5 million. This rapid deployment demonstrates the solution's cost-effectiveness and scalability.

    An insurance provider migrating historical data to the cloud processed about 1.5 TB of data daily, executing over 1,500 jobs per day. This showcases the architecture's efficiency in handling large data volumes and complex workloads.

    A leading financial institution testified: "This tool is the best-suited solution for our organization. We've seen significant improvements in data availability and quality, greatly enhancing our ability to make informed decisions and drive revenue growth."

    Revenue Impact
    This product has been successfully deployed in major organizations across various industries, including banking, finance, insurance, and logistics. It has significantly reduced the efforts required by data teams, leading to improved Total Cost of Ownership and enhanced revenue. By streamlining data onboarding and ensuring high data quality, organizations have experienced increased operational efficiency and better decision-making capabilities, directly contributing to their financial growth

  3. If you are providing written answers to the questions for this category, you must answer this first question: Briefly describe the nominated individual: history and past performance (up to 200 words):

    Total 197 words used.

    From the bustling halls of Satyam Computer Services to the high-stakes environment of Barclays Technology Center, and the innovation-driven corridors of JPMorgan Chase, my career has been a testament to relentless pursuit of excellence.

    At JPMorgan Chase, I took on the monumental task of developing one of the largest data warehouses in the financial sector, a staggering 1.5 PB serving over 300 analysts. My innovative approach in optimizing data loads and transitioning to Spark and Hadoop slashed $15 million in license costs, a feat that turned heads across the industry.

    At ERP Analysts, I orchestrated a revolution in data onboarding, reducing time and costs by 70% with a groundbreaking ingestion framework. The cloud-native architectures I designed saved over $3 million, boosting revenue by 23%.

    Now, at Apexon, I lead a diverse team of 17 across three countries, driving unparalleled innovation. I crafted an Enterprise Data Management Platform that cut costs by $50 million and introduced an inline Data Quality Engine, scanning over a billion records daily to ensure flawless regulatory audits.

    These achievements reflect my relentless drive to push the boundaries of what's possible in data engineering, making a lasting impact on every organization I've touched.

  4. If you are providing written answers to the questions for this category, you must answer this second question: Outline the nominee's thought leadership achievements since 1 January 2022 that you wish to bring to the judges' attention (up to 250 words):

    Total 256 words used.

    Innovative Solutions:
    With 18 years of experience in the data space, I led the design of an enterprise data management system at Apexon, utilizing AWS and Azure Cloud Native services, Spark, and Hadoop. This platform saved $8 million in costs and optimized 5 TB of data transfers to the cloud. Additionally, we implemented an inline data quality engine, enhancing regulatory audits by scanning over 1 billion records daily.

    Architecture:
    Developed a comprehensive data architecture featuring user interfaces and cloud-native microservices. This system streamlined the onboarding of 2,500 data assets in 3–4 months, saving $2.5 million, and processed 1.5 TB of data daily, ensuring high data quality and maintaining accurate catalog information.

    Leadership and Mentorship:
    I have successfully led teams across the US, Canada, and India, fostering collaboration and high performance. As a mentor at ADPList, I provide guidance to aspiring data professionals, helping them develop essential skills for career advancement.

    Contributions and Recognition:
    Recognized as an industry expert and judge for the 2024 Information Technology Awards, American Business Awards, and Fortress Cybersecurity Awards, I also serve as an active reviewer for IEEE and IGI Global. My thought leadership is evidenced by publications in reputable journals and authoring book chapters on AI and quantum technologies.

    Impact and cost savings:
    My innovative solutions have significantly impacted customer operations by reducing costs and enhancing efficiency. For instance, a major financial client saved $2.5 million by adopting my data architecture, allowing their data science team to focus on machine learning models instead of data quality issues.

  5. If you are providing written answers to the questions for this category, you must answer this third question: Explain why the achievements you have highlighted are unique or significant. If possible compare the achievements to the performance of other players in your industry and/or to the nominee's past performance (up to 250 words):

    Total 228 words used.

    The achievements highlighted are unique and significant due to their impact on operational efficiency, cost savings, and data quality in the data engineering domain. The development of a comprehensive data architecture featuring a user interface, cloud-native microservices, and a robust metadata model addresses critical challenges in onboarding new data assets into data lakes.

    Compared to other players in the industry, this achievement stands out as existing ETL tools often require extensive coding and manual intervention to handle diverse file formats and ensure data quality. The data architecture's parameterized ingestion services and data quality dashboard provide a significant advantage, streamlining processes and reducing manual efforts. Furthermore, this achievement surpasses past performance by demonstrating a rapid deployment capability and substantial cost savings, positioning it as a leading solution in the industry.

    Client testimonials further underscore the significance of these achievements. A major financial institution noted, "This tool is the best-suited solution for our organization. We've seen significant improvements in data availability and quality, greatly enhancing our ability to make informed decisions and drive revenue growth." An insurance provider highlighted the system's efficiency, stating, "By processing about 1.5 TB of data daily and executing over 1,500 jobs per day, we've experienced remarkable improvements in handling large data volumes and complex workloads." These endorsements from satisfied clients reinforce the unique and significant impact of the developed data architecture.

  6. You have the option to reference here any attachments of supporting materials throughout this nomination and how they provide evidence of the claims you have made in this nomination (up to 250 words):

    Total 20 words used.

    https://dzone.com/articles/comprehensive-guide-to-data-analysis

    https://dzone.com/articles/snowflake-data-sharing-capabilities

    https://www.ijcseonline.org/pdf_paper_view.php?paper_id=5686&2-IJCSE-09397.pdf

    https://www.ijcseonline.org/pdf_paper_view.php?paper_id=5677&6-IJCSE-09375.pdf

Attachments/Videos/Links:
Krishnamurty Raju Mudunuru
PDF [REDACTED FOR PUBLICATION]