Pooja Shinde

Pooja Shinde

Software Engineer

Followers of Pooja Shinde496 followers
location of Pooja ShindePune, Maharashtra, India

Connect with Pooja Shinde to Send Message

Connect

Connect with Pooja Shinde to Send Message

Connect
  • Timeline

  • About me

    Senior Consultant - Data Engineer at Thoughtworks | Unlocking the Power of Data | AWS Certified | Databricks Certified | Big Data & PySpark Expert | Transforming Insights into Action

  • Education

    • Kendriya Vidyalaya ,pune

      2005 - 2010
      Science
    • DY Patil College of Engineering, Pune

      2011 - 2014
      Bachelor of Engineering - BE Computer Software Engineering
    • Pune University

      2010 - 2014
      Bachelor of Engineering (BE) Computer Engineering
    • Malaviya National Insitute of Technology Jaipur

      2016 - 2018
      Master of Technology (MTech) Computer Engineering

      Activities and Societies: working on various security and privacy issues for big data in distributed environment. student placement coordinator at MNIT Jaipur

  • Experience

    • Persistent Systems

      Nov 2014 - Jun 2016
      Software Engineer

      Contributed in Drupal-based website development, integrating Apache Solr to implement advanced search functionalities such as faceting, highlighting, and stop-word removal. Assisted in managing Solr configurations in a clustered environment with ZooKeeper to ensure high performance and reliability.Contributed to the internal portal developments.Gained hands-on experience with Apache Solr, focusing on configuring and optimizing search features to improve user experience.Completed comprehensive full-stack training.Utilized a variety of technologies, including Apache Solr, AWS, Drupal, Java, JavaScript, and ZooKeeper, to contribute to the development of comprehensive solutions. Show less

    • BlazeClan Technologies

      Sept 2018 - Oct 2019
      Cloud Engineer

      Designed and developed a cloud-based data lake to facilitate reporting and analytics, showcasing my expertise in architecting scalable and efficient data storage solutions on AWS. Developed real-time Serverless solutions to enhance sales performance, effectively leveraging AWS services for immediate data processing.Implemented Python-based Lambda functions for efficient data processing, demonstrating my proficiency in serverless computing.Executed PySpark jobs on Amazon EMR for historical data loads, highlighting my expertise in big data processing and analysis.Integrated ElasticSearch for audit logging and reporting in Kibana, showcasing my ability to utilize AWS Elasticsearch Service for comprehensive log analytics.Created a UI-based testing automation framework for Data Lake projectsLeveraged Python libraries such as Pandas, NumPy, and Boto3 for efficient data processing and seamless interaction with AWS services. Show less

    • Saksoft

      Oct 2019 - Sept 2021
      Consultant

      Developed a robust Data Lake pipeline insurance client, integrating and processing data from upstream systems like Headless Manufacture for downstream accessibility. Conducted data cleaning and transformations across multiple channels, generating over 100 tailored reports to fulfill diverse business needs.Managed JSON data in ingestion pipelines, converting it to Parquet format for optimized storage and processing.Utilized AWS Step Functions to streamline workflows, enhancing operational efficiency.Created APIs for testing teams to fetch entity status and generate ad-hoc reports.Established Docker environments for efficient querying of Parquet data.Developed automated email solutions to disseminate client reports via AWS Glue jobs.Restructured existing jobs using design patterns and OOP principles without altering their external behavior.Leveraged AWS Textract for processing images, PDFs, and files aimed at report generation.Created dashboards on Amazon QuickSight, delivering insights from various data sources.Conducted reconciliation to identify inaccuracies in daily data transmissions.Managed client group products, collaborating directly with clients for new developments and issue resolution.Developed a Kinesis pipeline for logging data flow, ensuring relevance and integrity throughout the process.Utilized AWS services such as Lambda, Glue, SQS, SNS, Kinesis, DynamoDB, Aurora, and S3 to effectively address data processing and integration needs. Show less

    • Aligned Automation

      Sept 2021 - Dec 2024
      Senior Solutions Specialist

      Played a pivotal role in enhancing data accessibility, reliability, and security across various projects.Successfully orchestrated the migration of data from diverse sources, including legacy systems, to AWS Data Lake and RDS, utilizing AWS DMS and robust data processing techniques with Python and PySpark.Automated cloud-based data pipelines that support real-time reporting and analytics using advanced visualization tools like Power BI. Implemented event-driven architectures with AWS services like EventBridge and Lambda, ensuring efficient data handling.Led Agile ceremonies and streamlined processes to boost team productivity and deliver projects on time, establishing a culture of collaboration and continuous improvement.Implemented advanced data security measures, including PII data handling protocols and encryption techniques, ensuring compliance with regulatory standards.Designed auto-deployment processes for data management and established CI/CD practices using Git and CodePipeline, enhancing operational efficiency and reliability.Guided and mentored a team of developers, promoting best practices and fostering a collaborative environment that drives project success. Show less

    • Thoughtworks

      Dec 2024 - now
      Senior Consultant
  • Licenses & Certifications

    • ChatGPT and AI tools

      Be10x
      Sept 2024
    • AWS Certified Solutions Architect - Associate

      Amazon Web Services (AWS)
      Sept 2019
    • BigData Hadoop Certification

      Edureka
      Jul 2018
    • Databricks Certified Data Engineer Associate

      Databricks
      Feb 2025
      View certificate certificate
    • Business Intelligence using Power BI

      Skill Nation
      Sept 2014
  • Honors & Awards

    • Awarded to Pooja Shinde
      Innovation Award Saksoft Sep 2020
    • Awarded to Pooja Shinde
      Best Project Team Customer Obession award Blazeclan Jul 2019
    • Awarded to Pooja Shinde
      You made a difference award Blazeclan Sep 2018
    • Awarded to Pooja Shinde
      Got 2 CVE's registered for Kerberos Protocol [CVE-2018-5709 and CVE-2018-5710] NVD ( National Vulnerability Database US) Jan 2018 Kerberos is Network authentication Protocol developed by MIT. Kerberos Protocol is used for authentication Purposes. We detected 2 vulnerabilities in the latest version of this protocol (krb5-1.6.0) and got approved by MIT. We have given solution for resilient Kerberos and shown the impact of the detected vulnerability in distributed systems like Hadoop. The detected vulnerabilities got recognition from all products using it like Redhat, Microsoft Window, Fedora and from MIT itself.Work… Show more Kerberos is Network authentication Protocol developed by MIT. Kerberos Protocol is used for authentication Purposes. We detected 2 vulnerabilities in the latest version of this protocol (krb5-1.6.0) and got approved by MIT. We have given solution for resilient Kerberos and shown the impact of the detected vulnerability in distributed systems like Hadoop. The detected vulnerabilities got recognition from all products using it like Redhat, Microsoft Window, Fedora and from MIT itself.Work Done under the supervision of Emmanuel Pilli (Associate Professor at MNIT).With Special thanks to Sharwan Ram for continues support in my work. Show less