
Shreyank B.
About Me
Shreyank is a Senior Full-stack Python Engineer
Shreyank is a Senior Full-stack Python Engineer
Led the modernization of financial task workflows by developing distributed data pipelines using PySpark and transforming legacy systems into scalable, real-time processing frameworks.
Designed and implemented real-time ingestion flows, schema management strategies, and CI/CD automation using industry-standard orchestration and version control tools.
Delivered cost-optimized ETL and ELT pipelines using dbt and PySpark and built secure and high-performance APIs with FastAPI and Flask.
Created observability dashboards to monitor system health and performance using modern visualization platforms.
Modernized legacy data pipelines by building distributed PySpark jobs on AWS EMR, reducing job execution time by 40% while enabling cost-efficient parallel processing across large-scale financial queues.
Migrated application workflows to Amazon DynamoDB for NoSQL and Amazon RDS (PostgreSQL) for relational workloads using Django ORM and SQLAlchemy.
Refactored legacy services into Flask/FastAPI microservices and developed REST APIs with Django REST Framework and Flask, deployed via AWS Fargate.
Integrated Amazon SageMaker models for fraud detection with real-time scoring shown in both Vue and React dashboards.
Used Terraform and GitHub Actions to manage CI/CD pipelines, deploy EMR clusters, Fargate services, Redshift objects, and API endpoints.
Built reusable front-end components using TypeScript, HTML5, and CSS3, improving load performance by 25% and ensuring WCAG-compliant accessibility.
Designed and developed back-end infrastructure orchestration services using Python and Terraform to automate cloud resource provisioning on AWS (EC2, IAM, S3, VPC).
Integrated ML scoring APIs into dashboards using CloudWatch streaming and React hooks, surfacing insights in near real time.
Developed Jest unit tests and snapshot tests for React components, achieving 90%+ coverage and automating regressions via GitHub Actions.
Architected scalable batch and streaming data pipelines to support claims processing and supply chain analytics across large, high-volume datasets.
Built secure, compliant data lake architectures and deployed predictive ML models to improve operational forecasting.
Established automated CI/CD workflows for consistent and reliable deployments.
Developed secure APIs and interactive dashboards to deliver real-time visibility into inventory, logistics, and alerting metrics for key stakeholders.
Engineered scalable batch pipelines using Apache Spark (PySpark) on AWS EMR, improving throughput by 30% while reducing compute costs via auto-scaling and fine-tuned cluster configurations.
Developed real-time ingestion architecture with Amazon Kinesis Data Streams, Firehose, and AWS Lambda, enabling 95% classification accuracy for health claim data and reducing ingestion anomalies by 20%.
Established a HIPAA-compliant data lake on Amazon S3, centralizing structured and semi-structured data across providers, supply chains, and patient events, with lifecycle policies and encryption at rest.
Designed ETL/ELT pipelines using AWS Glue, Airflow (MWAA), and PySpark and developed automated ingestion pipelines with Snowflake.
Created dynamic Python modules and scripts for provisioning secure VPCs, IAM roles, and data pipelines, reducing infra setup time by 60%.
Engineered high-performance data access patterns across MySQL and PostgreSQL, including query rewriting, connection pooling, and custom profiling to improve performance by 40%.
Embedded monitoring hooks via OpenTelemetry and pushed metrics to Datadog, enabling fine-grained tracking of ingestion and API latencies.
Monitored data pipeline health and model performance using CloudWatch Logs, CloudTrail, and Slack-based alerts.
Built real-time and batch data pipelines to support financial transaction analytics using distributed messaging, orchestration tools, and scalable processing frameworks.
Developed robust ETL workflows using Spark and workflow automation platforms, trained and deployed ML models for fraud detection and behavioral insights, and exposed secure APIs using FastAPI.
Automated infrastructure provisioning through IaC practices and deployed containerized services on a managed orchestration platform for scalable, resilient operations.
Collaborated with analysts and data scientists to define ingestion cadence, reporting KPIs, and real-time metrics visualized through React-based dashboards, Amazon QuickSight, and Kibana.
Built cloud-native data warehouse solutions using Amazon Redshift, applying star and snowflake schema designs, partitioning, and sort key strategies to improve query performance by 50%+.
Engineered real-time and batch ingestion pipelines using Amazon Kinesis, Apache Kafka, and AWS Glue, capturing high-volume payment event streams into Amazon S3 for downstream processing.
Played a key role in back-end modernization by building FastAPI-based APIs and migrating legacy Django logic to async-first services with Pydantic models and SQLAlchemy ORM.
Deployed containerized microservices using Docker and managed deployments using CI/CD (GitHub Actions) with pre-deployment PyTest checks and rollback strategies.
Trained and deployed fraud and churn models using SageMaker and implemented NLP pipelines using spaCy and NLTK to classify and tag user support chats and memos.
Monitored health of data pipelines using CloudWatch, CloudTrail, and Slack-based alerting, reducing MTTR and improving SLA compliance.
Built scalable data models in Redshift using dbt and schema strategies like star/snowflake schema, optimizing query latency by 50%.
Created reusable widgets in TypeScript, enhanced with CSS animations and responsive design using HTML5 flex/grid layouts.
Employed Gradle for front-end tooling and Maven to handle hybrid integrations with external Java-based fraud scoring systems.
Maintained dashboard scalability using React Context, Redux, and lazy-loaded components, reducing initial load time by 40%.
Engineered scalable ETL and real-time streaming pipelines to transform and process large-scale insurance datasets for analytics, reporting, and business insights.
Designed and maintained high-performance data warehouse solutions and deployed ML inference pipelines to support predictive analytics use cases.
Automated infrastructure provisioning using IaC methodologies and managed containerized applications on a secure, orchestrated platform to ensure high availability and scalability.
Built batch ETL jobs using Apache Spark (PySpark) on Azure HDInsight, improving processing speed by 40% across multi-TB datasets.
Automated cloud infrastructure provisioning with Terraform, ARM templates, reducing deployment time and ensuring consistent resource creation.
Optimized data models in Azure SQL Database, PostgreSQL, MySQL, and Oracle (on Azure VMs) using indexing and caching.
Integrated and deployed ML models with Azure Machine Learning with Spark SQL, supporting advanced data science workflows with 90%+ accuracy.
Containerized services using Docker and AKS and implemented CI/CD with Azure DevOps pipelines.
Wrote reusable transformation logic in Python, PySpark, and Spark SQL, enabling scalable operations.
Built high-throughput batch and streaming frameworks using Spark, Hadoop, and Kafka for healthcare data pipelines.
Developed real-time ETL workflows with Informatica, exposed internal APIs using Flask, and created business dashboards in Tableau and Power BI.
Automated infrastructure deployments using AWS CloudFormation and Jenkins CI/CD.
Engineered scalable Big Data processing frameworks using Apache Spark, Apache Hadoop, and Apache Flink.
Built and optimized a cloud-native data warehouse using Snowflake, with schema evolution and performance tuning automated through Python-based orchestration script.
Managed and tuned high-volume Oracle databases and developed secure RESTful APIs with Flask.
Created Python-based data loaders and enrichment services to support interactive dashboards built in Tableau and Power BI.
Deployed and managed containerized microservices using Docker and Kubernetes.
Provisioned infrastructure using Terraform and AWS CloudFormation.
Implemented CI/CD pipelines with Jenkins and developed custom logging and monitoring tools in Python for Datadog.