Techno Talent Inc.
Job Description
Big Data Engineer with EKS: Positions x2 Location: Tysons Corner, VA Data Engineer Job Description Summary We are seeking a highly skilled and experienced Big Data Engineer to design, develop, and optimize large-scale data processing systems. In this role, you will work closely with cross-functional teams to architect data pipelines, implement data integration solutions, and ensure the performance, scalability, and reliability of big data platforms. The ideal candidate will have deep expertise in distributed systems, cloud platforms, and modern big data technologies such as Hadoop, Spark, and Kubernetes-based orchestration. Responsibilities Design, develop, and maintain large-scale data processing pipelines using Big Data technologies (e.g., Hadoop, Spark, Python, Scala). Architect and deploy containerized big data workloads on Amazon EMR on EKS (Elastic Kubernetes Service). Design and implement Kubernetes-based infrastructure for running Spark applications at scale. Implement data ingestion, storage, transformation, and analysis solutions that are scalable, efficient, and reliable. Stay current with industry trends and emerging Big Data technologies to continuously improve the data architecture. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. Optimize and enhance existing data pipelines for performance, scalability, and reliability. Develop automated testing frameworks and implement continuous testing for data quality assurance. Conduct unit, integration, and system testing to ensure the robustness and accuracy of data pipelines. Work with data scientists and analysts to support data-driven decision-making across the organization. Ability to write and maintain automated unit, integration, and end-to-end tests. Monitor and troubleshoot data pipelines in production environments to identify and resolve issues. Manage Kubernetes clusters, pods, services, and deployments for big data workloads. Essential Technical Skills: AI Tool Proficiency Hands-on experience with AI development tools (GitHub Copilot, Q Developer, ChatGPT, Claude, etc.) Big Data Technologies: Experience with Big data technologies such as Hadoop, Spark, Hive & Trino Understanding of common issues like data skew and strategies to mitigate it, working with massive data volumes in PetaBytes, and troubleshooting job failures due to resource limitations, bad data, and scalability challenges. Real-world experience with debugging and mitigation strategies. Container Orchestration & Kubernetes: Strong experience with Kubernetes architecture, concepts, and operations (pods, services, deployments, namespaces, ConfigMaps, Secrets) Hands-on experience with Amazon EMR on EKS (Kubernetes) for running Apache Spark workloads Experience with Kubernetes resource management, scheduling, and auto-scaling Knowledge of Helm charts for deploying and managing applications on Kubernetes Understanding of Kubernetes networking, storage (PVs, PVCs), and security best practices Experience with kubectl and Kubernetes YAML manifests Ability to troubleshoot Kubernetes cluster issues, pod failures, and resource constraints Experience integrating Spark with Kubernetes operators and dynamic allocation Cloud Technologies Experience with AWS services like S3, EMR, EMR on EKS, Glue, Lambda, Athena, etc. Hands-on experience using S3 with Spark (e.g., dealing with file formats, consistency issues) Strong experience with Amazon EKS (Elastic Kubernetes Service) architecture and best practices Experience with AWS IAM roles for service accounts (IRSA) for Kubernetes workloads Knowledge of AWS networking for EKS (VPC, subnets, security groups) Experience with AWS monitoring and logging tools (CloudWatch, CloudTrail) for Kubernetes workloads Serverless knowledge (Lambda, Fargate) Programming - Python or Scala: Ability to write clean, modular, and performant code Experience with functional programming concepts (e.g., immutability, higher-order functions) Real-world use cases where scalable data processing code was implemented Strong understanding of collections, concurrency, and memory management SQL Skills (Window Functions, Joins, Complex Queries): Proficiency with SQL window functions, multi-table joins, and aggregations Ability to write and optimize complex SQL queries Experience handling edge cases like NULLs, duplicates, and ordering