Description We are Cognizant Artificial IntelligenceDigital technologies, including analytics and AI, give companies a once-in-a-generation opportunity to perform orders of magnitude better than ever before. But clients need new business models built from analyzing customers and business operations at every angle to really understand them.With the power to apply artificial intelligence and data science to business decisions via enterprise data management solutions, we help leading companies prototype, refine, validate and scale the most desirable products and delivery models to enterprise scale within weeks*You must be legally authorized to work in the United States without the need for employer sponsorship, now or at any time in the future *Create and maintain optimal data pipeline architecture, Assemble large, complex data sets that meet functional / non-functional business requirementsIdentify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability etcBuild the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS 'Big data' technologiesBuild analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metricsWork with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needsCreate data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader5 years of experience (Sr-level) Strong Programming experience with object-oriented/object function scripting languages: Scala5 years of experience (Mid-level) Experience with big data tools: Hadoop, Apache Spark etc1 years of strong technical Experience with AWS cloud services and DevOps engineering: S3, IAM, EC2, EMR, RDS, Redshift, Cloudwatch with Docker, Kubernetes, GitHub, Jenkins, CICDExperience with stream-processing systems: Python, Spark-Streaming, etc. (Nice to have)1 Years of experience with relational SQL, Snowflake and NoSQL databases, like Postgres and CassandraMust Have SkillsSparkAWS EMRSparkSQLScalaPythonQualifications Technical SkillsSNoPrimary SkillProficiency Level *Rqrd./Dsrd.1SparkPL1Required2AWS EMRPL1Required3SparkSQLPL1Required4ScalaPL1Required5PythonPL1Required* Proficiency LegendsProficiency LevelGeneric ReferencePL1The associate has basic awareness and comprehension of the skill and is in the process of acquiring this skill through various channels.PL2The associate possesses working knowledge of the skill, and can actively and independently apply this skill in engagements and projects.PL3The associate has comprehensive, in-depth and specialized knowledge of the skill. She / he has extensively demonstrated successful application of the skill in engagements or projects.PL4The associate can function as a subject matter expert for this skill. The associate is capable of analyzing, evaluating and synthesizing solutions using the skill.