Position: Azure Pyspark
Pyspark + with Data warehousing and Azure will be add on.
· Must have low level design and development skills. Should able to design a solution for a given use cases.
· Agile delivery- Person must able to show design and code on daily basis
· Must be an experienced PySpark developer and Scala coding. Primary skill is PySpark
· Must have experience in designing job orchestration, sequence, metadata design, Audit trail, dynamic parameter passing and error/exception handling
· Good experience with unit, integration and UAT support
· Able to design and code reusable components and functions
· Should able to review design, code & provide review comments with justification
· Zeal to learn new tool/technologies and adoption
· Good to have experience with Devops and CICD
Nos of Resources required: 1 to 2
Work location: Bangalore
Experience: 8 yrs – 9 yrs
Mobilization Period in weeks: 2 weeks
SAS Data Integration Consultant:
Web Scrapping Programmers:
· As a Python Developer, your role is to apply your knowledge set to fetch data from multiple online sources, cleanse it and build APIs on top of it
· Develop a deep understanding of our vast data sources on the web and know exactly how, when, and which data to scrap, parse and store this data
· Work closely with Database Administrators to store data in SQL and NoSQL databases
· Develop frameworks for automating and maintaining the constant flow of data from multiple sources
· Work independently with little supervision to research and test innovative solutions
Skills and Qualifications:
· Strong coding experience in Python (knowledge of Java, Javascript is a plus)
· Experience with SQL databases
· Experience with multi-processing, multi-threading and AWS/Azure
· Strong knowledge of scraping frameworks such as Python(Request, BeautifulSoup), Web-Harvest and others
· Previous experience with web crawling is a must
Experience:
Job location:
Duration:
Joining:
Position: Hadoop Admin
Must Have Technical Skills: Hadoop Admin
Good to have Technical Skills: Linux Admin, ETL
· Extensive experience with RedHat Linux and Cloudera is mandatory.
· Experience in installing, configuring, upgrading and managing Hadoop environment.
· Responsible for deployments, and monitor for capacity, performance, and/or troubleshooting issues.
· Work closely with data scientists and data engineers to ensure the smooth operation of the platform.
· End-to-end performance tuning of the clusters.
· Maintains and administers computing environments including computer hardware, systems software, applications software, and all configurations.
· Defines procedures for monitoring and evaluates, diagnoses and establishes work plan to resolve system issues.
· Working knowledge of entire Hadoop ecosystem like HDFS, Hive, Yarn, Oozie, Kafka, Impala, Hive, Kudu, HBase, Spark and Spark Streaming.
· Knowledge of Private and Public cloud computing and virtualization platform.
Nos of Resources required: 2 to 3
Work location: Remote
Qualification: BTech
Experience: 4 yrs – 5 yrs
Mobilization Period in weeks: 1 week
Duration: 3 to 6 months
Machine Learning Programmers:
Responsibilities
Requirements
Experience:
Job location:
Duration:
Joining: