Average salary: Rs944,444 /yearly
More statsSearch Results: 914 vacancies
...with 5+ Years on Pyspark/NoSQL is Mandatory)1. Person should be strong in Pyspark2. Should have hands on in MWAA (Airflow) / AWS EMR(Hadoop, Hive) framework3. Hands on and working knowledge in Python4. Knowledge in AWS services like EMR , S3, Lamda, Step Function, Aurora -...
...identify improvement areas and suggest the technology solutions
One or two industry domain knowledge
Client Interfacing skills
Project and Team management Technical and Professional Requirements:
Primary skills:Technology- Big Data - Hadoop- Hadoop Administration
...bring your talent and ambition to make a difference. We will create a world of opportunities for you.
Job Details
Job Title : Hadoop Administrator
Location: Bangalore / Pune / Hyderabad / Noida / Kolkata
Quick joiners needed
Minimum Requirement :
Must...
Responsibilities:- Designing and implementing fine-tuned production-ready data/ML pipelines in Hadoop platform.- Driving optimization, testing, and tools to improve quality.- Reviewing and approving high-level & detailed design to ensure that the solution delivers to the business...
Job Description :- Must have working experience Designing, building, installing, configuring and supporting of Hadoop.- Good to have Teradata, Cloud & Snowflakes Knowledge- Must have working experience in IntelliJ IDEA, AutoSys(Control M), WinSCP, Putty & GitHub.- Translate...
...Position : Hadoop Developer
Experience: 6+ Years
Location : Chennai/Bangalore(Work from Office)
Mandatory Skill - pySpark,Hadoop,GCP, Python
Professional experience with a cloud platform
Developer must have sound knowledge in Apache Spark and Python programming...
CSA ID Hadoop CSA Status Active Title Hadoop developer Job Type Contract JD hadoop developer or big data engineer.
Advanced knowledge of the hadoop ecosystem and its components.
In-depth knowledge of Hive, HBase, and spark
Familiarity with MapReduce
Knowledge of...
...Iterators in Scala.- Have worked on multi-threading it will be helpful.- Experience in working with Kafka will be helpful.- Knowledge of Hadoop MapReduce, HDFS, Hbase, and Hive will be considered a plus.- Exposure to DevOps and SQL (Postgres, MS SQL) will be considered as an...
...statistical analysis, and data visualization tools and technologies- Experience with big data processing frameworks, such as Spark or Hadoop- Strong analytical and problem-solving skills, with the ability to analyze complex data sets and identify trends and insights- Excellent...
Mandatory Skills:- Data scientist exp- Hadoop, Scala- Python- Machine learning- SQL- Only from product OrgsRequirements from Past Experience :- Prior experience of working with large-scale datasets (10s of millions of documents) is strongly preferred.- End-to-end ownership of...
...leading on client-facing projects, including working in close-knit teams- 3+ years of experience and interest in Big Data technologies (Hadoop / Spark / Relational DBs)- 3+ years of experience working on projects within the cloud ideally AWS or Azure- Data Warehousing...
...BigData-Hadoop/Scala Engineer
Position Overview
Job Title: Data Engineer (ETL, Big Data, Hadoop, Spark, GCP)
Corporate Title:Assistant Vice President
Location- Bangalore, India
Role Description
Senior Engineer is responsible for developing and delivering...
3-4 years of experience with building big data applications and robust data pipelines running on Hadoop cluster with exposure to building frameworks related to batch and stream consumption using Big Data tech stack such as Spark, Kafka, Hive, HDFS, HBase, along with exposure...
Job Description :- Hadoop ecosystem (HDFS, Hive, Yarn, File formats like Avro/Parquet)- Python programming language is mandatory.- PySpark - Excellent with SQL- Excellent with Airflow is a plus.Good to Have:- Airflow- Good aptitude, strong problem-solving abilities, and analytical...
Skills : Hadoop, Java, Python, Scala, SqlRequirements :- 3-4+ yrs as a backend developer- Worked in a good startup or good engineering colleges- Working with an early-stage startup in a role that involves- Building scalable systems to extract information from 15Mn+ domains.-...
Job Description :Mandatory Skills : Microsoft Azure, Hadoop, Spark, Databricks, Airflow, Kafka, Py sparkRequirements :- Experience working with distributed technology tools for developing Batch and Streaming pipelines using SQL, Spark, Python, Airflow, Scala, Kafka- Experience...
...Hands on coder with good experience in programming languages like Java, Python or Scala.- Hands-on experience on the Big Data stack like Hadoop, Mapreduce, Spark, Hbase, and ElasticSearch.- Good understanding of programming principles and development practices like checkin...
...experience using Azure SQL or SQL DWH (synapse)- Knowledge of batch and streaming data architectures- Experience using big data technologies (Hadoop, Hive, HBase, Spark and others).- Strong knowledge of SQL (MSSQL or MySQL or PostgreSQL or Presto).- Demonstrated strength in data...
...Solutions Architect (Data & BI) with around 10 - 15 years of experience in the following areas,Mandatory Skills :- Tableau, Big Data, Hadoop, Data Warehousing, Legacy BI Migrations, Cloud technologies.- Experienced in BI Implementation, BI migration involving Tableau.Preferred...
...projects, including working in close-knit teams - Overall, 5 + years of experience, with at least 3+ years in Big Data technologies (Hadoop / Spark / Relational DBs) and similar experience in working on projects within the cloud ideally AWS or Azure - Data Warehousing experience...