13

Big Data Job Vacancies in Pune

filter
  • Location
  • Role
  • Functional Area
  • Qualification
  • Experience
  • Employer Type
  • 8 - 12 yrs
  • Kharadi Pune
Core Java Apache Beam Google Cloud Platforms Spring Boot Big Data Microservices Data Base
Requirements:8+ years of experience in Core Java and Spring Framework (Mandatory)Minimum 2 years of experience in Google Cloud Platform (GCP) (Mandatory)Hands-on experience with Apache Beam / Dataflow for building ETL/data pipelines (Mandatory)Strong expertise in big data processing on distributed systemsProficiency with RDBMS, NoSQL, and Cloud-native databasesExperience in handling multiple data formats (Flat file, JSON, Avro, XML, etc.) with schema/contract definitionsExperience in Microservices architecture and API integration patternsStrong understanding of data structures and data model design
View all details
  • 8 - 10 yrs
  • Pune
Kafka Scala Spark Hadoop Airflow Data Lakes Kappa Kappa ++ Architectures RDBMS NoSQL Cassandra Redis Oracle
Sr. Big Data Engineer Location: PuneExperience: 10+ years Mode: HybridRole Overview:We are seeking a talented Sr. Big Data Engineer to design, develop, and support a highly scalable, distributed SaaS-based Security Risk Prioritization product. You will lead the design and evolution of our data platform and pipelines, providing technical leadership to a team of engineers and architects.Key Responsibilities: Provide technical leadership on data platform design, roadmaps, and architecture. Design and implement scalable architecture for Big Data and Microservices environments. Drive technology explorations, leveraging knowledge of internal and industry prior art. Ensure quality architecture and design of systems, focusing on performance, scalability, and security. Mentor and provide technical guidance to other engineers.Required Skills & Technologies: Mandatory: Kafka, Scala, Spark. Big Data & Data Streaming: Spark, Kafka, Hadoop, Presto, Airflow, Data lakes, lambda architecture, kappa, and kappa ++ architectures with flink data streaming. Databases & Caching: RDBMS, NoSQL, Oracle, Cassandra, Redis. Search Solutions: Solr, Elastic. ML & Automation: Experience with ML models engineering and related deployment, scripting, and automation. Architecture: In-depth experience with messaging queues and caching components. Other Skills: Strong troubleshooting and performance benchmarking skills for Big Data technologies.Qualifications: Bachelors degree in Computer Science or equivalent. 8+ years of total experience, with 6+ years relevant. 2+ years in designing Big Data solutions with Spark. 3+ years with Kafka and performance testing for large infrastructure.
View all details

Big Data Engineer (Spark and Scala)

E2E Infoware Management Services

Scala Spark Pyspark
Role: Bigdata Developer - Scala SparkExp: 5+ YrsMode of Work: WFO All 5 daysLocation: Chennai/Bangalore/PuneInterview: Any one Level F2FJob Description: Total IT / development experience of 3+ years Experience in Spark (Scala-Spark ) developing Big Data applications on Hadoop, Hive and/or Kafka, HBase, MongoDB Deep knowledge of Scala-Spark libraries to develop and debug complex data engineering challenges Experience in developing sustainable data driven solutions with current new generation data technologies to drive our business and technology strategies Exposure to deploying on Cloud platforms At least 2 years of development experience on designing and developing Data Pipelines for Data Ingestion or Transformation using Spark-Scala At least 2 years of development experience in the following Big Data frameworks: File Format (Parquet, AVRO, ORC), Resource Management, Distributed Processing and RDBMS At least 2 years of developing applications in Agile with Monitoring, Build Tools, Version Control, Unit Test, Unix Shell Scripting, TDD, CI/CD, Change Management to support DevOps
View all details

Big Data Lead

Hexaware Technologies

Snowflake Python SQL
Must have 4-6 years of experience in Data warehouse, ETL, BI projects Must have atleast 4+ years of experience in Snowflake Expertise in Snowflake architecture is must. Must have atleast 3+ years of experience and strong hold in Python/PySpark Must have experience implementing complex stored Procedures and standard DWH and ETL concepts Proficient in Oracle database, complex PL/SQL, Unix Shell Scripting, performance tuning and troubleshoot Good to have experience with AWS services and creating DevOps templates for various AWS services. Experience in using Github, Jenkins Good communication and Analytical skills Snowflake certification is desirable
View all details

Get Personalized Job Matches

Based on your experience, skills, interests, and career goals to help you find the most relevant opportunities faster. Register Now!

IT Trainer

Vijaya Management Services

  • 2 - 8 yrs
  • 5.0 Lac/Yr
  • Pune
Ava Python Big Data Technologies Hadoop Spark PiSpark Kafka Airflow Machine Learning Deep Learning Tableau Power BI
Training on Java, Python, Big data technologies, Hadoop, Spark, PiSpark, Kafka, Airflow, Machine Learning, Deep learning, Tableau, Power BI, TableauMin Experience: 2 to 3 years of training experience.
View all details

Big Data & Cloud Senior Architect

Brid Tech Solutions Private Limited

Big Data Cloud Architect Senior Cloud Cloud Engineer Work From Home Walk in
Job DescriptionThe Big Data/Analytics Solution Architect is responsible for understanding emerging and evolving end user usage models and requirements in Big Data and Analytics, documenting those usage models and business, technical and user requirements and designing a solutions architecture to meet those requirements and specifying an implementation HW and SW solution stack. Solution architects document the solution architectures and solution requirements and, when needed, define end user proofs of concept to test the architectures, usage models and corresponding Intel technologies in testbed or real end user environments. They also work with end users and ecosystem partners to deploy those solutions in early adopter production environments. A strong candidate will have:Experience in Requirements Engineering, Solution Architecture, Design, Development and DeploymentA broad set of technical skills and knowledge across hardware, software, systems and solutions development and a across more than one technical domain.Demonstrated experience in real world IT or other solutions environments including creating (on your own or with a team) a product or IT solution in the area of Big Data/AnalyticsStrong communication skills including representing your company in industry standards organizations or industry technical forums or events in Cloud SecurityStrong technical team leadership, mentorship and collaboration.Ability to develop technical relationships with end-users, ISVs, OEMs and Intel platform architects and Proof-Of-Concept engineers.
View all details
  • 8 - 14 yrs
  • 100.0 Lac/Yr
  • Pune
Python Data Bricks SQL Cloud Technologies Big Data Technologies
We are hiring for one of our US MNC Client for Principal Data as well as Senior Data Engineer Position .Position : Senior Data Engineer & Principal Data EngineerExperience : 8 to 15 Years Location : Remote Extensive experience in developing data and analytics applications in geographically distributed teams. Hands-on experience in using modern architectures and frameworks, structured and unstructured data, and programming with Python. Hands-on SQL knowledge and experience with relational databases such as MySQL, PostgresSQL, and others. Hands-on ETL knowledge and experience. Knowledge of commercial data platforms (Data bricks, Snowflake) or cloud data warehouses (Redshift, Big Query). Knowledge of data catalog and MDM tooling (Atlan, Alation, Informatica, Collibra). Knowledge of how machine learning / A.I. workloads are implemented in batch and streaming, including the preparing of datasets, training models, and using pre-trained models. Excellent analytical and troubleshooting skills. Excellent communication skills . Excellent English (both verbal and written). B.S. in Computer Science or equivalent.
View all details

Corporate Trainer

The Magic Data

Java Full Stack Software Testing Life Cycle Machine Learning Python Trainer Big Data Robotic Process Automation RPA Developer Azure Deep Learning Corporate Trainer
We are looking for a trainer who can conduct training in any one or more of the topics mentioned below:1. Java Full Stack2. Software Testing3. Machine Learning (Python)4. Robotic Process Automation (RPA)5. Big Data
View all details
Hadoop Hive Kafa Python Big Data Engineer JSON Work From Home
Brief about the Company:AdZapier Corporation is a global technology and enablement services company with a vision to transform data into value for everyone. Through a simple open approach, in connecting systems and data, we provide the data foundation for the worlds best marketers. By making it safe and easy to activate, validate, enhance, and unify data. We provide marketers with the ability to deliver relevant messages at scale and tie those messages back to actual results. Our products and services enable individual-based marketing, allowing our clients to generate a higher ROI and drive better omni-channel customer experiences.Position Description:Join our Information Technology team where you will work on new technologies and find ways to meet our customers needs and make it easy for them to do business with us.You will use functional expertise to act as an advisor to management and make recommendations on more complex projects. You will use professional concepts and company policies & procedures to solve a wide range of difficult problems creatively and practically.ResponsibilitiesYou will be responsible for operations and administration of Cloudera Hadoop platform.You will work independently on day to day monitoring and operations of Data Analytics platform. You will be required to develop automation using scripting languages. After initial training, you will be able to handle critical operation tasks as well as on demand requests.Minimum Requirements: 5+ years of experience in Software Development including Big Data Analytics area Experience in Hadoop Big Data Platform Operations and Administration High Proficiency working with Hadoop platform including Hadoop, Hive, Spark/Scala, Java, Kafka, Flume etc. Experience with any scripting language such as BASH, Scala or Python Good understanding of file formats including JSON, Parquet, Avro, and otherWork Hours - 2.30 pm noon to 11.30 pm (Mon-Fri) (US Shift)
View all details
Big Data React JS Python AWS C++ Angular Spark Programming ETL SQL Work From Home
**Preference will be given to the candidates who can join on or before 1st of October, 2022**You will:Write excellent production code and tests and help others improve in code-reviewsAnalyze high-level requirements to design, document, estimate, and build systemsCoordinate across teams to identify, resolve, mitigate and prevent technical issuesCoach and mentor engineers within the team to develop their skills and abilitiesContinuously improve the team's practices in code-quality, reliability, performance, testing, automation, logging, monitoring, alerting, and build processesYou have:For (Full stack):2 - 10 Years of experienceStrong with DS & AlgorithmsHands on Experience in the Programming languages: JavaScript (React or Angular), Python, SQL.Experience with AWS.For (Backend):2 - 10 years of experienceHands on product development experience using Java/ C++/PythonExperience with AWS,SQL,GITStrong with Data structures and AlgorithmsAdditional nice to have skills/certifications:For Java skill set:Mockito, Grizzly, Netty, VertX, Jersey / JAX-RS, Swagger / Open API, Nginx, Protocol Buffers, Thrift, Aerospike, Redis, Kinesis, Sed, Awk, PerlFor Python skill set: Data Engineering experience, Athena, Lambda, EMR, Spark, Glue, Step Functions, Hadoop, Kinesis, Orc, Parquet, Perl, Awk, RedshiftFor (Data Engineering):2 - 10 years of experienceExperience with object-oriented/object function scripting languages: Python.Experience with AWS cloud services: EC2, RDS, Redshift,S3,Athena, GlueMust be proficient in GIT, Jenkins, CICD (Continuous Integration Continuous Deployment)Experience in big data technologies like Hadoop, Map Reduce, Spark, etcExperience with Amazon Web Services and DockersFor (Geo Team):4 - 10 years of experienceExperience with Big Data technologies like Hadoop, Spark, Map Reduce, Kafka, etcExperience using object-oriented languages (Java, Python)Experience in working with different AWS technologies.Experience in software
View all details
Python SCALA JAVA AWS - EMR Hadoop Spark Kafka SQL NoSQL Data Architecture Data Structures Storm Flink
ResponsibilitiesCreate and maintain optimal data pipeline architectureAssemble large, complex data sets that meet functional / non-functional business requirements.Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Open Source and AWS big data technologiesBuild analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics.Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.Work with data and analytics experts to strive for greater functionality in our data systems.QualificationsExperience building and optimizing big data pipelines, architectures and datasets.Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.Experience interacting with customers and various stakeholders.Strong analytical skills related to working with unstructured datasets.Build processes supporting data transformation, data structures, metadata, dependency and workload management.Working knowledge of message queuing, stream processing, and highly scalable big data lakes.Strong project management and organizational skills.Experience supporting and working with cross-functional teams in a dynamic environment.They should also have experience using the following software/tools:Big data technologies: Hadoop, Spark, Kafka, etc.Relational SQL and NoSQL databases, including Postgres and Cassandra.Data pipeline and workflow management tools: Airflow, NiFi etc.Cloud services: AWS - EMR, RDS, Redshift, Glue. Azure - Databricks, Data Factory. GCP - Dataproc, Pub/SubStream-processing systems: Storm, Spark Streaming, Flink etc.
View all details

Data Specialist

Perex Engineering Private Limited

Snowflake Python ETL Hadoop Big Data Data Specialist Work From Home
We have vacant of 20 Data Engineer Jobs in Hyderabad,Bangalore,Chennai,Pune Experience Required : 3 Years to 10 Years Educational Qualification : Other Bachelor Degree Skill Snowflake,Python,ETL,Hadoop,Big data etc.
View all details

Data Scientist

GSN Solutions LLC

  • 7 - 13 yrs
  • Pune
Statistics Python Machine Learning Business Intelligence Tableau SQL Programming Big Data Hadoop Private API JSON YAML XML Hypotheses Validation Tests Work From Home
What we expect? Excellent knowledge of Statistics, especially distributions, likelihood estimators, etc. Hands on experience with one of these programming languages: Python / R Extensive knowledge of Machine Learning concepts and techniques, along with knowledge of various algorithms and their use cases Experience in using Business Intelligence (BI) tools like Tableau (or equivalent) Hands on experience in using SQL programming with Microsoft SQL Server / MySQL Familiarity in using Big Data / Hadoop (or equivalent) Familiarity with a wide variety of data sources, including public or private APIs and standard data formats, like JSON, YAML and XML Ability to design and implement Validation Tests, Hypotheses with supporting documentation Excellent documentation skills, including advanced use of Microsoft Excel, Word, and PowerPoint Ability to work independently with minimum input or assistance Ability to effectively communicate with all levels of audience business and technical Good in spoken and written English languageResponsibilities Gather and analyze data, using various types of analytics and reporting tools to detect patterns, trends, and relationships in data sets Establish best practices for collecting data using analysis tools, and interpreting data Process, cleanse, and verify the integrity of data used for analysis Perform ad-hoc analysis and present results in a clear manner using appropriate medium Process / mine huge volumes of structured, semi-structured and unstructured data using state-of-the-art methods to derive meaningful insights in an appropriate format Recommend tools, techniques, and practices across organization to enhance knowledge Create complex predictive models using ML techniques and relevant tools
View all details