A student from various disciplines can Apply. Interested and eligible candidates can read more details below.
Primary Responsibilities:
- Deliver professional level technical work in support of the development of company products, tools, platforms and services related to big data, typically for an external customer or end user
- Operates within established methodologies, procedures, and guidelines
- Gather and analyse business requirements
- Be able to analyse Big Data infrastructure and solutions
- Develop technical specifications and design documents
- Develop, enhance, integrate, test and deploy Big Data Solutions
- Be able to solve technical and strategic challenges using innovative approaches
- Comply with the terms and conditions of the employment contract, company policies and procedures, and any and all directives (such as, but not limited to, transfer and/or re-assignment to different work locations, change in teams and/or work shifts, policies in regards to flexibility of work benefits and/or work environment, alternative work arrangements, and other decisions that may arise due to the changing business environment). The Company may adopt, vary or rescind these policies and directives in its absolute discretion and without any limitation (implied or otherwise) on its ability to do so
For Freshers & Software Jobs Visit our website
Apply for Pwc | Apply for oracle |
Apply for Zoho | Apply for Infosys |
Apply for Wipro | Apply for Genpact |
Apply for Apple | Apply for siemens |
Apply for Volvo | Apply for Microsoft |
Apply for Deloitte | Apply for solugenix |
Apply for Logitech | Apply for Accenture |
Apply for Mindtree | Apply for Capgemini |
Apply for J.P Morgan | Apply for ServiceNow |
Apply for Developer associate | Apply for Hexaware Technology |
Required Qualifications:
- B.Tech/B.E degree in applicable area of expertise or equivalent experience
- 3 – 5 years of Bigdata Engineering IT work experience as a developer, engineer or similar role (corporate employer full time)
- 3+ years of hands-on experience in technical stack
- Spark/Scala
- Hadoop Ecosystem (Hive, HBase, HDFS)
- any cloud (Azure/AWS/GCP)
- Apache Kafka
- RDBMS (Sound SQL knowledge)
- NoSQL
- Shell Scripting
- 2+ years of work Exposure in Azure/ AWS/ GCP cloud
- Experience in developing/coding software components in Bigdata, that includes various tools like Map Reduce, Hive, Swoop, HBase, Pig, Kafka, Spark, Scala
- Experience in Spark/Scala, pig, Oozie, streaming, No SQL DB
- Experience in writing solid unit tests and integration tests
- Expertise and deep understanding on designing and developing Big Data Pipelines
- Exceptional understanding of SRE, SLA, SLI, Resiliency, etc.
- Product/Platform centric view on things. Customer centric approach
- Ability to create operational dashboard, run retro analysis on issues, etc.
- Proven ability to perform POCs or run end to end operations for an emerging platform/Migrations/Upgrades
- Willing to learn Cloud Technologies, big data technologies and develop technical depth in quick timeframe
- Available for rotational weekend support
- Ability to work independently in a fast paced, agile environment
Preferred Qualifications:
- Experience working on a Big Data Engineering Data Pipelines
- Experience operating large-scale data systems
- Knowledge of modern DevOps practices, including automated testing, continuous integration, automated deployments and CICD pipelines
- Understanding of IAC, GitHub, Jenkins, etc.
- Understanding or work experience on Snowflake
- Opensource contributions to distributed systems
- Ability to work with vendors to resolve issues (or) find workaround solutions
- Proven to be highly productive, self-starter and self-motivated