Principal Software Engineer - Telemetry

Wal-Mart Sunnyvale, CA
Position Summary

* You will be developing data collection, distribution, and delivery systems that meet the realtime, near realtime and batched information needs of the Business.

* You'll work on bringing telemetry datasets in our data lake environment.

* You will help in design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

* Excellent written and verbal communication skills for diverse audiences on engineering subject matter

* Ability to document requirements, data lineage, subject matter in both business and technical terminology.

* Guide and learn from other team members.

* Demonstrated ability to transform business requirements to code, specific analytical reports and tools

* This role will involve coding, analytical modeling, root cause analysis, investigation, debugging, testing and collaboration with the business partners, product managers other engineering team.

* Be knowledgeable in all aspects of designing and constructing data warehouses and operational data stores in technologies like Teradata and Hadoop.

* Design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

* Ruthlessly prioritize and tweak where needed to ensure business needs are being met consistently.

* Your passion, communication skills, and ability to work well with others are skills that will be assessed and are critical to you being selected for this position.

* Promote and support company policies, procedures, mission, values, and standards of ethics and integrity

* Provide supervision and development opportunities for associates

* Work with enterprise architects, product owners, data scientists and engineers and business to bring big data and data science R&D projects into Production.

* Ability to learn, adopt, and apply new technologies.

Position Summary

* You will be developing data collection, distribution, and delivery systems that meet the realtime, near realtime and batched information needs of the Business.

* You'll work on bringing telemetry datasets in our data lake environment.

* You will help in design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

* Excellent written and verbal communication skills for diverse audiences on engineering subject matter

* Ability to document requirements, data lineage, subject matter in both business and technical terminology.

* Guide and learn from other team members.

* Demonstrated ability to transform business requirements to code, specific analytical reports and tools

* This role will involve coding, analytical modeling, root cause analysis, investigation, debugging, testing and collaboration with the business partners, product managers other engineering team.

* Be knowledgeable in all aspects of designing and constructing data warehouses and operational data stores in technologies like Teradata and Hadoop.

* Design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

* Ruthlessly prioritize and tweak where needed to ensure business needs are being met consistently.

* Your passion, communication skills, and ability to work well with others are skills that will be assessed and are critical to you being selected for this position.

* Promote and support company policies, procedures, mission, values, and standards of ethics and integrity

* Provide supervision and development opportunities for associates

* Work with enterprise architects, product owners, data scientists and engineers and business to bring big data and data science R&D projects into Production.

* Ability to learn, adopt, and apply new technologies.

Position Description

You will be developing data collection, distribution, and delivery systems that meet the realtime, near realtime and batched information needs of the Business.

You'll work on bringing telemetry datasets in our data lake environment.

You will help in design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

Excellent written and verbal communication skills for diverse audiences on engineering subject matter

Ability to document requirements, data lineage, subject matter in both business and technical terminology.

Guide and learn from other team members.

Demonstrated ability to transform business requirements to code, specific analytical reports and tools

This role will involve coding, analytical modeling, root cause analysis, investigation, debugging, testing and collaboration with the business partners, product managers other engineering team.

Be knowledgeable in all aspects of designing and constructing data warehouses and operational data stores in technologies like Teradata and Hadoop.

Design of new systems or the redesign of existing systems to meet business requirements, changing needs, or newer technology

Ruthlessly prioritize and tweak where needed to ensure business needs are being met consistently.

Your passion, communication skills, and ability to work well with others are skills that will be assessed and are critical to you being selected for this position.

Promote and support company policies, procedures, mission, values, and standards of ethics and integrity

Provide supervision and development opportunities for associates

Work with enterprise architects, product owners, data scientists and engineers and business to bring big data and data science R&D projects into Production.

Ability to learn, adopt, and apply new technologies.

Minimum Qualifications

Very Strong engineering skills. Should have an analytical approach and have good programming skills.

BS/MS in Computer Science or equivalent, with 12 or more years in software engineering, design and architecture

Demonstrated background in developing software systems and excellent SQL skills

Knowledge/experience on Physical Design and Implementation

Advanced SQL (preferably Teradata)

Experience working with large data sets, experience working with distributed computing (MapReduce, Hadoop, Hive, Pig, Apache Spark, etc.).

Strong Hadoop scripting skills to process petabytes of data

Experience in Unix/Linux shell scripting or similar programming/scripting knowledge

Experience in ETL processes

Real time data ingestion (Kafka)

Hands on experience with Cassandra

Real time data ingestion with Spark

Flexible, adaptable, and able to autonomously manage multiple tasks in dynamic, fastpaced,iterative development environment with short turnaround times

Additional Preferred Qualifications

Fluency in Java, JavaScript, Go, Bash, Ansible

Deep understanding and experience with Kafka, Cassandra, Solr/ElasticSearch

Deep understanding of HTTP semantics

Experience in data science & machine learning

Employment Type

Full Time