Sr Data Engineer

Posted 3 Days Ago
Be an Early Applicant
Bengaluru, Bengaluru Urban, Karnataka
Senior level
Insurance
The Role
The Sr Data Engineer will design and architect data lake solutions using AWS and Snowflake, manage the software development lifecycle of data, ensure data quality, and collaborate with cross-functional teams. Responsibilities include requirements gathering, documentation, and applying best practices for data governance and management.
Summary Generated by Built In

What if the work you did every day could impact the lives of people you know? Or all of humanity?

At Illumina, we are expanding access to genomic technology to realize health equity for billions of people around the world. Our efforts enable life-changing discoveries that are transforming human health through the early detection and diagnosis of diseases and new treatment options for patients.

Working at Illumina means being part of something bigger than yourself. Every person, in every role, has the opportunity to make a difference. Surrounded by extraordinary people, inspiring leaders, and world changing projects, you will do more and become more than you ever thought possible.

JOB DESCRIPTION:

The position is an exciting opportunity to be a member of the Data Integration & Analytics team within GIS Application & Platform Services Dept. The team’s scope includes data services on enterprise data platforms like Snowflake cloud data platform, SAP HANA analytics and Denodo data virtualization. The team is responsible for managing the full software development lifecycle of data, its data quality, and operations. This role will support strategic solutions like the enterprise data lake on AWS/ Snowflake and the enterprise data warehouse on Snowflake. The role is responsible for collaborating with cross functional teams, planning, and coordinating requirements, providing data engineering services and helping build trust in the data being managed.

JOB DUTIES:

  • Design and architect data lake solutions on AWS S3 and Snowflake, considering scalability, automation, security, and performance requirements.
  • Responsible for architecting, building, and optimizing data lake to store, process and analyze large volumes of structure and unstructured data.
  • Requirements gathering and analysis, development planning and co-ordination, data testing, documentation and ensuring high levels of data quality is maintained.
  • documentation.
  • Work with project leads, stakeholders, and business SMEs to define technical specifications to develop data integration requirements, code design and functional and technical documentation. 
  • Work with specifications to design, develop, and maintain data infrastructure. 
  • Work with cross functional teams and process owners on the development of test cases and scripts, test models and solutions to verify that requirements are met and ensuring high levels of data quality. 
  • Promote and apply quality assurance best practices.
  • Understand and comply with data governance and compliance practices as defined for risk management. This includes data encryption practices, RBAC and security policies.
  • Promote and apply metadata management best practices supporting enterprise data catalogs.
  • Excellent problem-solving skills and ability to troubleshoot complex data engineering issues.
  • Benchmark application operational performance periodically, track (metrics) and fix issues.
  • Support change and release management processes.
  • Support incident and response management including problem solving and root cause analysis, documentation.
  • Support automation and on-call processes (Tier 1 / Tier 2).  

SPECIFIC SKILLS OR OTHER REQUIREMENTS:

Requires 5 + years of experience with

Primary Experience: Data Lakes and Data Integrations

Data Engineering:

  • AWS: 4+ years of required expertise with AWS services like S3, Glue, Lambda, EMR, Athena and proficiency in programming languages/ tools like Python, Scala, SQL.
  • Apache Software: 2+ years of required expertise in data processing with Spark & Flink, message broking using Kafka, orchestration using Airflow, processing high data volumes in open table formats like Iceberg.
  • Snowflake: 3+ years of required expertise with Snowflake SnowSQL, Snowpipe, (integrated with AWS S3), Streams and Tasks, Stored Procedures (expertise working with merge statements), Compute and Storage usage optimizations, Performance optimizations.
  • HVR (Fivetran): Near Real-time Data Replications with HVR is a plus.
  • SnapLogic: Data integrations with SnapLogic integration platform is a plus.
  • Certifications: AWS and Snowflake data engineering certifications is a plus.

​Data Platforms:

  • Required: AWS S3 data lake store platform, Snowflake cloud data platform
  • Preferred: Denodo, SAP HANA

Programming/ Scripting:

  • Snowflake: SQL Scripting (Snowpipes, Tasks, Streams, Merge Statements, Stored Procedures, Functions, Security Policies – DDM, Row-Access), SQL Scripting, Python
  • AWS: Apache spark, Python

Data Orchestration:

  • Preferred: Control-M, Apache Airflow

Cloud Storage Platforms:

  • Required: Amazon Web Services
  • Preferred: Microsoft Azure

Source Systems:

  • Required: Knowledge and experience integrating data from SAP ERP (On premises, Cloud), Salesforce CRM, Workday, Planisware, Team Center PLM, Relational databases, REST APIs
  • Preferred: ServiceNow, MFG MES, IOT/ Sensor data platforms, Unstructured data, Other SaaS Applications

Data Operations: Excellent understanding of Data Ops practices for data management.

Stakeholder Engagement: Ability to take the lead and drive project activities collaborating with Analytics stakeholders and ensure the requirement is completed.

Solution Design: Good understanding of end-to-end solution architecture and design practices, ability to document solutions (maintain diagrams)

Governance: Good understanding of working with companies having regulated systems and processes for data. Adherence to data protection practices using tagging, security policies and data security (object-level, column-level, row-level).

Data Warehousing: Good understand on how data onboarded in the data lakes is used to support data warehouses. Impact of risks and change to a data warehouse. Fundamental concepts of dimensional modeling, experience working on data warehouse solutions, requirement gathering, design & build.

Data As-A-Product: Preferred knowledge and experience working with data treated as data products. Illumina is following a hybrid data mesh architecture which promotes data as a product for data lifecycle management.

Operating Systems: Windows, Linux

EDUCATION & EXPERIENCE:

Bachelor’s degree equivalent in Computer Science/Engineering or equivalent degree.

#LI-HYBRID

#illuminacareers


Illumina believes that everyone has the ability to make an impact, and we are proud to be an equal opportunity employer committed to providing employment opportunity regardless of sex, race, creed, color, gender, religion, marital status, domestic partner status, age, national origin or ancestry, physical or mental disability, medical condition, sexual orientation, pregnancy, military or veteran status, citizenship status, and genetic information.

Top Skills

Python
Scala
SQL
The Company
Mechanicsburg, , PA
317 Employees
On-site Workplace
Year Founded: 1885

What We Do

RCM&D is ranked among the top independent insurance advisory firms in the United States. Our specialized teams provide strategic solutions and consulting for risk management, insurance and employee benefits. Leveraging more than 135 years of experience and strong local, national and global reach, we partner with you to meet all of your business objectives.

RCM&D Self-Insured Services Company (SISCO) is a Third Party Administrator and a trusted partner in claim management, consulting and advisory services for clients that are self-insured, fully insured, or utilize captive or high deductible programs. Our expertise coupled with our technological resources ensure budget certainty, reduced costs and the overall efficiency of your claim process.

Established in 2020 through the strategic merger of RCM&D and Oswald Companies, Unison Risk Advisors (a dba of the JBO Holding Company) is majority employee owned. On June 20, 2023, NSI became the third partner of Unison Risk Advisors. Unison Risk Advisors is made up of these three primary operating companies and all their existing subsidiaries and sub-brands. Today it is one of the largest privately held and employee-owned insurance brokerage and risk management firms in the U.S.

Similar Jobs

Atlassian Logo Atlassian

Senior Data Engineer

Cloud • Information Technology • Productivity • Security • Software • App development • Automation
Remote
Bengaluru, Karnataka, IND
11000 Employees

Takeda Logo Takeda

Senior Data Engineer

Healthtech • Software • Analytics • Biotech • Pharmaceutical • Manufacturing
Hybrid
Bengaluru, Karnataka, IND
50000 Employees

Cisco Meraki Logo Cisco Meraki

Senior Data Engineer

Hardware • Information Technology • Security • Software • Cybersecurity • Conversational AI
Easy Apply
Remote
Hybrid
Bengaluru, Karnataka, IND
3000 Employees

Atlassian Logo Atlassian

Senior Data Engineer

Cloud • Information Technology • Productivity • Security • Software • App development • Automation
Remote
Bengaluru, Karnataka, IND
11000 Employees

Similar Companies Hiring

Flume Health Thumbnail
Software • Insurance • Healthtech
US
22 Employees
Spark Advisors Thumbnail
Software • Sales • Other • Insurance • Healthtech
New York, NY
73 Employees
MassMutual India Thumbnail
Insurance • Information Technology • Fintech • Financial Services • Big Data
Hyderabad, Telangana

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account