Important Information
- Experience: 7–10+ years
- Job Mode: Full-time
- Work Mode: Hybrid
Job Summary
- Assemble large, complex datasets that meet functional and non-functional business requirements.
- Plan, create, and maintain data architectures aligned with business goals.
- Create and maintain optimal data pipeline architecture, focusing on automation and scalability.
- Identify, design, and implement internal process improvements, including automation of manual processes and optimization of data delivery.
- Propose infrastructure for optimal extraction, transformation, and loading (ETL) of data from diverse sources using SQL and Big Data technologies.
- Continuously audit data management systems to ensure performance, address breaches or gaps, and report findings to stakeholders.
- Recommend analytics tools to generate actionable insights into business performance metrics, including customer acquisition and operational efficiency.
- Collaborate with stakeholders (executives, product teams, and data teams) to resolve technical issues and support data infrastructure needs.
- Build and maintain strong relationships with senior stakeholders to help them leverage Big Data technologies for business solutions.
Responsibilities and Duties
- Assemble, optimize, and maintain large datasets tailored to business needs.
- Design and implement scalable, high-quality data architectures and pipelines.
- Automate workflows, optimize performance, and ensure scalability in infrastructure design.
- Conduct continuous performance audits of data systems and implement improvements as needed.
- Design tools to deliver actionable insights for business intelligence and analytics.
- Collaborate with cross-functional teams to address technical issues and enhance data operations.
- Support data migrations, including integration with platforms like MS Dynamics CRM or SharePoint.
- Actively participate in Agile delivery frameworks (Scrum, DSDM) to ensure quality results.
Qualifications and Skills
- Education: BS/MS in Computer Science, Engineering, Information Technology, or related field with programming experience.
- Proven experience (7–10+ years) in engineering, database modeling, design, and architecture for large-scale analytics projects.
- Expertise in SQL and relational database management, as well as Big Data technologies (Apache Spark, Databricks, Kafka, Hadoop).
- Deep knowledge of modern data architectures (e.g., Lambda architecture, Streaming, Delta Lake).
- Experience with data pipeline tools (Azure Data Factory, Airflow) and Business Intelligence tools (SSAS, Power BI, Tableau).
- Familiarity with cloud services (Azure, AWS).
- Proficiency in programming languages such as Python, R, C#, or Java.
- Knowledge of Data Science, Machine Learning, and Artificial Intelligence trends.
- Strong understanding of industry best practices in data design, integration, and architecture.
- Experience working with Agile methodologies (Scrum, DSDM).
- Excellent English communication skills, both written and spoken.
Role-specific Requirements
- Extensive experience building and optimizing Big Data pipelines and architectures.
- Knowledge of Business Intelligence, analytics, and reporting technologies.
- Experience with data migrations and platforms such as MS Dynamics CRM and SharePoint.
- Strong knowledge of data trends, modern architectures, and scalable design.
- Customer-centric approach to explain technical concepts to non-technical stakeholders.
- Strong communication and collaboration skills in an international and virtual team setting.
- Proven ability to deliver quality results and foster strong client relationships.
Technologies
- Big Data: Apache Spark, Databricks, Snowflake, Kafka, Hadoop
- Data Pipeline Tools: Azure Data Factory, Airflow
- Business Intelligence Tools: SSAS, Power BI, Tableau
- Cloud Services: Azure, AWS
- Programming Languages: Python, R, C#, Java
Skillset Competencies
- Advanced SQL and Big Data pipeline optimization.
- Expertise in modern data architectures and ETL processes.
- Strong data migration and integration experience.
- Proficiency in analytics and reporting technologies.
- Excellent problem-solving, negotiation, and communication skills.
- Ability to work effectively in cross-functional, international teams.
- Strong client relationship management and quality delivery focus.
About Encora
Encora is a trusted partner for digital engineering and modernization, working with some of the world’s leading enterprises and digital-native companies. With over 9,000 experts in 47+ offices worldwide, Encora offers expertise in areas such as Product Engineering, Cloud Services, Data & Analytics, AI & LLM Engineering, and more. At Encora, hiring is based on skills and qualifications, embracing diversity and inclusion regardless of age, gender, nationality, or background.
Top Skills
What We Do
Headquartered in Santa Clara, California, and backed by renowned private equity firms Advent International and Warburg Pincus, Encora is the preferred technology modernization and innovation partner to some of the world’s leading enterprise companies. It provides award-winning digital engineering services including Product Engineering & Development, Cloud Services, Quality Engineering, DevSecOps, Data & Analytics, Digital Experience, Cybersecurity, and AI & LLM Engineering. Encora's deep cluster vertical capabilities extend across diverse industries, including HiTech, Healthcare & Life Sciences, Retail & CPG, Energy & Utilities, Banking Financial Services & Insurance, Travel, Hospitality & Logistics, Telecom & Media, Automotive, and other specialized industries.
With over 9,000 associates in 47+ offices and delivery centers across the U.S., Canada, Latin America, Europe, India, and Southeast Asia, Encora delivers nearshore agility to clients anywhere in the world, coupled with expertise at scale in India. Encora’s Cloud-first, Data-first, AI-first approach enables clients to create differentiated enterprise value through technology