Your browser does not support javascript! Please enable it, otherwise web will not work for you.

Data Warehouse Informatica & ETL Job in @ Uplers

Home >

 Data Warehouse Informatica & ETL Job in

Job Description

    5 - 7 Years 1 Opening Bangalore Role description Role Proficiency: This role requires proficiency in data pipeline development including coding and testing data pipelines for ingesting wrangling transforming and joining data from various sources. Must be skilled in ETL tools such as Informatica Glue Databricks and DataProc with coding expertise in Python PySpark and SQL. Works independently and has a deep understanding of data warehousing solutions including Snowflake BigQuery Lakehouse and Delta Lake. Capable of calculating costs and understanding performance issues related to data solutions. Outcomes: Act creatively to develop pipelines and applications by selecting appropriate technical options optimizing application development maintenance and performance using design patterns and reusing proven solutions.rnInterpret requirements to create optimal architecture and design developing solutions in accordance with specifications. Document and communicate milestones/stages for end-to-end delivery. Code adhering to best coding standards debug and test solutions to deliver best-in-class quality. Perform performance tuning of code and align it with the appropriate infrastructure to optimize efficiency. Validate results with user representatives integrating the overall solution seamlessly. Develop and manage data storage solutions including relational databases NoSQL databases and data lakes. Stay updated on the latest trends and best practices in data engineering cloud technologies and big data tools. Influence and improve customer satisfaction through effective data solutions. Measures of Outcomes: Adherence to engineering processes and standards Adherence to schedule / timelines Adhere to SLAs where applicable # of defects post delivery # of non-compliance issues Reduction of reoccurrence of known defects Quickly turnaround production bugs Completion of applicable technical/domain certifications Completion of all mandatory training requirements Efficiency improvements in data pipelines (e.g. reduced resource consumption faster run times). Average time to detect respond to and resolve pipeline failures or data issues. Number of data security incidents or compliance breaches. Outputs Expected: Code Development: Develop data processing code independently ensuring it meets performance and scalability requirements. Define coding standards templates and checklists. Review code for team members and peers. Documentation: Create and review templates checklists guidelines and standards for design processes and development. Create and review deliverable documents including design documents architecture documents infrastructure costing business requirements source-target mappings test cases and results. Configuration: Define and govern the configuration management plan. Ensure compliance within the team. Testing: Review and create unit test cases scenarios and execution plans. Review the test plan and test strategy developed by the testing team. Provide clarifications and support to the testing team as needed. Domain Relevance: Advise data engineers on the design and development of features and components demonstrating a deeper understanding of business needs. Learn about customer domains to identify opportunities for value addition. Complete relevant domain certifications to enhance expertise. Project Management: Manage the delivery of modules effectively. Defect Management: Perform root cause analysis (RCA) and mitigation of defects. Identify defect trends and take proactive measures to improve quality. Estimation: Create and provide input for effort and size estimation for projects. Knowledge Management: Consume and contribute to project-related documents SharePoint libraries and client universities. Review reusable documents created by the team. Release Management: Execute and monitor the release process to ensure smooth transitions. Design Contribution: Contribute to the creation of high-level design (HLD) low-level design (LLD) and system architecture for applications business components and data models. Customer Interface: Clarify requirements and provide guidance to the development team. Present design options to customers and conduct product demonstrations. Team Management: Set FAST goals and provide constructive feedback. Understand team members" aspirations and provide guidance and opportunities for growth. Ensure team engagement in projects and initiatives. Certifications: Obtain relevant domain and technology certifications to stay competitive and informed. Skill Examples: Proficiency in SQL Python or other programming languages used for data manipulation. Experience with ETL tools such as Apache Airflow Talend Informatica AWS Glue Dataproc and Azure ADF. Hands-on experience with cloud platforms like AWS Azure or Google Cloud particularly with data-related services (e.g. AWS Glue BigQuery). Conduct tests on data pipelines and evaluate results against data quality and performance specifications. Experience in performance tuning of data processes. Expertise in designing and optimizing data warehouses for cost efficiency. Ability to apply and optimize data models for efficient storage retrieval and processing of large datasets. Capacity to clearly explain and communicate design and development aspects to customers. Ability to estimate time and resource requirements for developing and debugging features or components. Knowledge Examples: Knowledge Examples Knowledge of various ETL services offered by cloud providers including Apache PySpark AWS Glue GCP DataProc/DataFlow Azure ADF and ADLF. Proficiency in SQL for analytics including windowing functions. Understanding of data schemas and models relevant to various business contexts. Familiarity with domain-related data and its implications. Expertise in data warehousing optimization techniques. Knowledge of data security concepts and best practices. Familiarity with design patterns and frameworks in data engineering. Additional Comments: UST is looking for Data warehousing Developers with Informatica: Responsibilities Creates and supports the ETL process to extract the data from source systems and place it into the data warehouse Performs data warehouse design and testing, including data design, database architecture, metadata, and repository creation Designs, develops, and maintains relational databases for data storage and data mining Works with business and IT stakeholders to optimize reporting information by defining business requirements and populating the data warehouse table structure Creates data warehouse schematics and layouts; assists with developing capacity planning Recommends improvements and applies standards, processes, and procedures for data warehousing systems Seeks continuous data warehouse performance improvement and ensures security of data Creates and implements effective metrics and monitoring processes May lead small projects with manageable risks and resource requirements Acts as a resource and impacts the quality of work for colleagues with less experience Performs other duties as assigned Requirements 4+ years Data Warehouse or relevant experience Experience in Kimball dimensional modeling 4+ years ETL, database, and scheduling experience Demonstrated ability to quickly learn new technologies 1+ years of Data Analysis experience Advanced ability to solve complex problems, drawing upon creativity, experience, and multiple sources of information Advanced interpersonal skills, including the ability to build consensus Advanced communication skills, including the ability to communicate complex information in straightforward manner Regular and predictable attendance Bachelors degree or equivalent education/ experience Preferred Insurance and/or reinsurance industry knowledge Masters degree and/or LOMA certification Experience as an ETL team lead Technical Skills Required Informatica Power Center & Data Validation Option Enterprise Scheduler (Tidal preferred) Erwin Data Modeler Database technologies (one or more of the following) o Oracle SQL & PL/SQL o Vertica Preferred Talend Tidal Enterprise Scheduler SQL Serve Exp : 5-7 Years Location : Bangalore Role Overview UST is seeking a Data Warehousing Developer with expertise in Informatica, ETL development, and data modeling. This role involves designing and optimizing data warehouse solutions, developing ETL processes, and collaborating with business and IT stakeholders to enhance reporting and data integration. Key Responsibilities Develop and maintain ETL processes to extract, transform, and load data into the data warehouse. Perform data warehouse design and testing, including database architecture, metadata management, and repository creation. Design and develop relational databases for data storage and mining. Collaborate with business and IT teams to define data requirements and optimize reporting. Create data warehouse schematics, layouts, and capacity planning. Apply and recommend best practices, standards, and processes for data warehousing solutions. Ensure data security, performance optimization, and monitoring. Lead small-scale projects with minimal risks and resource requirements. Provide guidance and mentorship to less experienced colleagues. Perform additional data analysis and problem-solving tasks as needed. Requirements 4+ years of experience in Data Warehousing and ETL development. Strong knowledge of Kimball dimensional modeling. 4+ years of experience with ETL tools, databases, and scheduling systems. Experience in data analysis and reporting optimization. Ability to quickly learn new technologies and apply them effectively. Strong problem-solving skills and ability to work with multiple data sources. Excellent communication and interpersonal skills for collaboration. Bachelors degree or equivalent experience. Preferred Qualifications Insurance or reinsurance industry experience. Masters degree and/or LOMA certification. Experience leading ETL development teams. Technical Skills Required Informatica PowerCenter & Data Validation Option Enterprise Scheduler (Tidal preferred) Erwin Data Modeler Database technologies: Oracle SQL & PL/SQL Vertica Preferred Technical Skills Talend Tidal Enterprise Scheduler SQL Server Skills ETL-Data Warehouse,Informatica Power Center About UST UST is a global digital transformation solutions provider. For more than 20 years, UST has worked side by side with the worlds best companies to make a real impact through transformation. Powered by technology, inspired by people and led by purpose, UST partners with their clients from design to operation. With deep domain expertise and a future-proof philosophy, UST embeds innovation and agility into their clients organizations. With over 30,000 employees in 30 countries, UST builds for boundless impacttouching billions of lives in the process.,

Employement Category:

Employement Type: Full time
Industry: IT Services & Consulting
Role Category: Not Specified
Functional Area: Not Specified
Role/Responsibilies: Data Warehouse Informatica & ETL Job in

Contact Details:

Company: UST Global
Location(s): Other Karnataka

+ View Contactajax loader


Keyskills:   Informatica Power Center SQL Python

 Fraud Alert to job seekers!

₹ Not Specified

Similar positions

S&C Global Network - AI - CG&S

  • Accenture
  • 5 to 9 Yrs
  • Other Haryana
  • 4 hours ago
₹ Not Specified

System Security Engineer 3 Job in Oracle

  • Oracle
  • 5 to 9 Yrs
  • Other Karnataka
  • 5 hours ago
₹ Not Specified

Uplers

Uplers is a one-stop digital services company delivering end-to-end web, design, digital marketing, and email production services to businesses and agencies across 52+ Nations. Backed by a team of 550+ digital expert.

Plugin template missing! Fix or contact support.