Analyze and organize raw data to build scalable data systems and pipelines, handling data ingestion, cleansing, transformation, mapping, import/export, and integration using Python. Perform end-to-end ETL processes including data extraction, transformation, loading and integration to support enterprise data architectures such as data warehouses and operational data stores. Migrate data from multiple sources including Hive, Teradata, DB2, SQL Server, Excel, and flat files to Google BigQuery through GCS buckets, ensuring data consistency and performance optimization. Automate CI/CD pipelines by creating and scheduling workflows, developed Python wrapper scripts to trigger workflows based on refresh frequency. Create and manage DAGs in Airflow using Cloud Composer to schedule, orchestrate, and monitor workflows across various data pipelines. Build and maintain BigQuery authorized views to implement row-level security, supporting secure data access and compliance requirements. Work under supervision. Travel and/or Relocation to various unanticipated client sites throughout USA is required.
Bachelor's degree in Computer Science/IT/IS/ Engineering (any) or closely related field with Twelve (12) months of experience in the job offered or as an IT Consultant or Analyst or Programmer or Developer or Engineer or closely related field.
Experience of Twelve (12) months working with Python and Hadoop is required. Travel and/or relocation is required to unanticipated client sites within USA. International travel is not required. The frequency of travel is currently not known as it depends on the client and project requirement that cannot be currently anticipated. Employer provides Information technology services to various clients in USA and hence implementing projects will require such travel.
Standard Company Benefits.