Data Engineer Mid

Detalles de la oferta

¡¡We are looking for your talent!! ?

**Data Engineer Mid**

**? **The desired profile should have at least 3 years hands-on experience in designing, establishing, and maintaining data management and storing systems. Skilled in collecting, processing, cleaning, and deploying large datasets, understanding ER data models, and integrating with multiple data sources. Efficient in analyzing, communicating, and proposing different ways of building Data Warehouses, Data Lakes, End-to-End Pipelines, and Big Data solutions to clients, either in batch or streaming strategies.

It will be very important that you have the following skills/experience:
**English B2 or higher**

**Technical Proficiencies**:

- SQL:
Data Definition Language, Data Manipulation Language, Intermediate/advanced queries for analytical purpose, Subqueries, CTEs, Data types, Joins with business rules applied, Grouping and Aggregates for business metrics, Indexing and optimizing queries for efficient ETL process, Stored Procedures for transforming and preparing data, SSMS, DBeaver
- Python:
Experience in object-oriented programming, Management and processing datasets, Use of variables, lists, dictionaries and tuples, Conditional and iterating functions, Optimization of memory consumption, Structures and data types, Data ingestion through various structured and semi-structured data sources, Knowledge of libraries such as pandas, numpy, sqlalchemy, Must have good practices when writing code
- Databricks / Pyspark:
Intermediate knowledge in

Understanding of narrow and wide transformations, actions, and lazy evaluations

How DataFrames are transformed, executed, and optimized in Spark

Use DataFrame API to explore, preprocess, join, and ingest data in Spark

Use Delta Lake to improve the quality and performance of data pipelines

Use SQL and Python to write production data pipelines to extract, transform, and load data into

tables and views in the Lakehouse

Understand the most common performance problems associated with data ingestion and how to

mitigate them

Monitor Spark UI: Jobs, Stages, Tasks, Storage, Environment, Executors, and Execution Plans

Configure a Spark cluster for maximum performance given specific job requirements

Configure Databricks to access Blob, ADL, SAS, user tokens, Secret Scopes and Azure Key Vault

Configure governance solutions through Unity Catalog and Delta Sharing

Use Delta Live Tables to manage an end-to-end pipeline with unit and integrations test
- Azure:
Intermediate knowledge in

Azure Storage Account:
Provision Azure Blob Storage or Azure Data Lake instances

Build efficient file systems for storing data into folders with static or parametrized names, considering possible security rules and risks

Experience identifying use cases for open-source file formats like parquet, AVRO, ORC

Understanding optimized column-oriented file formats vs optimized row-oriented file formats

Implementing security configurations through Access Keys, SAS, AAD, RBAC, ACLs

Azure Data Factory:
Provision Azure Data Factory instances

Use Azure IR, Self-Hosted IR, Azure-SSIS to establish connections to distinct data sources

Use of Copy or Polybase activities for loading data

Build efficient and optimized ADF Pipelines using linked services, datasets, parameters, triggers, data movement activities, data transformation activities, control flow activities and mapping data flows

Build Incremental and Re-Processing Loads

** What benefits will you have?**

WELLNESS: We will promote your integral wellbeing through personal, professional and economic balance. Our legal and additional benefits will help you achieve it.
LET'S RELEASE YOUR POWER: You will have the opportunity to specialize in a comprehensive manner in different areas and technologies, thus achieving an interdisciplinary development. We will push you to take on new challenges and surpass yourself.
WE CREATE NEW THINGS: We like to think outside the box. You will have the space, confidence and freedom to create and the training required to achieve it.
WE GROW TOGETHER: You will participate in cutting-edge, multinational technology projects with foreign teams.

**Where will you do it?**

We are a great team working in a remote scheme, we are flexible and structured; providing the necessary equipment to work with and internal communication tools that facilitate our operation and that of our clients.

If you meet most of the requirements and you are interested in the profile do not hesitate to apply, our Talent team will contact you!

Become derevian & develop your superpower!


Salario Nominal: A convenir

Fuente: Whatjobs_Ppc

Requisitos

Qa Automation (Selenium, C#) Tester

**Role: QA Automation (Selenium, C#) Tester** **Location: REMOTE (6 Months remote then Hybrid)** **Type of Hire: Full Time** **Start Date: Immediate** **...


Desde Epsilon Staffing - Jalisco

Publicado 5 days ago

Fogonero Carreta Técnica Mantenimiiento Sueldo

**POR EXPANSIÓN SOLICITAMOS**: **FOGONERO | TÉCNICO - CON EXPERIENCIA EN CALDERAS** **Perfil**: - Preparatoria concluida (indispensable), Ingeniería a fin ...


Desde Caramelos De La Rosa S.A. De C.V - Jalisco

Publicado 5 days ago

Analista Del Sistema De Gestión De Inocuidad

Empresa 100% mexicana con más de 40 años de experiência en la alimentación y nutrición animal está buscando tu talento para el área de Sistema de Gestión de ...


Desde Pronua - Jalisco

Publicado 5 days ago

Soporte Técnico, Analista Gps

**Somos una empresa líder dedicados a la seguridad GPS y estamos en búsqueda de talento** - **Requisitos**: Edad: 18 a 50 años Sexo: Indistinto Estado civi...


Desde Gpsi Latinoamerica, S. De R. L. De C.V. - Jalisco

Publicado 5 days ago

Built at: 2024-11-06T00:42:32.054Z