Current jobs related to Azure Databricks - Iselin - Diverse Lynx


  • Iselin, United States E-Solutions Full time

    Role : Lead Databricks Engineer Location : Iselin NJ / New York, NY Onsite Hire type : Contract Must have : Databricks , Azure , spark , pyspark , Azure Data Factory , Python , SQL.Experience with Azure Databricks Medallion, onshore -offshore coordination JD:Strong knowledge on data bricks architecture and tools Have experience of task and wf jobs creations...

  • Databricks Engineer

    2 days ago


    Iselin, New Jersey, United States E-Solutions Full time

    Job Title: Databricks ArchitectWe are seeking a highly skilled Databricks Architect to join our team at E-Solutions. As a key member of our data engineering team, you will be responsible for designing, developing, and maintaining data pipelines using Databricks and Spark.Key Responsibilities:Design and develop data pipelines using Databricks and...


  • Iselin, United States Persistent Systems Full time

    About PersistentWe are a trusted Digital Engineering and Enterprise Modernization partner, combining deep technical expertise and industry experience to help our clients anticipate what’s next. Our offerings and proven solutions create a unique competitive advantage for our clients by giving them the power to see beyond and rise above. We work with many...


  • Iselin, New Jersey, United States Brains Workgroup Full time

    Job Title: ETL DeveloperWe are seeking an experienced ETL Developer to join our team at Brains Workgroup. As an ETL Developer, you will be responsible for designing and implementing highly performant data ingestion pipelines from multiple sources using Azure Databricks.Key Responsibilities:Migrate existing SSIS ETL scripts to Python and develop new ETL...


  • Iselin, New Jersey, United States Brains Workgroup Full time

    ETL Developer Job DescriptionWe are seeking an experienced ETL Developer to join our team at Brains Workgroup. As an ETL Developer, you will be responsible for designing, implementing, and maintaining data pipelines using Python and Snowflake.Key Responsibilities:Migrate existing SSIS ETL scripts to Python and develop new ETL scriptsSupport existing SSIS SQL...

  • Python ETL Engineer

    4 weeks ago


    Iselin, New Jersey, United States Brains Workgroup Full time

    Position Overview:We are seeking a skilled ETL Developer with expertise in Python to support our client's data integration initiatives within the financial sector. This role involves a hybrid work model, requiring two days in the office and three days remote.Key Responsibilities:• Transition existing ETL processes from SSIS to Python, while also developing...

  • Python ETL Engineer

    4 weeks ago


    Iselin, New Jersey, United States Brains Workgroup Full time

    Position Overview:We are seeking a skilled ETL Developer with a strong background in banking or financial services to join our team. This role involves a hybrid work model, requiring two days in the office and three days remote.Contract Duration:This is a long-term contract opportunity with potential for extensions or conversion to a permanent position.Key...

  • Data Architect

    5 days ago


    Iselin, New Jersey, United States Zortech Solutions Full time

    Job DescriptionJob Title: Solution Data ArchitectLocation: Wilmington, NC (Hybrid), Iselin, NJJob Summary:We are seeking a highly skilled and experienced Solution Data Architect to join our team at Zortech Solutions. The ideal candidate will have a strong background in designing and implementing data solutions, with a focus on scalability, efficiency, and...

Azure Databricks

4 months ago


Iselin, United States Diverse Lynx Full time

Develop deep understanding of the data sources, implement data standards, maintain data quality and master data management.• Expert in building Databricks notebooks in extracting the data from various source systems like DB2, Teradata and perform data cleansing, data wrangling, data ETL processing and loading to AZURE SQL DB.• Expert in building Ephemeral Notebooks in Databricks like wrapper, driver and config for processing the data, back feeding the data to DB2 using multiprocessing thread pool.• Expert in developing JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data.• Expert in using Databricks with Azure Data Factory (ADF) to compute large volumes of data.• Performed ETL operations in Azure Databricks by connecting to different relational database source systems using jdbc connectors.• Developed Python scripts to do file validations in Databricks and automated the process using ADF.• Analyzed the SQL scripts and designed it by using Pyspark SQL for faster performance.• Worked on reading and writing multiple data formats like JSON, Parquet, and delta from various sources using Pyspark.• Developed an automated process in Azure cloud which can ingest data daily from web service and load in to Azure SQL DB.• Expert in optimizing the Pyspark jobs to run on different Cluster for faster data processing.• Developed spark applications in python (Pyspark) on distributed environment to load huge number of CSV files with different schema in to Pyspark Dataframes and process them to reload in to Azure SQL DB tables.• Analyzed data where it lives by Mounting Azure Data Lake and Blob to Databricks.• Used Logic App to take decisional actions based on the workflow and developed custom alerts using Azure Data Factory, SQLDB and Logic App.• Developed Databricks ETL pipelines using notebooks, Spark Dataframes, SPARK SQL and python scripting.• Developed Spark applications using Pyspark and Spark-SQL for data extraction, transformation and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.• Good Knowledge and exposure to the Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.• Involved in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.• Expert in understanding current production state of application and determine the impact of new implementation on existing business processes.• Involved in Migration of data from On-prem server to Cloud databases (Azure Synapse Analytics (DW) & Azure SQL DB).• Good Hands on experience in setting up Azure infrastructure like storage accounts, integration runtime, service principal id, and app registrations to enable scalable and optimized utilization of business user analytical requirements in Azure.• Expert in ingesting streaming Digital : Databricks 10 & Above • Develop deep understanding of the data sources, implement data standards, maintain data quality and master data management.• Expert in building Databricks notebooks in extracting the data from various source systems like DB2, Teradata and perform data cleansing, data wrangling, data ETL processing and loading to AZURE SQL DB.• Expert in building Ephemeral Notebooks in Databricks like wrapper, driver and config for processing the data, back feeding the data to DB2 using multiprocessing thread pool.• Expert in developing JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data.• Expert in using Databricks with Azure Data Factory (ADF) to compute large volumes of data.• Performed ETL operations in Azure Databricks by connecting to different relational database source systems using jdbc connectors.• Developed Python scripts to do file validations in Databricks and automated the process using ADF.• Analyzed the SQL scripts and designed it by using Pyspark SQL for faster performance.• Worked on reading and writing multiple data formats like JSON, Parquet, and delta from various sources using Pyspark.• Developed an automated process in Azure cloud which can ingest data daily from web service and load in to Azure SQL DB.• Expert in optimizing the Pyspark jobs to run on different Cluster for faster data processing.• Developed spark applications in python (Pyspark) on distributed environment to load huge number of CSV files with different schema in to Pyspark Dataframes and process them to reload in to Azure SQL DB tables.• Analyzed data where it lives by Mounting Azure Data Lake and Blob to Databricks.• Used Logic App to take decisional actions based on the workflow and developed custom alerts using Azure Data Factory, SQLDB and Logic App.• Developed Databricks ETL pipelines using notebooks, Spark Dataframes, SPARK SQL and python scripting.• Developed Spark applications using Pyspark and Spark-SQL for data extraction, transformation and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.• Good Knowledge and exposure to the Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.• Involved in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.• Expert in understanding current production state of application and determine the impact of new implementation on existing business processes.• Involved in Migration of data from On-prem server to Cloud databases (Azure Synapse Analytics (DW) & Azure SQL DB).• Good Hands on experience in setting up Azure infrastructure like storage accounts, integration runtime, service principal id, and app registrations to enable scalable and optimized utilization of business user analytical requirements in Azure.• Expert in ingesting streaming • Develop deep understanding of the data sources, implement data standards, maintain data quality and master data management.• Expert in building Databricks notebooks in extracting the data from various source systems like DB2, Teradata and perform data cleansing, data wrangling, data ETL processing and loading to AZURE SQL DB.• Expert in building Ephemeral Notebooks in Databricks like wrapper, driver and config for processing the data, back feeding the data to DB2 using multiprocessing thread pool.• Expert in developing JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data.• Expert in using Databricks with Azure Data Factory (ADF) to compute large volumes of data.• Performed ETL operations in Azure Databricks by connecting to different relational database source systems using jdbc connectors.• Developed Python scripts to do file validations in Databricks and automated the process using ADF.• Analyzed the SQL scripts and designed it by using Pyspark SQL for faster performance.• Worked on reading and writing multiple data formats like JSON, Parquet, and delta from various sources using Pyspark.• Developed an automated process in Azure cloud which can ingest data daily from web service and load in to Azure SQL DB.• Expert in optimizing the Pyspark jobs to run on different Cluster for faster data processing.• Developed spark applications in python (Pyspark) on distributed environment to load huge number of CSV files with different schema in to Pyspark Dataframes and process them to reload in to Azure SQL DB tables.• Analyzed data where it lives by Mounting Azure Data Lake and Blob to Databricks.• Used Logic App to take decisional actions based on the workflow and developed custom alerts using Azure Data Factory, SQLDB and Logic App.• Developed Databricks ETL pipelines using notebooks, Spark Dataframes, SPARK SQL and python scripting.• Developed Spark applications using Pyspark and Spark-SQL for data extraction, transformation and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.• Good Knowledge and exposure to the Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.• Involved in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.• Expert in understanding current production state of application and determine the impact of new implementation on existing business processes.• Involved in Migration of data from On-prem server to Cloud databases (Azure Synapse Analytics (DW) & Azure SQL DB).• Good Hands on experience in setting up Azure infrastructure like storage accounts, integration runtime, service principal id, and app registrations to enable scalable and optimized utilization of business user analytical requirements in Azure.• Expert in ingesting streaming • Develop deep understanding of the data sources, implement data standards, maintain data quality and master data management.• Expert in building Databricks notebooks in extracting the data from various source systems like DB2, Teradata and perform data cleansing, data wrangling, data ETL processing and loading to AZURE SQL DB.• Expert in building Ephemeral Notebooks in Databricks like wrapper, driver and config for processing the data, back feeding the data to DB2 using multiprocessing thread pool.• Expert in developing JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data.• Expert in using Databricks with Azure Data Factory (ADF) to compute large volumes of data.• Performed ETL operations in Azure Databricks by connecting to different relational database source systems using jdbc connectors.• Developed Python scripts to do file validations in Databricks and automated the process using ADF.• Analyzed the SQL scripts and designed it by using Pyspark SQL for faster performance.• Worked on reading and writing multiple data formats like JSON, Parquet, and delta from various sources using Pyspark.• Developed an automated process in Azure cloud which can ingest data daily from web service and load in to Azure SQL DB.• Expert in optimizing the Pyspark jobs to run on different Cluster for faster data processing.• Developed spark applications in python (Pyspark) on distributed environment to load huge number of CSV files with different schema in to Pyspark Dataframes and process them to reload in to Azure SQL DB tables.• Analyzed data where it lives by Mounting Azure Data Lake and Blob to Databricks.• Used Logic App to take decisional actions based on the workflow and developed custom alerts using Azure Data Factory, SQLDB and Logic App.• Developed Databricks ETL pipelines using notebooks, Spark Dataframes, SPARK SQL and python scripting.• Developed Spark applications using Pyspark and Spark-SQL for data extraction, transfo

Diverse Lynx LLC is an Equal Employment Opportunity employer. All qualified applicants will receive due consideration for employment without any discrimination. All applicants will be evaluated solely on the basis of their ability, competence and their proven capability to perform the functions outlined in the corresponding role. We promote and support a diverse workforce across all levels in the company.