Get new jobs by email
- ...the Databricks Lakehouse Platform building Bronze/Silver/Gold layers using Delta Lake and Delta Live Tables (DLT). Strong in PySpark, Spark SQL, and Databricks Workflows for orchestration. Proficient with Unity Catalog for governance, lineage, and access control, and...SuggestedRemote work
- ...years of experience building data pipelines in cloud environments ~4+ years of experience with Big Data technologies (e.g., Spark, Hadoop) and cloud architecture ~3+ years of experience with reporting and analytics tools (e.g., Tableau, Power BI) ~ Hands...SuggestedLong term contract
- ...Data bricks Lakehouse Platform - building Bronze/Silver/Gold layers using Delta Lake and Delta Live Tables (DLT). Strong in PySpark, Spark SQL, and Data bricks Workflows for orchestration. Proficient with Unity Catalog for governance, lineage, and access control, and...SuggestedRemote work
- ...Proficient SQL; relational (PostgreSQL, MySQL) and NoSQL (MongoDB) experience Cloud & Big Data: AWS/Azure/Google Cloud Platform, Spark, Hadoop, scalable storage (S3, Blob, HDFS) Ready to Apply? Take the next step in your data engineering career with this exciting...SuggestedContract work
$69 - $74 per hour
...cloud computing experience. 3+ years of experience with Google BigQuery. Preferred Qualifications 5+ years of experience with Spark and Python. 3+ years of experience with secure DevOps practices and deployment automation in cloud environments. 3+ years of experience...SuggestedHourly pay- ...Must be Local to Reston, NO RELO - OnSite 3 days a week. Top 5 Technical Skills: Python (Big Data Pipeline) AWS Hadoop, Spark, Hive EMR Terraform Job Description: Strong Python development to build a big-data pipeline for data processing and analysis Need strong experience...SuggestedContract workWork experience placementLocal area3 days per week
- ...and SQL ~ Hands-on experience with AWS services (S3, Glue, Redshift, Lambda, EMR) ~ Experience with big data tools like Apache Spark, Hadoop, or Kafka ~ Knowledge of Machine Learning concepts and model lifecycle ~ Experience with ETL/ELT frameworks and data...SuggestedRemote work
- ...Petabyte-level data systems. Experience with cloud-native data tools and architectures (e.g., Redshift, Glue, Airflow, Apache Spark). Proficient in automated testing frameworks (PyTest, Playwright or Jest) and testing best practices. Experience developing...SuggestedLong term contractRemote work
$75 - $80 per hour
...Qualifications Knowledge of MLOps practices and ML pipelines Experience with data platforms such as Snowflake, Databricks, or Spark Familiarity with AI frameworks like TensorFlow or PyTorch Cloud certifications such as AWS Certified Solutions Architect...SuggestedFull timeContract workTemporary workWork experience placementImmediate startWorldwideFlexible hours- ...platform engineer to design and build a containerized API layer that abstracts and governs interactions with engines such as Apache Spark through a well-defined API contract. This role focuses on building platform capabilities, not simply consuming existing data tools enabling...SuggestedContract work
- ...related discipline Minimum of 5+ years of relevant industry experience Minimum 5 years of experience developing in Hadoop echo system ( Spark, PySpark, MapReduce, Hive, Impala ) Minimum 5 years of experience with common application frameworks (JEE Spring Boot, Struts,...SuggestedRemote work
- ...and SQL ~ Hands-on experience with AWS services (S3, Glue, Redshift, Lambda, EMR) ~ Experience with big data tools like Apache Spark, Hadoop, or Kafka ~ Knowledge of Machine Learning concepts and model lifecycle ~ Experience with ETL/ELT frameworks and data...SuggestedRemote work
- ...with: Cleansing, deduplication, parsing, and merging of high-volume datasets ~ Parsing EBCDIC/COBOL-formatted VSAM files using Spark-Cobol Library ~ Connecting to Db2 databases using JDBC drivers for ingestion For applications and inquiries, contact: hirings...SuggestedRemote work
- ..., improving database performance, and developing quality audits. Proficient with the Azure stack, including Synapse Analytics, Spark/Python, Azure SQL, Azure Data Factory (ADF), Kusto, among others. Knowledgeable in ETL tools, particularly Azure Data Services,...SuggestedContract work
- ...Google Cloud Platform) ~ Familiarity with REST APIs and microservices architecture ~ Experience with data processing tools (Spark, Pandas, etc.) Preferred Qualifications Experience working in data security / cybersecurity domain Knowledge of MLOps...SuggestedContract work
- ...Responsibilities : Expertise in big data processing, Core Java and Apache spark particularly within finance domain. Should have a strong experience working with financial instruments, market risk and large-scale distributed computing systems. Develop...
- ...data mesh and data federation as the architectural core is highly desirable. Experience in big data technologies like PrestoTrino | Spark Flink| Airflow Prefect| Red Panda Kafka| Iceberg Delta Lake| Snowflake Databricks| MemGraph Neo4J as well as modern security tooling...Work experience placement
- ...develop scalable data transformation pipelines using DBT Cloud Architect and implement Databricks-based data solutions (Delta Lake, Spark) Build and optimize data models (star/snowflake schemas) for analytics Develop ETL/ELT pipelines using modern data stack...
- ...equivalent): AWS or Azure Data Platform Services, Postgres/Oracle/DB2, Collibra, Databricks, Delta Lake, Python, Snowflake, ETL tools (Spark, etc.), CI/CD pipelines supporting Data Lakehouse Expertise in real time and batch data ingestion architectures (Kafka/Event...Hourly payContract work
- ...Experience with cloud platforms and containerization (Docker, Kubernetes). ~ Familiarity with data engineering tools (e.g., Airflow, Spark) and ML Ops frameworks. ~ Solid understanding of software engineering principles and DevOps practices. ~ Ability to...
- ...~7+ years of experience in Software Engineering ~4+ years building big data pipelines ~4+ years of experience with: Apache Spark (PySpark / Spark SQL) Hive and Iceberg tables SQL / SQL Server or other RDBMS ~ Strong programming experience in: Python PySpark...Contract workRemote work
- ...resume and contact details. Core Technical Skills: SQL Server SQL Server Integration Services Azure Synapse Spark Microsoft Fabric Required Skills & Experience: ~ Bachelor's or Master's degree in computer science, information...Contract workWork at officeRemote work
- ...ingestion, transformation, and data reliability Real-time systems Build streaming solutions using Kafka or Azure Event Hubs Use Spark Structured Streaming for high-volume data processing API and integration Lead integrations using Spring Boot, Dell Boomi, and...
- ...Preferred qualifications: Experience with streaming API s like Kafka Understanding of Big Data/Data Lake technologies (Spark, Hadoop, Databricks etc) Understanding of Design patterns and clean coding Understanding of technical aspects of Analytic applications...Local area
- ...generation, and intelligent automation. - Work with large structured and unstructured datasets using tools like Pandas, NumPy, and Spark. - Implement models in production environments using Python, TensorFlow, PyTorch, or scikit-learn. - Conduct exploratory...
- ...RapidFuzz, jellyfish Experience with LLM APIs (OpenAI, Anthropic) and prompt pipelines Strong SQL skills and experience with Spark or Dask for distributed processing Familiarity with vector databases and embedding-based retrieval Experience with ML...Work experience placement
- ...Forest, PSI/KSI drift detection, model retraining. Vision Models: ViT, Grad-CAM, explainability methods. Data Engineering: Spark, Python, Delta Lake, REST APIs. Governance & Security: RBAC, lineage, masking, compliance policies. Soft Skills:...Remote work
- ...Terraform Containerization & orchestration with Docker and Kubernetes Data Engineering & Analytics Advanced experience with Spark , SQL , data modeling , and distributed data processing Proficient in Snowflake for data warehousing and ELT pipelines...ApprenticeshipTraineeship
- ...team is seeking a Data Engineer experienced in implementing modern data solutions in Azure, with strong hands-on skills in Databricks, Spark, Python, and cloud-based DataOps practices. The Data Engineer will analyze, design, and develop data products, pipelines, and...
- ...workloads Knowledge of MLOps practices and ML pipelines Experience with data platforms such as Snowflake, Databricks, or Spark Familiarity with AI frameworks like TensorFlow or PyTorch Cloud certifications such as AWS Certified Solutions Architect or...Full timeContract workImmediate start