Pyspark
8 job openings found.
NETCONOMY
In this role, you will build scalable ETL/ELT processes on Google Cloud, optimize BigQuery queries, and implement data architectures and real-time processing using Dataflow and Pub/Sub.
In this role, you will develop powerful data pipelines and ETL/ELT processes using Databricks and Google BigQuery, work on cloud integrations, and foster data modeling in the team to solve complex data challenges.
Erste Bank
In this position, you will build data solutions on the Databricks platform using PySpark and Delta Lake, maintain data pipelines, and optimize data quality through monitoring while mentoring junior team members.
Österreichische Lotterien
In this role, you will develop ML-based solutions, integrate models into existing systems, and optimize workflows while applying MLOps practices like CI/CD and automated monitoring.
Siemens Personaldienstleistungen GmbH
In this role, you will design robust data pipelines on AWS, define data contracts with teams, and enable optimal ML workflows. You will implement CI/CD practices and ensure the scalability and security of the data infrastructure.
Porsche Informatik GmbH
In this role, you will develop data pipelines on Azure Databricks, integrate data from various sources, and optimize performance and monitoring to provide high-quality data for different business departments.
Fronius International GmbH
In this role, you develop AI-based features for products and implement them from prototype to deployment while designing scalable data processing pipelines and monitoring model data.
Porsche Holding GmbH
In this role, you will develop data pipelines on Azure Databricks, optimize their performance, and orchestrate them with Azure Data Factory to provide high-quality data for business departments.
Receive new Pyspark Jobs in Austria by email.
Sportradar Media Services GmbH
In this role, you will develop scalable data models for audience segmentation and measurement while collaborating with modern technologies to provide strategic insights for advertisers.
PwC Österreich
In this role, you will be responsible for designing and implementing scalable data platforms, developing data pipelines using Apache Spark, and optimizing REST APIs for cloud solutions like Azure and AWS.
You design and optimize scalable data platforms, develop cloud-native data pipelines, and implement REST APIs with Python, while acting as a technical sparring partner for process optimization.
In this role, you will develop scalable data models to support audience segmentation and activation, conduct data analyses for strategic insights, and design advanced dashboards for stakeholder engagement.
Leinweb Data Solutions
In this role, you will develop data pipelines, model data structures, and create analytical solutions using Power BI and Microsoft Fabric, while collaborating closely with client teams.
In this role, you will design and operate scalable data platforms, develop ETL processes using Apache Spark, and orchestrate data flows in the cloud or on-premises while implementing everything in Python and modern APIs.
In this role, you will interpret data, develop scalable models, and build advanced dashboards. You will collaborate closely with various teams to create data-driven solutions for audience strategy.