Superset (nice to have)
Power BI (nice to have)
Tableau (nice to have)
Apache Spark (junior)
Scandinavian software industry company where cutting edge technology solutions are being developed for automotive, telecom and industrial domains.
Greatly skilled at ADAS, telematics, analytics and machine learning solutions, software and algorithms. The team consists of about 400 devs with its own R&D branch and products.
Perform Data Engineering tasks required to analyse data from CDIT+D clusters. This includes setting-up data pipelines, storing data in client premises and transforming collected data to data structures optimized for efficient browsing, execute testing, perform troubleshooting of delivered data pipelines.
Excellent Python knowledge (as a plus combined with libs like : pandas, NumPy)
Familiarity with Apache Spark and column-oriented data formats (parquet / orc)
Familiarity with partitioning, indexing and retention strategies
Good knowledge of SQL in the context of working with >
1mln rows tables
Ability to work with hdfs and s3 storages
Experience in Kubernetes for containerization
Experience in orchestrating data pipelines with Airflow
nice to have : Azure in the context of data processing (data bricks, adls, data factory, synapse)
Nice to have :
Knowledge of BIs for data visualization (Tableau, PowerBI, Superset)
20.000-29.000 PLN B2B depends on experience
An experienced and ambitious community that is looking for new makers for our challenging and changing customer projects.
An opportunity to develop your skills in a superb group of people and in varying projects : our own product-related and done for our customers.
Wage highly correlated with your skills and mindset. And fair.
A deep and flexible organization that looks after its own.