Consulting and Services company specializing in Big Data and Cloud Platforms

Read More

Data Science

+

R, Matlab, Spark-Mlib, H2O, Python, Deep Learning Frameworks

Read More

Advanced Analytics

+

Visualization Tools, Tableau, Spotfire, D3.js

Read More

Big Data

+

Cloudera, Hortonworks, MapR, Hadoop tools, NoSQL Databases, Scala, Java, C++, Python, Perl, Bash, SQL, H-SQL

Read More

Cloud Solutions

+

Google Cloud Platform, Amazon AWS and Azure

Read More

Airflow

+

Airflow

Read More

Data Science

+

Data Science

R, Matlab, Spark-Mlib, H2O, Python, Deep Learning Frameworks

Read More

Advanced Analytics

+

Advanced Analytics

Visualization Tools, Tableau, Spotfire, D3.js

Read More

Big Data

+

Big Data

Cloudera, Hortonworks, MapR, Hadoop tools, NoSQL Databases, Scala, Java, C++, Python, Perl, Bash, SQL, H-SQL

Read More

Cloud Solutions

+

Cloud Solutions

Google Cloud Platform, Amazon AWS and Azure

Read More

Airflow

+

Airflow

Full support of modern orchestration tool

Read More

Our main differentiation is a flexible approach to constantly changing business requirements and needs.

DS Stream is an IT consulting company specializing in Data Engineering and Data Science. We use a wide range of Big Data technologies to deliver our services to a variety of industries including CPG, Telecommunications and Banking.

Our highly qualified engineers and data scientists provide insightful expertise which help us deliver real added value to our customers.

Big Data Solutions

  • Large scale Big Data solutions.
  • Infrastructure Design and Data Architectures.
  • Development, maintenance, testing and evaluation of Big Data solutions.
  • Data Flow automation with the usage of robust data wrangling techniques.
  • Process Automation: Oozie, Airflow.

Data Science

  • Development of sophisticated analytical programs using machine learning and statistical methods.
  • Inventing new algorithms to solve business challenges.
  • Exploration and examination of data to determine hidden patterns, trends and/or spot new opportunities.
  • Predictive and descriptive analytics, data mining, big data analytics.

Cloud Solutions

  • Big Data Storage & Databases – management of scalable Big Data and archive solutions.
  • IoT – usage of cloud services to gather, process and store data.
  • Computing Services – serverless computing services using Lambda Architecture.
  • Process Automation – automated e2e data flow with self-adjusting ability of the computation power.

Advanced Analytics

  • Expertise knowledge in Fraud Detection systems, Cyber Security, Marketing Campaigns.
  • Data Exploration – manual and automatic techniques to explore data and find business insights.
  • Hypothesis generation and evaluation.
  • Forecasting future events and behaviors using multi-genre analytics.
  • Predictive and descriptive analytics, data mining, big data analytics.

Airflow

  • Airflow deployment on Virtual Machines, Company on-premises machines with or without Docker, with all kinds of available executors, Kubernetes, Celery, Local, etc.
  • Designing and creating Airflow workflows(DAGs), moving current solution(ETL processes/Workflows) to Apache Airflow.
  • Investigating and resolving performance, security issues also all kinds of problems that not always are dependent on Airflow , why it was restarted, why Airflow is not starting up, why tasks are not getting executed, why there is high CPU and memory USAGE on Airflow Machine.
  • Upgraiding and support AirFlow platform.
  • Integrating Airflow with external services, writing custom operators, hoooks, fixing bugs in existing ones, integrating with company on-premises autenthication services(LDAP, ActiveDirectory, etc.).
  • Existing code review whether it is in compliance with Airflow good practices, whether all kinds of Airflow features are used correctly and not replaced by custom unnecessary ones.

We hire people, not positions!

With a lot of experience in Big Data
You will be responsible for:
  • Designing and implementing Big Data systems (e.g. data warehouses, data lakes, streaming applications, control systems etc.) while working with multiple Open Source technologies
  • Building high-quality Big Data products and providing an ongoing support
  • Applying statistical methods to organize, analyze, and interpret data
  • Detecting anomalies and recommending data quality solutions
  • Conducting data-driven analyses and reports; applying qualitative and quantitative data analysis methods; preparing statistical and non-statistical data exploration.
You will need one of these:
  • Good programming skills (Python, SQL, Java are preferred)
  • Prior experience in designing and implementing large scale systems based on Big Data technologies such as Cloudera and Hortonworks
  • Hands-on experience in using machine learning and implementing data mining algorithms
  • Prior experience with Hadoop, Spark, Hive, Kafka or other related technologies
  • Knowledge of Open Source projects is a plus
For the not so experienced in Big Data
You will be responsible for:
  • Analyzing what sort of data is required for dashboards, modeling, statistical analysis and data mining
  • Conducting data-driven analyses and reports in cooperation with ETL experts; applying qualitative and quantitative data analysis methods; preparing statistical and non-statistical data exploration
  • Designing, analyzing and validating data requirements as well as project deliverables
  • Acting as a liaison between technical developers and business users to provide model tuning support
You will need to have:
  • Degree or specialization in Technical Science/Mathematics/Statistics/Quantitative Methods (3rd-5th year of study)
  • Programming and statistical/data analytical skills (hands-on experience in R, SAS 4GL, SQL, Python, Spark, Java are preferred)
  • Basic knowledge of database design and data architecture
  • Basic experience in system automation
  • Basic experience in data mining


Application form

Consent for data processing can be withdrawn at any time, which will not affect the lawfulness of processing based on consent before its withdrawal.

According to the article 13. (1) and (2) of General Data Protection Regulation (EU) 2016/679 - hereinafter referred to as the GDPR, please, be informed that:
    1) The data controller of your personal data is DS STREAM sp. z o. o., with registered seat in Warsaw (03 - 743), at Grochowska 306/308;
    2) Your personal data will be processed with the purpose of recruitment process pursuant to the art. 6 (1)(a) GDPR, i.e. based on your freely given consent;
    3) Your personal data will be stored until the recruitment process is completed;
    4) You have the right to access your data and the right to rectify, delete, restrict the data processing, the right to data transfer, the right to object, the right to withdraw your consent to processing at any time. The withdrawal of consent you shall send in writing to postal address of the data controller or by e-mail to the e-mail address career@dsstream.com;
    5) You have the right to file a complaint before Polish Data Protection Authority, if you think that the processing of your personal data violates the provisions of the GDPR;
    6) Sharing with us your personal data is voluntary, but necessary for the purposes related to the recruitment process.
Send us a message

* These fields are required.

Registered Address

Grochowska 306/308, 03-840 Warszawa

DS Stream sp. z o.o. z siedzibą w Warszawie przy Grochowska 306/308, 03-840 Warszawa, zarejestrowaną w Rejestrze Przedsiębiorców Krajowego Rejestru Sądowego prowadzonego przez Sąd Rejonowy dla m. st. Warszawy w Warszawie, XII Wydział Gospodarczy, pod numerem KRS 0000652483, NIP: 5272789783, REGON: 366076146