Big Data Developer (Media Measurement sphere)

Company Background

Our client is the company which pioneers the future of cross-platform media measurement, arming organizations with the insights they need to make decisions with confidence. Central to this aim are our people who work together to simplify the complex on behalf of our clients & partners.


It is a trusted partner for planning, transacting and evaluating media across platforms. With a data footprint that combines digital, linear TV, over-the-top and theatrical viewership intelligence with advanced audience insights, its platform allows media buyers and sellers to quantify their multiscreen behavior and make business decisions with confidence.

Project Description

You’ll be responsible for building next generation data warehouse together with migrating the data and processes from existing one (which would be deprecated later this year). The data is main supplier for a broad range of clients and products, including industry leading ad agencies, national television networks and other products. As a member of this fast-moving team you’ll have large impact on the evolution and adoption of the data processing as well as on the success of the business. It’s worth mentioning that this company processes and stores dozens of petabytes of data which is coming from TV/Web and their current infrastructure processes more than 15 bln requests per day.


Existing data warehouse is organized in Greenplum database and different job/procedures implemented via different set of tools/programming languages. The goal is to migrate it to AWS/Snowflake together with new warehouse design, improved usability and performance optimizations. Jobs should be reviewed, optimized and implemented using Spark (additional tools to be defined).

What You’ll Do
  • Work within an agile team to develop new ETL processes and DW design;
  • Recommend and implement creative solutions for performance improvement;
  • Increase scalability and maintainability to support rapid usage growth;
  • Collaborate openly with stakeholders and clients to continuously improve the data usage experience;
Apache Spark
Apache Parquet
Job Requirements
  • Experience with SQL-like databases and data warehouses;
  • Having worked with DW design and establishing ETL (ELT) processes would be beneficial to help with successful migration and have the users switch to new system;
  • Experience with Java/Scala/Python languages would be a benefit in future design of DW;
  • Strong communication skills (written and verbal) along with a track record of success delivering large software projects;
  • Demonstrated knowledge of commonly used software engineering concepts, practices, and procedures;
What Do We Offer
  • 100+ technical and non-technical trainings to develop and expand your horizons;
  • Internal meetings and conferences (we like sharing our expertise);
  • English courses (both with Russian-speaking teachers and native speakers);
  • Free medical insurance in 80+ private clinics;
  • Sports activities (soccer, basketball, hockey, tennis, swimming, etc.) and corporate events (corporate parties, team building, happy hours, canoeing, etc.);
  • Free hot lunches in the office.
Feel free to contact us!
We will answer you!