Ogy Cangoz
- 570 877 1832
- ocangoz2@gmail.com
- github
- New York

Senior Analytics Engineer with over 6 years of experience. Specializing in optimizing large-scale processes to ensure swift data delivery and meet cost-cutting objectives. As a motivated engineer, I have consistently pursued opportunities for ongoing professional development to refine both technical and interpersonal skills, allowing me to take on a wide variety of projects.
Work Experiences
Senior Analytics Engineer
Business Intelligence Engineer II
As a Senior Analytics Engineer with Warner Bros Discovery, my role revolves around Looker reporting initiatives. I support Ad Sales stakeholders with reporting needs for all WBD products. My main objective is the conversion of raw data into self-serving analytics solutions that empower data consumers.
- Reduced a daily ETL process from 1.5 hr. to 40 min by implementing best practices, including UDFs, CTEs, numeric joins, and parallel execution, saving the Ad Sales org about 40% on execution time daily on the 2 largest ETL’s.
- Secured management approval for the implementation of OPEN AI API, leading to enhancements in Data Quality Checks, Documentation, Code Reviews/Recommendations, and the creation of an LLM Chat Bot for our Technical Support Slack channel, boosting team efficiency, scalability, and code security.
- Developed a unified Revenue Pacing metric for both legacy WarnerMedia and Discovery campaigns, working meticulously with stakeholders before the launch of our unified product “MAX”.
- Consistently writing scalable Apache Airflow DAGs, as well as maintaining 50+ DAGs daily. I have implemented data quality checks within Airflow which indicate the location of the issue when DAGs fail.
- Established leadership role in technical reporting by collaborating with key stakeholders like CNN, Bleacher Report, and Looker users in weekly meetings to address reporting, business needs, and data quality.
Senior Data Analyst
Data Analyst
As a Senior Data Analyst on the Decision Sciences team at 3Q Digital, I was a member of a specialized analytics team that developed data science models and communicated findings to enterprise clients.
- Identified optimal media spend allocation across channels by leveraging statistical forecast models (Prophet Model), informing our highest paying client on how best to optimize spending each quarter.
- Developed Python models using statistics to inform clients on budget allocation, incrementality, and forecasting.
Planning Analyst
Eat Club is a food-tech company that provides corporate lunch. As a Planning Analyst, I was responsible for demand planning and overseeing the supply chain process, as well as managing inventory of 100,000+ units.
- Time-series forecasted to manage supply across 3 New York City distribution centers, making crucial daily decisions in a fast-paced environment that directly impacted the company’s profitability.
- Leveraged Looker for analytics reporting and dashboard creation for the non – analytical leadership team.
Recent Projects
Innovative Ad Sales Pacing Calculation
- Seamlessly integrated final output into Looker, showcasing technical and business logic proficiency.
- Tasked with developing a unified campaign pacing report between legacy WarnerMedia and legacy Discovery.
- This project required meticulous coordination with stakeholders to gather:
- Reporting requirements
- Project deadlines
- Data sources and development of methodology
- Seamlessly integrated final output into Looker, showcasing technical and business logic proficiency.
- This project was crucial for reporting needs ahead of the MAX product launch.
Data Quality Enhancement
- Developed a process of creating Data Quality checks within our workflows.
- When a data quality issue is present, the query would fail by a ‘division by zero’ and airflow would trigger a slack alert to the specified engineer.
- This process has been built to scale across all workflows and consists of the following checks (at minimum):
- ROW_COUNT – Is row count within standard deviation?
- COMPLETENESS – Has the data been processed?
- DUPLICATION – Are there duplicate records?
- VALIDATION – Does impressions from raw_layer = reporting_layer = source UI?
Self-Service Analytics and AI Implementation:
- Leveraging OpenAI API, I developed a process to generate documentation for all workflows, cataloging:
- joins, architecture, data sources, transformations, and ETL workflows.
- Partnered with Principal Engineer to develop and AI-generated support bot for the Revenue Analytics team. This bot acts as a self-service resource to help users answer reporting questions, resolve issues, and generate new reports.
Reengineer primary workflows for WBD:
- Due to consolidating WarnerMedia and Discovery data, I was tasked with optimizing our primary workflows to process data faster.
- By removing inefficient joins, consolidating steps within the ETL, running tasks concurrently, and processing smaller amounts of data at a greater frequency, I optimized our workflow and exposed the most recently available data within our warehouse to our end users.