I have over two decades of deep experience in the technology space with a unique blend of leadership, systems design, programming, distributed systems, data engineering, business intelligence, and devops. I bring a proven record of solving high impact business problems with a focus on quality and cost effectiveness.
Experience
2023 — Now
New York City Metropolitan Area
* Defined technical vision for new unified semantic customer data platform to provide heterogeneous data access to client data across all of the company’s product offerings and sold vision to CPO/CTO and other leaders. This new platform was incorporated as a key item in the company’s 3 year plan because it was designed to scale the business by significantly reducing custom solutions work and speeding up onboarding new customers by 3x.
* Designed and built a novel relationship discovery approach to automatically identify connections in client data.
* Built production grade data platform ahead of schedule and with better than company average quality that exposed a dynamic GraphQL interface based on customer defined semantic metadata utilizing F#, ASP.NET Core, PostgreSQL, HotChocolate, and Kubernetes.
* Utilized Claude Code and GitHub CoPilot for agentic AI assisted development.
* Wrote efficient fault tolerant background job processing logic for huge file ingestions with flat memory usage.
* Backend storage system was made pluggable with PostgreSQL/SQLite/DuckDB backends available.
Implemented performance testing and regression testing using randomized data generation.
* Grew team by recruiting to fill critical software engineering gaps for the team.
2019 — 2023
New York City Metropolitan Area
* Saved $MMs by skillfully managing the relationship between AWS Redshift/PostgreSQL teams and Movable Ink.
* Reduced company-wide infrastructure costs by over 15% by partnering with the CTO to implement a multitude of optimizations & technical architectural improvements. This included migrating a multi-terabyte self-hosted PostgreSQL to AWS Aurora to handle the critical holiday rush and improving system-wide message structure to reduce data volume by 60%, both of which improved quality, scalability, performance, and latency.
* Provided CFO/FinOps actionable reporting by replacing patchwork data swamp with comprehensive multi-region data and BI strategy and a data platform acting as a single source of truth. This involved GDPR compliance, business plan development, updating company processes, and coordinating with teams across the company.
* Convinced CTO of benefits and led migration from NSQ to a durable log-based messaging system (Pulsar selected but evaluated Kafka also). This included implementing a data-loss protection wrapper for Pulsar Node.js client, adding disk-based durability and creating a Golang message bus repeater to ease the transition.
* Ensured scalability and reliability of databases and data processing systems as the subject matter expert for data engineering, stream processing, distributed computing, concurrency for company. This included applying a novel caching solution to read-only SQL queries, reducing load by 80% & latency by 90%.
* Grew data team from inception to 7 team members supporting all of the business' data and analytics needs. This involved defining standards and processes and training new team members including creating tooling such as a unit and integration test framework for Apache Airflow using Python/AWS Redshift/EMR/S3.
* Designed Redshift schemas and ETL jobs supporting 3 PB scale and tens of billions of messages per day, optimizing for reliability and query latency to enable real-time reporting and analytics.
2015 — Now
2015 — Now
* Implemented several enhancements to the Pony programming language as a core contributor
* Improved efficiency of memory allocation, garbage collection, and cycle detection by up to 10000x to enable more uses of Pony in high performance apps
* Developed a dynamic suspend and resume capability for the scheduler threads and improved performance & reliability of the Epoll based AsyncIO (ASIO) subsystem & the buffered IO package in the standard library to enable uses of Pony in high efficiency apps
* Enhanced observability and debuggability of Pony by implementing detailed memory, GC, and system / application CPU usage metrics collection and systematic/deterministic/repeatable testing ability for the runtime
* Improved cross platform reliability by creating CI pipelines to compile for non-x86-64 archs using QEMU
2019 — 2019
2019 — 2019
New York City Metropolitan Area
* Advised clients on streaming data platforms, systems architecture, and event driven architecture
* Evangelized and supported clients on Confluent Platform and Kafka
* Educated clients on security, automation, performance, monitoring, reliability, connectors
2015 — 2018
2015 — 2018
New York City Metropolitan Area
* Core contributor to Wallaroo, an open source, high-throughput, low-latency distributed real-time event processing framework written from the ground up in the Pony language
* Wallaroo Up tool for automating the installation and configuration of Wallaroo system on multiple Linux distros
* Makefile system for building, testing, and deploying Wallaroo, as well as testing documentation code samples
* Perf tuning and optimization, as many OS/Environment compatibility fixes, including: Nix/OSX LTO support, bcc/eBPF, perf, systemtap, dtrace, flame graphs for identifying and resolving bottlenecks in Pony, C, and Java
* Orchestration framework using Terraform, Ansible, Docker, and Mesos, supporting multiple user profiles
Education
New York University
Bachelor of Arts
1999 — 2002
Bronx High School of Science
High School Diploma
1994 — 1998