Ricky K.

Developer Relations (DevRel)

Matia

OtherPLG AssistedConsultative
Posted by Ricky K.•

Overview

You're evangelizing Matia's unified DataOps approach to data engineers and analytics engineers. This means writing technical content, giving talks at data conferences, building example projects, and being active in data engineering communities (dbt Slack, Locally Optimistic, r/dataengineering). Your success is measured by community growth, content engagement, and ultimately how much qualified inbound pipeline you generate.


Role Snapshot

AspectDetails
Role TypeDeveloper Relations / Advocacy
Sales MotionCommunity-led / Content-driven inbound
Deal ComplexityN/A (you enable sales, don't close)
Sales CycleN/A
Deal SizeN/A
Quota (est.)Community metrics + inbound pipeline contribution

Company Context

Stage: Early stage (41 employees)

Size: 41 employees

Growth: Building out GTM motion—you're the first DevRel hire

Market Position: Category consolidator competing against established point solutions. Need to build credibility and awareness in data engineering community.


GTM Reality

Your Role in Pipeline:

  • You don't sell directly, but everything you do should make selling easier
  • Create content that SDRs can reference in outreach
  • Generate inbound interest from engineers who discover Matia through your content/talks
  • Build relationships that turn into warm intros for sales team

What DevRel Means Here:

  • First DevRel hire = you're building this function from scratch
  • You'll report to leadership (likely CEO/CTO given company size)
  • No team yet—you're a one-person show initially

Competitive Landscape

Main Competitors: Fivetran, Airbyte, Monte Carlo, Metaplane, Atlan, Collibra, Census, Hightouch

How They Differentiate: Unified platform vs. tool sprawl—you need to make this message resonate with practitioners.

Your Job: Make "unified DataOps" a concept people understand and want. Right now, data teams don't think in those terms—they think about ETL, then observability, then catalog as separate problems.

Community Challenges:

  • Data engineers are skeptical of new vendors
  • "Best-of-breed vs. platform" debates are real
  • Switching costs are high—you need to make the pain of current state obvious

What You'll Actually Do

Time Breakdown

Content Creation (40%) | Community Engagement (30%) | Speaking/Events (20%) | Internal Collaboration (10%)

Key Activities

  • Technical Content: Writing blog posts, tutorials, and guides on data engineering best practices. Topics like "Building Reliable Data Pipelines," "Modern Data Stack Anti-Patterns," "How to Debug Data Quality Issues." You need to be legitimately helpful, not just promotional.
  • Community Presence: Active in dbt Slack, Data Engineering Slack, Locally Optimistic, r/dataengineering. Answering questions, sharing insights, building relationships. This is slow-burn work—you're not selling, just being helpful and visible.
  • Speaking: Submitting to data conferences (Data Council, Coalesce, local meetups). Giving talks on data observability, pipeline reliability, etc. Most talks won't be about Matia directly—you're establishing credibility.
  • Example Projects: Building open-source examples, sample pipelines, reference architectures. Showing how Matia fits into modern data stacks.
  • Documentation: Working with product/engineering to make docs actually useful. DevRel often owns developer experience.
  • Sales Enablement: Creating demo repos, technical FAQs, competitive positioning docs. Joining calls when prospects have deep technical questions.

The Honest Reality

What's Hard

  • Measuring Impact: DevRel ROI is squishy. You can track community growth and content views, but tying your work directly to revenue takes months/years.
  • Balancing Promotion vs. Education: Data engineers hate being sold to. If your content feels like marketing, you lose credibility fast. But you still need to drive awareness and pipeline.
  • Building from Zero: You're the first DevRel. No playbook, no processes, no existing community. You're figuring out what works.
  • Technical Depth Required: You need to actually understand data engineering. If you can't code, build pipelines, or debug data quality issues, you won't be credible.
  • Long Feedback Loops: A blog post you write today might generate a lead in 6 months. A conference talk might pay off a year later. Patience required.
  • Early-Stage Chaos: At 41 people, priorities shift quickly. Product might change. Messaging might pivot. You adapt or get frustrated.

What Success Looks Like

  • Growing an engaged community (Slack members, newsletter subscribers, GitHub stars)
  • Content that ranks well and gets shared in data engineering circles
  • Speaking slots at recognized data conferences
  • Inbound demo requests that mention your content
  • Engineers saying "I've seen your posts" on sales calls
  • Building repeatable content/event engines that scale

Who You're Influencing

Primary Audience:

  • Data Engineers (IC level)
  • Analytics Engineers
  • Staff/Principal Engineers (technical decision-makers)
  • Heads of Data (managers who follow technical trends)

What They Care About:

  • Practical Solutions: They want actionable advice, not vendor fluff
  • Technical Credibility: Can you actually build data systems or are you just marketing?
  • Community: They trust peers more than vendors—you need to be seen as a peer
  • Staying Current: New tools/approaches emerge constantly—they want to know what actually matters

Requirements

  • Hands-on data engineering experience (can write SQL, understand data pipelines, have built ETL/ELT workflows)
  • Strong technical writing skills—you'll be publishing frequently
  • Public speaking experience or willingness to learn (conferences, meetups, webinars)
  • Active in data engineering communities already (credibility exists before you start)
  • Comfortable with ambiguity—you're building this role from scratch
  • Self-directed—no manager will tell you what content to create or which events to attend
  • Understanding of modern data stack (dbt, Snowflake/Databricks, Fivetran/Airbyte, etc.)
  • Genuine interest in data engineering as a discipline, not just as a job