About Moon
An ambitious and independent stealth SaaS company incubated by Home Organizers, a market leader with decades of proven success in designing and delivering exceptional, innovative home organization solutions through its subsidiaries Closet World, Closets by Design, Brio Water Technology, and others. Backed by their deep industry experience and a commitment to be Home Organizer’s critical SaaS provider for its 6000+ employees, our team is building innovative solutions to solve universal problems that most businesses face — yet are not addressed by a single, unified tool.
Our mission is to transform the entrepreneurial experience and deliver operational excellence for businesses across the world through a unified platform supercharged with proprietary AI agents. We want to unleash the creativity of billions and inspire the world to dream big and build fast. We’re a rapidly growing team of forward-thinking and, most importantly, committed builders. We are driven by the opportunity to push boundaries, reimagine the foundations of human work, and shape tools that power the next generation of “business operations.” The way the world views and does business is changing, and we are committed to leading this change responsibly.
Role Overview
Python is central to Moon’s roadmap. Our data and ML layer powers core home services workflows— surfacing operational insights for service company owners and enabling predictive features that help users make better decisions. The data is real operational data at meaningful scale; the problems are genuinely interesting, and mistakes have real downstream consequences.
This is not a data science internship where you run notebooks in isolation. You’ll ship code that connects to a real backend and reaches real users. The year-round track is intentional: meaningful data and ML work takes time to build, validate, and integrate into a production product. You’ll go deeper here than you could in 12 weeks.
About The Role
You’ll join the data and ML engineering track with a dedicated mentor who works across data
engineering and applied ML — weekly 1:1s, pipeline reviews, and structured ramp milestones.
The code quality bar is the same as the rest of the engineering team. Mentorship is how we help
you get there — not a reason to lower it.
You’ll collaborate directly with the .NET team on data contracts between systems — the work
does not exist in isolation.
We expect you to be 3 days on-site in Glendale, with flexibility around your academic schedule.
Fully remote is not offered.
AI-assisted development is the default here — across EDA, pipeline development, debugging, and documentation. You’re expected to come in already working this way.
What You'll Do
Data Engineering & Pipelines
Build and maintain Python ETL pipelines: ingestion, transformation, validation, and reporting.
Write data validation and quality checks — bad data in production is a customer-facing problem,
not a technical inconvenience.
Instrument and monitor data pipelines; silent failures are often worse than loud ones.
Collaborate with the .NET team on data contracts between systems.
Write tests for pipeline outputs and model behavior; data pipelines have bugs just like
application code does — they’re just harder to find.
Applied ML & AI Integration
Prototype and develop ML features in production or active development — applied to home
services operational data.
Integrate LLM capabilities into application features using LangChain, direct API calls, or agent
orchestration patterns.
Use AI tools actively across the whole workflow: EDA, code generation, debugging,
documentation, and multi-step automated pipelines. AI-assisted development is your default
mode, not an occasional tool.
Document data models and transformation logic as part of the definition of done.
Qualifications
Required
Solid Python — functions, classes, error handling, and code that someone else can read.
Data manipulation with pandas, polars, or equivalent — load a dataset, clean it, answer
questions from it without fighting the tools.
SQL — non-trivial queries and a real understanding of what a join is doing.
AI tool usage that is habitual and specific: you’ve used LLMs to accelerate EDA, write boilerplate,
or debug data issues, and you can describe exactly how. This is evaluated explicitly.
Genuine intellectual curiosity about data — you want to know why a number looks wrong, not
just make the error go away.
Nice to Have
ML library exposure: scikit-learn, PyTorch, or similar. You don’t need production model
experience, but you should know what a train/test split is and why it matters.
Data pipeline tooling: Airflow, Prefect, dbt, or similar.
LangChain, OpenAI/Anthropic API integration, or agent workflow experience.
Cloud data services on Azure, AWS, or GCP.
FastAPI or Python-based API experience.
Statistics coursework — not required, but genuinely use
INTERN
intern
4/30/2026
You will be redirected to the job posting on LinkedIn.
Sign in and we'll score your resume against this role.