at sync. we're making video as fluid and editable as a word document.
how much time would you save if you could record every video in a single take?
no more re-recording yourself because you didn't like what you said, or how you said it.
just shoot once, revise yourself to do exactly what you want, and post. that's all.
this is the future of video: AI modified >> AI generated
we're playing at the edge of science + fiction.
our team is young, hungry, uniquely experienced, and advised by some of the greatest research minds + startup operators in the world. we're driven to solve impossible problems, impossibly fast.
our founders are the original team behind the open sourced wav2lip — the most prolific lip-sync model to date w/ over 9k+ GitHub stars.
we’re at a stage today in computer vision where we were w/ NLP two years ago — have a bunch of disparate, specialized models (eg. Sentiment classification, translation, summarization, etc), but LLMs (a generalized large language model) displaced them.
we’re taking the same approach – curating high quality datasets + training a series of specialized models to accomplish specific tasks, while building up to towards a more generalized approach for one model to rule them all.
post batch our growth is e^x – we need help asap to scale up our infra, training, and product velocity.
we look for the following: [1] raw intelligence [2] boundless curiosity [3] exceptional resolve [4] high agency [5] outlier hustle
sync. is a team of artists, engineers, and scientists building foundation models to edit and modify people in video. Founded by the creators of Wav2lip and backed by legendary investors, including YC, Google, and visionaries Nat Friedman and Daniel Gross, we've raised 6 million dollars in our seed round to evolve how we create and consume media.
Within months of launch our flagship lipsync API scaled to millions in revenue and powers video translation, dubbing, and dialogue replacement workflows for thousands of editors, developers, and businesses around the world.
That's only the beginning, we're building a creative suite to give anyone Photoshop-like control over humans video – zero-shot understanding and fine-grained editing of expressions, gestures, movement, identity, and more.
Everyone has a story to tell, but not everyone's a storyteller – yet. We're looking for talented and driven individuals from all backgrounds to build inspired tools that amplify human creativity.
About the role
We're seeking an exceptional ML Engineer to expand the boundaries of what's possible with AI video editing. You'll work with the creators of Wav2lip to build and extend computer vision pipelines giving users unprecedented control over humans in video.
What you'll do
Create novel CV features that unlock new forms of video manipulation
Build ML pipelines that understand and modify humans in video
Transform research breakthroughs into production capabilities
Design systems that make complex AI feel like magic to users
Pioneer new approaches to fine-grained video control
What you'll need
5+ years implementing computer vision and ML systems that users love
Deep expertise in PyTorch and video processing pipelines
Track record of shipping novel ML features from concept to production
Ability to bridge cutting-edge research with practical applications
Strong collaboration skills across research and engineering teams
Preferred qualifications
Experience with face/human detection and tracking
Background in generative AI or video understanding
History working with large-scale video datasets
Open source contributions to CV/ML projects
Our goal is to keep the team lean, hungry, and shipping fast.
These are the qualities we embody and look for:
[1] Raw intelligence: we tackle complex problems and push the boundaries of what's possible.
[2] Boundless curiosity: we're always learning, exploring new technologies, and questioning assumptions.
[3] Exceptional resolve: we persevere through challenges and never lose sight of our goals.
[4] High agency: we take ownership of our work and drive initiatives forward autonomously.
[5] Outlier hustle: we work smart and hard, going above and beyond to achieve extraordinary results.
[6] Obsessively data-driven: we base our decisions on solid data and measurable outcomes.
[7] Radical candor: we communicate openly and honestly, providing direct feedback to help each other grow.
next.js nest.js python pytorch aws/gcp/azure kubernetes
fulltimeSan FranciscoFull stack$165K - $250K0.10% - 1.30%6+ years
fulltimeBengaluruFull stack₹1M - ₹10M INR0.10% - 0.50%3+ years
fulltimeSan FranciscoFull stack$130K - $200K0.30% - 1.00%6+ years
fulltimeBengaluru₹1M - ₹10M INR0.15% - 0.75%3+ years
fulltimeSan FranciscoFull stack$165K - $240K0.20% - 1.20%6+ years