About the Role
As an ML Research Engineer at SEMRON, you will design the algorithms and quantization schemes that unlock efficient, high-accuracy inference on our analog in-memory compute platform. Your work will bridge cutting-edge quantization research, mathematical modeling, and hardware-aware algorithm design, ensuring that deep neural networks execute with maximal accuracy and throughput on our custom silicon.
Why us?
We’re building at the intersection of math, hardware, and machine learning, pushing the boundaries of what's possible in compute. If you’ve implemented your own MVM kernels just to see what happens, trained quantized models for fun, or love thinking deeply about efficiency, sparsity, and how to make models run faster and better, you’ll feel right at home. As a small, technical team, early work defines the future of the stack, and we treat it that way. You'll own critical pieces of what we build, with equity to match. No hierarchy, no bureaucracy, just ideas, experiments, and real impact. You’ll grow as fast as you can grow.
About us
At SEMRON, we’re redefining what’s possible in AI hardware. Our core innovation lies in analog in-memory computing for deep neural network acceleration, enabling us to build compute architectures that scale vertically into the third dimension, much like NAND flash revolutionized memory. This leap in physical density allows us to deploy models with billions of parameters on chip areas as small as a few square millimeters.