JOB DETAILS

AI Research Engineer

CompanyEnCharge AI
LocationMunicipio Sena
Work ModeOn Site
PostedMarch 18, 2026
About The Company
EnCharge AI is a leader in advanced hardware and software systems for AI computing. EnCharge’s robust and scalable next-generation in-memory computing technology provides orders-of-magnitude higher compute efficiency and density compared to today’s best-in-class solutions, at a fraction of the cost. The high-performance architecture is coupled with seamless software and will enable the immense potential of AI to be accessible in power -, energy-, and space-constrained applications. EnCharge AI launched in 2022 and is led by veteran technologists with backgrounds in semiconductor design and AI systems.
About the Role

EnCharge AI is a leader in advanced AI hardware and software systems for edge-to-cloud computing. EnCharge’s robust and scalable next-generation in-memory computing technology provides orders-of-magnitude higher compute efficiency and density compared to today’s best-in-class solutions. The high-performance architecture is coupled with seamless software integration and will enable the immense potential of AI to be accessible in power, energy, and space constrained applications. EnCharge AI launched in 2022 and is led by veteran technologists with backgrounds in semiconductor design and AI systems.

About the Role

EnCharge AI is looking for an experienced AI Research Engineer to optimize deep learning models for deployment on edge AI platforms. You will work on model compression, quantization strategies, and efficient inference techniques to improve the performance of AI workloads. 

Responsibilities

  • Research and develop quantization-aware training (QAT) and post-training quantization (PTQ) techniques for deep learning models.

  • Implement low-bit precision optimizations (e.g., INT8, BF16).

  • Design and optimize efficient inference algorithms for AI workloads, focusing on latency, memory footprint, and power efficiency.

  • Work with frameworks such as PyTorch, ONNX Runtime, and TVM to deploy optimized models.

  • Analyze accuracy trade-offs and develop calibration techniques to mitigate precision loss in quantized models.

  • Collaborate with hardware engineers to optimize model execution for edge devices, and NPUs.

  • Contribute to research on knowledge distillation, sparsity, pruning, and model compression techniques.

  • Benchmark performance across different hardware and software stacks.

  • Stay updated with the latest advancements in AI efficiency, model compression, and hardware acceleration. 

Qualifications

  • Master’s or Ph.D. in Computer Science, Electrical Engineering, or a related field.

  • Strong expertise in deep learning, model optimization, and numerical precision analysis.

  • Hands-on experience with model quantization techniques (QAT, PTQ, mixed precision).

  • Proficiency in Python, C++, CUDA, or OpenCL for performance optimization.

  • Experience with AI frameworks: PyTorch, TensorFlow, ONNX Runtime, TVM, TensorRT, or OpenVINO.

  • Understanding of low-level hardware acceleration (e.g., SIMD, AVX, Tensor Cores, VNNI).

  • Familiarity with compiler optimizations for ML workloads (e.g., XLA, MLIR, LLVM). 

EnchargeAI is an equal employment opportunity employer in the United States.

 

Key Skills
Deep LearningModel OptimizationQuantizationPythonC++CUDAOpenCLPyTorchTensorFlowONNX RuntimeTVMTensorRTOpenVINOHardware AccelerationCompiler OptimizationsModel Compression
Categories
TechnologyEngineeringScience & ResearchData & AnalyticsSoftware
Job Information
📋Core Responsibilities
The AI Research Engineer will research and develop quantization techniques for deep learning models and implement optimizations for efficient inference algorithms. Collaboration with hardware engineers is essential to optimize model execution for edge devices.
📋Job Type
full time
📊Experience Level
5-10
💼Company Size
88
📊Visa Sponsorship
No
💼Language
English
🏢Working Hours
40 hours
Apply Now →

You'll be redirected to
the company's application page