
Product Application Engineer - AI Models, AI Frameworks, Python, C++, Embedded Systems
- San Jose, CA
- Permanent
- Full-time
- Facilitate adoption of AMD’s ML inferencing solutions by facilitating customer on-boarding and delivering best-in-class customer experience
- Work with R&D, marketing, and key customers to drive the evolution of the product
- Support strategic business engagements leveraging AMD’s ML inferencing solutions
- Evaluate end-to-end ML inferencing pipelines optimized for AMD’s Embedded devices. This role involves conducting performance benchmarking, memory profiling, and bottleneck analysis in addition to extensive flow, and usability, analysis for model inferencing flows.
- Develop use cases, examples, tutorials, and methodology collaterals for AI inference software components and ecosystems, including Compiler, Quantizer, Optimizer, Runtime, Profiler, Visualizer, Debugger, and more targeting embedded devices.
- Engage with engineering teams to collaborate on product development, analyze product/feature specifications and usability to provide early feedback, and understand product/feature usage holistically to identify potential customer pain points.
- Interact with internal and external customers to understand their issues, assist them in debugging their workflows to meet critical model and flow-specific requirements, and create use cases to reproduce issues, driving successful adoption.
- Collaborate with the sales and marketing teams on strategic business engagements.
- Work closely with R&D to prioritize issues and manage escalations.
- Experience working with AI frameworks such as PyTorch and ONNX.
- Proven experience deploying and optimizing ML models on embedded hardware(microcontrollers, custom SOCs or similar platforms)
- Work experience with state-of-the-art models in one or more domains: CNN, Transformers, LLMs, Generative AIs.
- Experience in model analysis (CNN, LLM), performance, benchmarking, and accuracy testing.
- Experience in quantizing and pruning ML networks is preferred.
- Experience working on and managing open-source AI projects is preferred.
- Hands-on experience with Python, C, and C++.
- Expertise with popular development tools such as Git, CMake, Make, Conda, Docker, VSCode, bash, Linux.
- Demonstrated expertise in performance analysis and debugging across software/hardware boundaries
- Ability to think from the end-user's perspective, understand their needs, and advocate for them.
- Documentation skills with the ability to present features, solutions clearly and concisely.
- Excellent analytical, problem-solving, and communication skills
- Bachelor’s or Master’s in computer engineering or computer science or electrical engineering, or comparable disciplines
- San Jose, CA OR
- Longmont, CO