Skip to content

Add 16bit xtensa depthwise conv kernel support#3481

Open
narrietal wants to merge 1 commit intotensorflow:mainfrom
narrietal:Add_16x8_deptwise_conv_xtensa_opt_kernel
Open

Add 16bit xtensa depthwise conv kernel support#3481
narrietal wants to merge 1 commit intotensorflow:mainfrom
narrietal:Add_16x8_deptwise_conv_xtensa_opt_kernel

Conversation

@narrietal
Copy link

@narrietal narrietal commented Feb 16, 2026

This PR adds support for the optimized Xtensa depthwise convolution kernel when using 16-bit activations and 8-bit weights. Previously, this configuration would fall back to the reference implementation.

Changes:

  • Removed hardcoded if-else logic in the Prepare function that restricted inputs to int8 activations only
  • Removed TF_LITE_ENSURE_EQ assertion enforcing int8-only inputs
  • Renamed the existing int8 evaluation function for clarity
  • Added a new evaluation function to support int16 activations with int8 weights

bug=fixes #3484

@narrietal narrietal requested a review from a team as a code owner February 16, 2026 15:39
@rameshkunasi rameshkunasi added the ci:full Triggers the comprehensive cross-platform test suite. label Mar 4, 2026
@rameshkunasi
Copy link
Contributor

Hi @narrietal,

Thank you for this PR. Can you please resolve the failed test cases for Hifi3z and Fusion F1 platforms and update the PR?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci:full Triggers the comprehensive cross-platform test suite.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Missing depthwise conv 16 bit xtensa kernel

2 participants