Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove null pointer from error message in mlir::odml #67556

Open
apach301 opened this issue May 14, 2024 · 6 comments · May be fixed by #67557
Open

Remove null pointer from error message in mlir::odml #67556

apach301 opened this issue May 14, 2024 · 6 comments · May be fixed by #67557
Assignees
Labels
comp:lite TF Lite related issues stat:awaiting tensorflower Status - Awaiting response from tensorflower TF 2.16 type:bug Bug

Comments

@apach301
Copy link
Contributor

Issue type

Bug

Have you reproduced the bug with TensorFlow Nightly?

No

Source

source

TensorFlow version

2.16.1

Custom code

No

OS platform and distribution

No response

Mobile device

No response

Python version

No response

Bazel version

No response

GCC/compiler version

No response

CUDA/cuDNN version

No response

GPU model and memory

No response

Current behavior?

Possible null pointer dereference in mlir::odml::MatchInverseScalesOperand(): undefined variable inverse_scale_constant_op is printed in error message.

auto inverse_scale_constant_op =
dyn_cast_or_null<stablehlo::ConstantOp>(inverse_scales.getDefiningOp());
if (!inverse_scale_constant_op) {
llvm::dbgs()
<< "Inverse scales should be a constant. Instead, it was defined by: "
<< inverse_scale_constant_op << ".\n";
return failure();

Standalone code to reproduce the issue

Bug was found with Svace static analyzer

Relevant log output

No response

@AsbestosLampshade
Copy link

Hi @apach301 is this issue still open?
If so, could you tell me how to replicate the bug so that I may have a go at it?

@apach301
Copy link
Contributor Author

Hi @apach301 is this issue still open?

If so, could you tell me how to replicate the bug so that I may have a go at it?

This bug was detected by static analysis, therefore I have no input to reproduce it during execution. But I propose a fix in PR #67557

@AsbestosLampshade
Copy link

Hi @apach301 the PR looks great, if you would like any help with CI, do let me know.

Thanks and regards

@Venkat6871 Venkat6871 added TF 2.16 comp:lite TF Lite related issues labels May 16, 2024
@Venkat6871 Venkat6871 assigned sawantkumar and unassigned Venkat6871 May 17, 2024
@sawantkumar sawantkumar assigned pkgoogle and unassigned sawantkumar May 23, 2024
@pkgoogle
Copy link

will wait for review.

@pkgoogle pkgoogle added the stat:awaiting tensorflower Status - Awaiting response from tensorflower label May 29, 2024
@apach301
Copy link
Contributor Author

will wait for review.

Hi @pkgoogle How to enable CI tests run in PR #67557? Is it wait any approval to run?

@pkgoogle
Copy link

Hi @apach301, we'll have to wait for review, we run CI/CD internally and review the PR.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
comp:lite TF Lite related issues stat:awaiting tensorflower Status - Awaiting response from tensorflower TF 2.16 type:bug Bug
Projects
None yet
Development

Successfully merging a pull request may close this issue.

5 participants