Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

snnn/p202310301 #18169

Closed
wants to merge 15 commits into from
Closed

snnn/p202310301 #18169

wants to merge 15 commits into from

Conversation

snnn
Copy link
Member

@snnn snnn commented Oct 30, 2023

centwang and others added 15 commits August 28, 2023 12:34
…7456)

### Description
<!-- Describe your changes. -->

Use name of temporary provisioning profile.

### Motivation and Context
<!-- - Why is this change required? What problem does it solve?
- If it fixes an open issue, please link to the issue here. -->

The old provisioning profile no longer works. Switched to a temporary
one that we can use before a new one is available. The temporary one has
a different name.

Alternative to microsoft#17454.
Disable QNN QDQ test for release branch

### Description
Disable QNN QDQ test for release branch to get rid of model test failure
caused by new model update in build image.
…icrosoft#17461)

### Description
Remove 52 from CMAKE_CUDA_ARCHITECTURES to reduce Nuget package size. 

### Motivation and Context
PR microsoft#17227 increased binary size by 20%. Right the package size is about
260MB. However, nuget has a hard limit of 250MB. Without this change we
cannot publish the package.
Cherry-pick microsoft#17507  for rel-1.16.0.

Note: The PR 17507 contains the part of engine decryption refactor that
we don't want to include it in ORT 1.16 release. This cherry pick PR
excludes this part.
### Description
1. Delete Prefast tasks (microsoft#17522)
2. Disable yum update (microsoft#17551)
3. Avoid calling patchelf (microsoft#17365 and microsoft#17562) we that we can validate
the above fix

The main problem I'm trying to solve is: our GPU package depends on both
CUDA 11.x and CUDA 12.x . However, it's not easy to see the information
because ldd doesn't work with the shared libraries we generate(see issue
microsoft#9754) . So the patchelf change are useful for me to validate the
"Disabling yum update" was successful. As you can see we call "yum
update" from multiple places. Without some kind of validation it's hard
to say if I have covered all of them.
The Prefast change is needed because I'm going to update the VM images
in the next a few weeks. In case of we need to publish a patch release
after that.

### Motivation and Context
Without this fix we will mix using CUDA 11.x and CUDA 12.x. And it will
crash every time when we use TensorRT.
Cherry-pick the following PRs to the release branch:

Fix: Fail to skip disabledmodel in winml (microsoft#17728) 
Move dotnet build and test into docker in Linux CPU CI (microsoft#17417) 
Run Nuget_Test_Linux_GPU in container (microsoft#17452) 
Run Final_Jar_Testing_Linux_GPU in docker (microsoft#17533) 
TreeEnsemble speed up (microsoft#17449) 
Remove onnxruntime extensions from list of gitmodules (microsoft#17615) 
Include onnxruntime_float16.h in the package. (microsoft#17637) 
Fix static quantization for QDQ and Percentile distribution (microsoft#17649) 
[TensorRT EP] Back out the PerThreadContext (microsoft#17690) 
Update nodejs to 18.x (microsoft#17657) 
Update linux-wasm-ci.yml: remove the ln command (microsoft#17735)
Remove the condition to allow an empty provide list.

Co-authored-by: Randy Shuai <rashuai@microsoft.com>
### Description
fix session option access in Node.js binding


### Motivation and Context
This is a bug that affect transformer.js using ONNX Runtime Node.js
binding. Issue: microsoft#17377

This bug is already fixed in main branch, but it is not picked in 1.16
release.
### Description

Python package pipeline fails due to "tokenizers" compilation. Since
"tokenizers" is a dep of "transformers", we update its version and hope
a new solution had been there.

```
error: casting `&T` to `&mut T` is undefined behavior, even if the reference is unused, consider instead using an `UnsafeCell`
--> tokenizers-lib/src/models/bpe/trainer.rs:517:47
```



### Motivation and Context
Cherry-pick from microsoft#17823
…8070)

1. Increase version number for preparing the 1.16.2 release (microsoft#18070)
2. cherry-pick 18034
…soft#18034)" (microsoft#18150)

This reverts commit 99b8dca.

### Description
<!-- Describe your changes. -->



### Motivation and Context
Restore the dml stage in windows GPU  pipeline.
Agent issue is solved by adding Feature.DisableGpuDriver in pool
properties.
@snnn snnn requested a review from a team as a code owner October 30, 2023 16:40
@snnn snnn closed this Oct 30, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

9 participants