* Centralize Microsoft.Net.Sdk package version
Note: Templates were omitted as their version needs to be static.
* Unifying additional missmatched versions
* prefercliruntime
whitespace threw off ReplaceAll
* Additional missed globs
* Revert SDK version for performance tests
* PR Feedback
* Roll back VSTestXunitDesktopAndNetCore.csproj SDK version
* Updating the Microsoft.Net.Sdk & Microsoft.Net.Sdk.Web versions
* Fixed merge conflicts. Had to re-update the Sdk version in one place.
* re-migrate dotnet.dll
* Revert Performance Test Projects
* Fix test test
* Add missing WithRuntime
* Disable failing test test
* Update README.md
* Moving Ubuntu 16.04 to be next to Ubuntu 14.04
* Adding Oracle Linux and Linux Mint to the titles
Adding Oracle Linux and Linux Mint to the titles next to their compatible binaries.
* Remove showing firsttime eula for non verbs.
* Add Serviceable assembly attribute and nuspec attributes for all shipping CLI assemblies.
Fix#3345
* Use NugetCache Sentinel for Telemetry setting.
* Fix Oracle Linux version in README.md
Oracle Linux 7 -> Oracle Linux 7.1
* Fix README to use hostfxr download links (#3622)
Also fix a rebase error from b524fd079e6dcdd744faeb6061ccbfe99d1f810f#diff-04c6e90faac2675aa89e2176d2eec7d8
* Remove the VS2012 CRT dependency from docs (#3632)
* fix typo in dotnet-install file
This was needed to rebuild the CLI with the updated Roslyn NuGet packages.
Thanks goes to @akoken for the fix.
Previously I attempted to save the cost of reading/decompressing files
by keeping a copy in memory. That resulted in too much memory pressure
so I later swapped in temporary files. This defeated the initial goal
of keeping the streams alive to some extent since the files would
be flushed to disk.
In practice since all the nupkgs we are packing also have raw files so
we weren't even saving much CPU by avoiding a second decompression.
Switch the archiver to just first index all files and save the source
information to read at a later point. After this when building the
archive we'll reopen the files/zips and copy from there.
My measurements show that this actually improve the archiving perf and
the lack of temp files means we don't hit the ulimit restriction on OSX.
This adds LZMA to the THIRD-PARTY-NOTICES list for CLI. Since we've
modified the SDK in porting it to .NET Core and minor bugfixes we've
also been advised to add the .NET Foundation copyright header.
Previously we'd keep track of any file that we extracted once and try
to reuse that file (by copying it) if we needed the same file later at a
different destination. The reason was that it's theoretically faster to
a file copy than a createfile and write, since the copy can happen
entirely in the kernel. In practice we were foiled by AV scanners.
This happens to be the only time during extraction where we let a file
close after writing it and then try and use it again. Sure enough on
fast machines we were seeing that as soon as we closed it MsMpEng would
map the file for exclusive access causing our copy to fail with a
sharing violation.
To fix this, I've removed the copy optimization and will just copy the
file from the in-memory archive every time.
Renaming destinationPath to destinationRelativePath for clarity. This
string represents the path relative to the extract directory to which
files will be written.
We were missing an Interlocked.Increment during a parallel operation.
This type will archive multiple files de-duplicated in a zip with a
central manifest that describes how to recreate the actual layout.
In addition, any other zip file will be expanded and deduplicated
so that we can further reduce the size of those zips/nupkgs.
All of these are placed in a zip with no-compression, essentially
using the zip as only a container. We then LZMA compress that
container to achieve maximum compression.
We're builiding a library and don't need the application-related source
so I've removed this in addition to fixing the source to compile as
NETStandard1.0.