For example, while searching for low-latency open source apps I found this OpenPDC application on Github which seems to be the de-facto standard application for "high-performance data collection" used by numerous utilities around the world. It's even used in academic articles as the tool for assessing network delays for the measurement devices. And yet, the application is written in .NET, and runs on Windows.
The number of applications that have hard real time requirements may be fewer than you think. Also keep in mind that high-performance may mean high throughput, not low latency. These are two different metrics, sometimes at odds.
If you prefer low-latency, disable Nagle's algorithm:
https://msdn.microsoft.com/en-us/library/system.net.sockets.socket.nodelay(v=vs.110).aspx .
I am aware of "techniques" used to reduce GC in general (less garbage generation, static allocation, and similar stuff), but I still had the general idea that that Windows is "not a real-time OS", and that nothing can prevent the GC from pausing your app - you can only delay the inevitable.
Concurrent collectors can collect some garbage in background threads without pausing the world. Many GCs also give you some ability to control when GC pauses can happen, including .NET's:
https://msdn.microsoft.com/en-us/library/system.gc.trystartnogcregion.aspx
Even before this, unless your generating a lot of garbage, GC pauses are often short enough to not be a problem - especially in the context of networking where 100ms+ delays just routing your packets over the internet through several routers are not uncommon. The GC potentially pausing your app for even a few milliseconds in this context is probably not a problem. General purpose operating systems can suspend your threads for longer, just to share your CPU with other processes.
Has something changed in recent .NET/Windows Server versions which would allow this kind of applications to run with near zero latency? Is it possible that this and other similar applications are written in a way which completely prevents "stop-the-world" garbage collection/blocking due to Windows non-real-time nature, or is it unrealistic to expect this to be guaranteed?
There's a big difference between "running with near zero latency most of the time", and "guaranteed to run with near zero latency all the time, under any and all circumstances."
I'd argue the former has been quite possible in Windows and .NET for quite some time - the most recent change I'm aware of being the adoption of concurrent GCs - whereas the latter might mean TCP itself hasn't ever been an option for you, for not guaranteeing enough.