Hacker Newsnew | past | comments | ask | show | jobs | submit | textlapse's commentslogin

If someone could please give an octopus a waterproof keyboard, perhaps we could have a kernel, a compiler and a new internet protocol all in one.


To really confuse it, ask it to take that tricycle with the platypus on it to a car wash.


What's the performance like? What would the benefits be of converting a streaming multiprocessor programming model to this?


We aren't focused on performance yet (it is often workload and executor dependent, and as the post says we currently do some inefficient polling) but Rust futures compile down to state machines so they are a zero-cost abstraction.

The anticipated benefits are similar to the benefits of async/await on CPU: better ergonomics for the developer writing concurrent code, better utilization of shared/limited resources, fewer concurrency bugs.


warp is expensive - essentially it's running a 'don't run code' to maintain SIMT.

GPUs are still not practically-Turing-complete in the sense that there are strict restrictions on loops/goto/IO/waiting (there are a bunch of band-aids to make it pretend it's not a functional programming model).

So I am not sure retrofitting a Ferrari to cosplay an Amazon delivery van is useful other than for tech showcase?

Good tech showcase though :)


I think you're conflating GPU 'threads' and 'warps'. GPU 'threads' are SIMD lanes that are all running with the exact same instructions and control flow (only with different filtering/predication), whereas GPU warps are hardware-level threads that run on a single compute unit. There's no issue with adding extra "don't run code" when using warps, unlike GPU threads.


My understanding of warp (https://docs.nvidia.com/cuda/cuda-programming-guide/01-intro...) is that you are essentially paying the cost of taking both the branches.

I understand with newer GPUs, you have clever partitioning / pipelining in such a way block A takes branch A vs block B that takes branch B with sync/barrier essentially relying on some smart 'oracle' to schedule these in a way that still fits in the SIMT model.

It still doesn't feel Turing complete to me. Is there an nvidia doc you can refer me to?


That applies inside a single warp, notice the wording:

> In SIMT, all threads in the warp are executing the same kernel code, but each thread may follow different branches through the code. That is, though all threads of the program execute the same code, threads do not need to follow the same execution path.

This doesn't say anything about dependencies of multiple warps.


It's definitely possible, I am not arguing against that.

I am just saying it's not as flexible/cost-free as you would on a 'normal' von Neumann-style CPU.

I would love to see Rust-based code that obviates the need to write CUDA kernels (including compiling to different architectures). It feels icky to use/introduce things like async/await in the context of a GPU programming model which is very different from a traditional Rust programming model.

You still have to worry about different architectures and the streaming nature at the end of the day.

I am very interested in this topic, so I am curious to learn how the latest GPUs help manage this divergence problem.


I would love to see a nutritional facts label on how many prompts / % of code / ratio of human involvement needed to use the models to develop their latest models for the various parts of their systems.


There is a big difference between paperbacks and TikTokification:

Paperbacks required authors to spend the same amount of time/effort to create content with a vastly expanded market and distribution mechanism.

TikTok and Insta created N creators to M consumers where N is nearly the same as M. Making the distribution channels bigger but effortless to create content doesn’t magically equate quality paperbacks with short form hummingbird-attention videos.


Not sure about your N = M

TikTok isn’t much bigger than open mics?


This is… not even remotely true or even descriptive of the nature of TikTok.


His consistency and craftsmanship is amazing.

Being an engineer and coding at this stage/level is just remarkable- sadly this trade craft is missing in most (big?) companies as you get promoted away into oblivion.


On the contrary, Jeff Bezos and Steve Jobs were very customer focused - they listened to customers more so than other big tech companies. You could actually have emailed them and gotten a response (in the case of Bezos a legendary ‘?’ forward to the team).

Not sure how apocryphal a tale this is but it does speak volumes to how customer obsessed these companies were.


Jobs responded to email back when he ran a tiny company in an overall much smaller industry with far fewer customers. I'm not sure he was responding so much by the time he had resurrected Apple and the iPhone launched.


I once emailed Tim Cook and sure enough, he responded. Seemed like it was him, too.


This is awesome… but: On the Web, is Avalonia using Skia to render inside a Skia Canvas?

**insert inception meme here**

Joking aside: this points to MSFT moving away from the whole Mono/Maui investments and into Aspire or whatever they call it. Without MSFT backing this I am not sure if there is much more future left for MAUI (or dotnet on mobile in general).

Avalonia is great though.


> On the Web, is Avalonia using Skia to render inside a Skia Canvas?

Yes. They're also looking at offering Impeller as a render option:

https://avaloniaui.net/blog/avalonia-partners-with-google-s-...


I wish C++ did some sane things like if I have a const member variable, allow me to initialize it as I wish in my constructor - it's a constructor for crying out loud.

Don't be silly and assume if I assign it multiple times in an if condition it's mutable - it's constructing the object as we speak, so it's still const!!!

C# gets this right among many other things (readonly vs const, init properties, records to allow immutability by default).

And the funny thing is the X thread has lots of genuine comments like 'yeah, just wrap a lambda to ensure const correctness' like that's the okay option here? The language is bad to a point it forces good sane people into seeing weird "clever" patterns all the time in some sort of an highest ELO rating for the cleverest evilest C++ "solution".

I was hoping Carbon was the hail mary for saving C++ from itself. But alas, looks like it might be googlified and reorged to oblivion?

Having said that, I still like C++ as a "constrained C++" language (avoid clever stuff) as it's still pretty good and close to metal.


It does feel like Apple is firing on all cylinders for their core competencies.

Software (iOS26), services (Music/Tv/Cloud/Apple Intelligence) and marketing (just keep screaming Apple Intelligence for 3 months and then scream Liquid Glass) ---- on the other hand seem like they are losing steam or very reactive.

No wonder John Ternus is the widely anticipated to replace Tim Cook (and not Craig).


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: