←back to thread

311 points melodyogonna | 2 comments | | HN request time: 0.64s | source
Show context
Cynddl ◴[] No.45137883[source]
Anyone knows what Mojo is doing that Julia cannot do? I appreciate that Julia is currently limited by its ecosystem (although it does interface nicely with Python), but I don't see how Mojo is any better then.
replies(8): >>45137977 #>>45137999 #>>45138029 #>>45138039 #>>45138393 #>>45138420 #>>45138513 #>>45143187 #
thetwentyone ◴[] No.45137999[source]
Especially because Julia has pretty user friendly and robust GPU capabilities such as JuliaGPU and Reactant[2] among other generic-Julia-code to GPU options.

1: https://enzymead.github.io/Reactant.jl/dev/ 2: https://enzymead.github.io/Reactant.jl/dev/

replies(1): >>45138080 #
jb1991 ◴[] No.45138080[source]
I get the impression that most of the comments in this thread don't understand what a GPU kernel is. These high-level languages like Python and Julia are not running on the kernel, they are calling into other kernels usually written in C++. The goal is different with Mojo, it says at the top of the article:

> write state of the art kernels

You don't write kernels in Julia.

replies(5): >>45138165 #>>45138182 #>>45138372 #>>45138558 #>>45138820 #
1. jakobnissen ◴[] No.45138165[source]
Im pretty sure Julia does JIT compilation of pure Julia to the GPU: https://github.com/JuliaGPU/GPUCompiler.jl
replies(1): >>45138233 #
2. actionfromafar ◴[] No.45138233[source]
” you should use one of the packages that builds on GPUCompiler.jl, such as CUDA.jl, AMDGPU.jl, Metal.jl, oneAPI.jl, or OpenCL.jl”

Not sure how that organization compares to Mojo.