#gpuprogramming kết quả tìm kiếm

Each common operation is implemented as its own .cu file—modular. intriguing. #CUDA #NVIDIA #GPUProgramming #libcudf

yasunrik's tweet image. Each common operation is implemented as its own .cu file—modular. intriguing.
#CUDA #NVIDIA #GPUProgramming #libcudf

Day 2 of #GPUProgramming: >read an article about shared memory >learnt about registers……..Global memory >almost blacked out from elaboration of L1 & L2 iykyk >”repetition”to digest what I just learnt for about 900k milliseconds

tatavishnurao's tweet image. Day 2 of #GPUProgramming:
>read an article about shared memory 
>learnt about registers……..Global memory
>almost blacked out from elaboration of L1 & L2 iykyk
>”repetition”to digest what I just learnt for about 900k milliseconds

"Need better CUDA textbooks. 'Programming Massively Parallel Processors' is a good intro. I've created C/CUDA C implementations for first 3 chapters. Check book & my GitHub repo for details. #CUDA #GPUprogramming"

taras_y_sereda's tweet image. "Need better CUDA textbooks. 'Programming Massively Parallel Processors' is a good intro. I've created C/CUDA C implementations for first 3 chapters. Check book & my GitHub repo for details. #CUDA #GPUprogramming"

Day 3 of GPU programming At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. #CUDA #GPUProgramming #MachineLearning

LearnInShadows's tweet image. Day 3 of GPU programming

At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. 
#CUDA #GPUProgramming #MachineLearning
LearnInShadows's tweet image. Day 3 of GPU programming

At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. 
#CUDA #GPUProgramming #MachineLearning

Day 2 of GPU programming Never knew addition needs so much code 😂 Starting to get the hang of program_id. Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code. Lets hope this momentum continues

LearnInShadows's tweet image. Day 2 of GPU programming

Never knew addition needs so much code 😂

Starting to get the hang of program_id.  
Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code.
Lets hope this momentum continues
LearnInShadows's tweet image. Day 2 of GPU programming

Never knew addition needs so much code 😂

Starting to get the hang of program_id.  
Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code.
Lets hope this momentum continues


"10 days into CUDA, and I’ve earned my first badge of honor! 🚀 From simple kernels to profiling, every day is a step closer to mastering GPU computing. Onward to 100! #CUDA #GPUProgramming #100DaysOfCUDA"

limbizzz11's tweet image. "10 days into CUDA, and I’ve earned my first badge of honor! 🚀 From simple kernels to profiling, every day is a step closer to mastering GPU computing. Onward to 100! #CUDA #GPUProgramming #100DaysOfCUDA"

#GPUProgramming - Day 07: 🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not! #COA #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic

#GPUProgramming - Day 02: 🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency. #COA #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 02:
🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency.
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 02:
🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency.
#COA #LearnInPublic

#GPUProgramming - Day 03: 🧠 CPUs: Processors adapt with DISA. #CPU's core duo - Control Unit & Datapath. Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 🔄 Follow the Instruction Execution Cycle: Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 03:
🧠 CPUs: Processors adapt with DISA. 
#CPU's core duo - Control Unit & Datapath. 
Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 
🔄 Follow the Instruction Execution Cycle: 
Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ 
#LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 03:
🧠 CPUs: Processors adapt with DISA. 
#CPU's core duo - Control Unit & Datapath. 
Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 
🔄 Follow the Instruction Execution Cycle: 
Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ 
#LearnInPublic

#GPUProgramming - Day 01: 🚀 Exploring RISC architecture: Simplified, optimized instructions in one clock cycle. 🔄 Bye, CISC complexity! 🏎️ Registers rule, boosting speed. 🤖💡 Compiler-friendly design, slick pipelining for simultaneous processing! 🕵️‍♂️ #COA #RISC #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 01:
🚀 Exploring RISC architecture: Simplified, optimized instructions in one clock cycle. 🔄 Bye, CISC complexity! 🏎️ Registers rule, boosting speed. 🤖💡 Compiler-friendly design, slick pipelining for simultaneous processing! 🕵️‍♂️
#COA #RISC #LearnInPublic

For maximum performance, firms often develop custom CUDA kernels. This involves writing low-level code to directly program the GPU's parallel cores, squeezing out every drop of efficiency for critical tasks. #CUDA #GPUProgramming


Day 3 of #GPUProgramming: >in-depth of what shared memory is capable of >read about techniques by which these concepts optimize their performance >synchronization of threads during matmul >realized it took ~1.5 hours to digest this stuff >tried to code matrix multiplication…

tatavishnurao's tweet image. Day 3 of #GPUProgramming:
>in-depth of what shared memory is capable of
>read about techniques by which these concepts optimize their performance
>synchronization of threads during matmul
>realized it took ~1.5 hours to digest this stuff
>tried to code matrix multiplication…

#GPUProgramming - Day 08: 🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 08:
🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 08:
🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic

⚡ Built my own graphics engine: Asthrarisine Sounds fun? Reality = invisible meshes, memory bugs & shader headaches. But here’s what made it work: #OpenGL #GraphicsEngine #GPUProgramming #GLTF #GameDev #ShaderProgramming


#GPUProgramming - Day 06: 🔍 Diving into computer architecture! 🖥️ Structural hazards arise when hardware resources are in high demand, causing contention among instructions. Data hazards? RAW, WAR, WAW – the battle for data paths and registers! 💡 #COA #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 06:
🔍 Diving into computer architecture! 🖥️ Structural hazards arise when hardware resources are in high demand, causing contention among instructions. Data hazards? RAW, WAR, WAW – the battle for data paths and registers! 💡
#COA #LearnInPublic

#GPUProgramming - Day 04: 🕰️ Dive into processor architectures! 🧠 Single-cycle execution, one clock cycle per instruction, demands a versatile datapath. 🔄 Multi-cycle instructions break it down for a more intricate dance with time. ⏳ #ComputerArchitecture #LearnInPublic 🚀

hridoy_bashir's tweet image. #GPUProgramming - Day 04:
🕰️ Dive into processor architectures! 🧠 Single-cycle execution, one clock cycle per instruction, demands a versatile datapath. 🔄 Multi-cycle instructions break it down for a more intricate dance with time. ⏳ #ComputerArchitecture #LearnInPublic 🚀
hridoy_bashir's tweet image. #GPUProgramming - Day 04:
🕰️ Dive into processor architectures! 🧠 Single-cycle execution, one clock cycle per instruction, demands a versatile datapath. 🔄 Multi-cycle instructions break it down for a more intricate dance with time. ⏳ #ComputerArchitecture #LearnInPublic 🚀

#GPUProgramming - Day 05: 🚀 Pipelining in computer architecture boosts performance by dividing instruction execution into stages. Techniques like forwarding, branch prediction, and superscalar processors enhance parallelism.💻🌐 #ComputerArchitecture #Pipelining #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 05:
🚀 Pipelining in computer architecture boosts performance by dividing instruction execution into stages. Techniques like forwarding, branch prediction, and superscalar processors enhance parallelism.💻🌐
#ComputerArchitecture #Pipelining #LearnInPublic

Programming Tensor Cores in Unity with WMMA (Warp Matrix Multiply Accumulate) API. #gpuprogramming #unity3d I have written minimal example: github.com/przemyslawzawo…

pzaworski90's tweet image. Programming Tensor Cores in Unity with WMMA (Warp Matrix Multiply Accumulate) API. #gpuprogramming #unity3d I have written minimal example: github.com/przemyslawzawo…

🚀Struggling to set up Nvidia's OpenCL on Linux? This step-by-step guide covers it all: ✅What is ICD Loader? ✅Installing Nvidia’s OpenCL & headers ✅Installing clinfo for validation 🎥Watch now: youtu.be/BIQOz5dfyoY #OpenCL #Linux #GPUProgramming #Nvidia #TheWolfAround

The_Wolf_Around's tweet image. 🚀Struggling to set up Nvidia's OpenCL on Linux? This step-by-step guide covers it all:
✅What is ICD Loader?
✅Installing Nvidia’s OpenCL & headers
✅Installing clinfo for validation
🎥Watch now: youtu.be/BIQOz5dfyoY
#OpenCL #Linux #GPUProgramming #Nvidia #TheWolfAround

Không có kết quả nào cho "#gpuprogramming"

Day 3 of GPU Programming -implemented ReLU activation for 1D array -Did potd leetcod -explored how GPUs handle element-wise operations in parallel #100DaysOfGPU #CUDA #GPUProgramming #ParallelComputing #AI #DeepLearning #100DaysOfCode #MachineLearning #NVIDIA #CodingJourney

cuda_programmer's tweet image. Day 3 of GPU Programming

-implemented ReLU activation for 1D array
-Did potd leetcod
-explored how GPUs handle element-wise operations in parallel 

#100DaysOfGPU #CUDA #GPUProgramming #ParallelComputing #AI #DeepLearning #100DaysOfCode #MachineLearning #NVIDIA #CodingJourney
cuda_programmer's tweet image. Day 3 of GPU Programming

-implemented ReLU activation for 1D array
-Did potd leetcod
-explored how GPUs handle element-wise operations in parallel 

#100DaysOfGPU #CUDA #GPUProgramming #ParallelComputing #AI #DeepLearning #100DaysOfCode #MachineLearning #NVIDIA #CodingJourney

Day 3 of GPU programming At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. #CUDA #GPUProgramming #MachineLearning

LearnInShadows's tweet image. Day 3 of GPU programming

At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. 
#CUDA #GPUProgramming #MachineLearning
LearnInShadows's tweet image. Day 3 of GPU programming

At this rate I'll be writing custom inference kernels for AI by next month. The gap between PyTorch abstractions and bare metal isn't as wide as it seemed. 
#CUDA #GPUProgramming #MachineLearning

Day 2 of GPU programming Never knew addition needs so much code 😂 Starting to get the hang of program_id. Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code. Lets hope this momentum continues

LearnInShadows's tweet image. Day 2 of GPU programming

Never knew addition needs so much code 😂

Starting to get the hang of program_id.  
Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code.
Lets hope this momentum continues
LearnInShadows's tweet image. Day 2 of GPU programming

Never knew addition needs so much code 😂

Starting to get the hang of program_id.  
Used Gemini 3.0 to generate pseudocode since I'm new to GPU programming and didn't want full code.
Lets hope this momentum continues


"Need better CUDA textbooks. 'Programming Massively Parallel Processors' is a good intro. I've created C/CUDA C implementations for first 3 chapters. Check book & my GitHub repo for details. #CUDA #GPUprogramming"

taras_y_sereda's tweet image. "Need better CUDA textbooks. 'Programming Massively Parallel Processors' is a good intro. I've created C/CUDA C implementations for first 3 chapters. Check book & my GitHub repo for details. #CUDA #GPUprogramming"

Each common operation is implemented as its own .cu file—modular. intriguing. #CUDA #NVIDIA #GPUProgramming #libcudf

yasunrik's tweet image. Each common operation is implemented as its own .cu file—modular. intriguing.
#CUDA #NVIDIA #GPUProgramming #libcudf

Day 2 of #GPUProgramming: >read an article about shared memory >learnt about registers……..Global memory >almost blacked out from elaboration of L1 & L2 iykyk >”repetition”to digest what I just learnt for about 900k milliseconds

tatavishnurao's tweet image. Day 2 of #GPUProgramming:
>read an article about shared memory 
>learnt about registers……..Global memory
>almost blacked out from elaboration of L1 & L2 iykyk
>”repetition”to digest what I just learnt for about 900k milliseconds

Finally, it has arrived!! I have got my #Nvidia Jetson Xavier NX :D #cuda #gpuProgramming

mario21ic's tweet image. Finally, it has arrived!! I have got my #Nvidia Jetson Xavier NX :D #cuda #gpuProgramming

#GPUProgramming - Day 07: 🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not! #COA #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 07:
🔧 #CPU Hazards 101 🚧: Ever heard of #Register Renaming & Out-of-Order Execution? They tackle structural hazards, ensuring smooth sailing for instructions. Watch out for Data Hazards (#RAW, #WAR, #WAW) in #MIPS, but fear not!
#COA #LearnInPublic

#GPUProgramming - Day 02: 🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency. #COA #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 02:
🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency.
#COA #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 02:
🔄 Exploring CPU architectures! #RISC, like #ARM & #Power, opts for efficiency with many registers. #CISC, exemplified by #Intel 8086, prioritizes simplicity, offering diverse, complex instructions. RISC excels in energy efficiency.
#COA #LearnInPublic

#GPUProgramming - Day 03: 🧠 CPUs: Processors adapt with DISA. #CPU's core duo - Control Unit & Datapath. Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 🔄 Follow the Instruction Execution Cycle: Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 03:
🧠 CPUs: Processors adapt with DISA. 
#CPU's core duo - Control Unit & Datapath. 
Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 
🔄 Follow the Instruction Execution Cycle: 
Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ 
#LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 03:
🧠 CPUs: Processors adapt with DISA. 
#CPU's core duo - Control Unit & Datapath. 
Datapath: Registers, ALU, Buses, Multiplexers – a data symphony! 
🔄 Follow the Instruction Execution Cycle: 
Fetch ➡️Decode➡️Execute➡️Store➡️ Update PC. 🕹️ 
#LearnInPublic

#GPUProgramming - Day 01: 🚀 Exploring RISC architecture: Simplified, optimized instructions in one clock cycle. 🔄 Bye, CISC complexity! 🏎️ Registers rule, boosting speed. 🤖💡 Compiler-friendly design, slick pipelining for simultaneous processing! 🕵️‍♂️ #COA #RISC #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 01:
🚀 Exploring RISC architecture: Simplified, optimized instructions in one clock cycle. 🔄 Bye, CISC complexity! 🏎️ Registers rule, boosting speed. 🤖💡 Compiler-friendly design, slick pipelining for simultaneous processing! 🕵️‍♂️
#COA #RISC #LearnInPublic

#GPUProgramming - Day 08: 🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic

hridoy_bashir's tweet image. #GPUProgramming - Day 08:
🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic
hridoy_bashir's tweet image. #GPUProgramming - Day 08:
🚀 Explored #computerarchitecture today! 🖥️ Control Hazards tackle branch prediction, #Pentium FDIV bug a classic example. 💡 Memory #Hierarchy is key—#RAM, #cache levels (L1, L2, L3), and storage devices play crucial roles. 🔄🌐 #Memory #LearnInPublic

Loading...

Something went wrong.


Something went wrong.


United States Trends