Alex Ziskind is a tech enthusiast and YouTuber who specializes in AI hardware performance, developer hardware, and mini PCs. He delves into topics like local LLMs, Apple Silicon, and benchmarking GPU efficiency, offering insights that cater to both tech enthusiasts and professionals. His style is engaging and informative, often presenting complex topics in an accessible manner.
How this creator's focus and perspective has shifted over time
General hardware reviews and benchmarks
Apple Silicon and local LLMs
Advanced AI performance and benchmarking
Over time, Alex's content has evolved from a focus on general hardware reviews to more specialized topics like local LLMs and GPU efficiency. His recent videos highlight his deep dive into Apple Silicon, benchmarking, and innovative solutions for AI performance.
Every video we've tracked from this creator, newest first
This video discusses the successful setup of a four-node NVIDIA cluster, highlighting both operational success and potential vendor restrictions.
"I've got a cluster of four of these working together."
Cloud code have very popular tool can now use local models. In other words, I can be running an LLM right on my laptop, and Cloud code ...
This video compares high-end graphics cards for running Large Language Models, arguing that the new RTX Pro 6000 with 96GB VRAM is a superior choice over the M3 Ultra and RTX 5090.
"This is the brand new RTX Pro 6000, and it's heavier than it looks."
This video reviews GMK Tech's new EVO X2 Mini PC, highlighting its groundbreaking 128GB RAM capacity designed for running local large language models.
"DJX Park will be ready, will be available shortly, probably in a few weeks."
This video tutorial demonstrates how to significantly improve local LLM inference speed using a draft model approach, specifically targeting performance bottlenecks.
"Alright, you're gonna like this. Watch this."
The DJX Spark isn't the only player in town when it comes to having one petaflop of AI supercomputer on your desk. There's a bunch of t...
This video demonstrates the feasibility of running a massive 1 trillion parameter AI model locally on a Mac computer, highlighting recent advancements in model efficiency.
"KK 2.5 is out and it's the new big hot model."
You might have heard that open video is here and it's the first time an actual open waste release shows up with the whole stack. The mo...
This video documents the process of upgrading and resolving issues within a self-hosted private AI framework cluster to enhance performance.
"I made some upgrades to my cluster."
This video explores why local LLM inference often underperforms and provides actionable steps to optimize speed by adjusting backend configurations and hardware utilization.
"You're probably familiar with Olama, right? This is what it looks like, you can launch it, you can talk to it."
This video reviews the GeekOM A9 Max mini PC, evaluating its performance for developer workloads and local LLMs while highlighting its sub-$1000 price point.
"The AMD Strix Point chip is finally in a mini PC that's under $1000 in this GeekOM A9 Max that many of you have been asking about."
This video challenges the common perception that NVIDIA DGX Spark systems are slow when running AI models, arguing that standard benchmarks may not reflect real-world performance.
"What if I told you that what you've seen so far about the DJX Sparking's slower is wrong?"
I have a course site that I've had for many years and I need to migrate away from my current provider. Unfortunately, because their pri...
This video reviews developer laptop options, comparing historical safe bets with current hardware suitable for local AI inference versus traditional coding workflows.
"I feel like there's two types of developers right now."
This video benchmarks different machines to determine which can generate 1 million tokens the fastest, comparing performance and costs.
"I wanted to answer one simple question. Which of these machines can generate 1 million tokens the fastest?"
Alex Ziskind discusses a new Apple update that potentially solves historical issues with Mac cluster performance and longevity.
"Every Mac cluster I've built in the past has had the same painful ending considerably worse."
If your Apple Silicon machine like Macbook or Little Mac Mini doesn't have much RAM and you still want to run large language models that are decently sized, you can dow...
AI News & Strategy Daily | Nate B Jones
36 videos
Nate presents his content in a clear, concise, and engaging manner, often using real-world examples to illustrate complex concepts.
Wes Roth
27 videos
Roth presents his content with a mix of informative and provocative commentary, often challenging conventional views on AI.
TheAIGRID
21 videos
TheAIGRID presents content with a mix of detailed analysis, provocative headlines, and engaging storytelling.
Matthew Berman
20 videos
Matthew Berman presents his content in a dynamic and interactive manner, often using live demonstrations and personal anecdotes to explain complex AI concepts.
WorldofAI
13 videos
WorldofAI presents content in a clear, concise manner with a focus on practical applications and hands-on demonstrations.
Julia McCoy
11 videos
Julia presents with a mix of enthusiasm and clarity, breaking down complex topics into understandable segments.