Тёмный

Get started Gemma 2 Locally on Mac using MLX 

Prince Canuma
Подписаться 968
Просмотров 884
50% 1

In this video, we'll explore how to convert and run Google's Gemma 2 language model locally on your Mac using the MLX framework. You'll learn:
What Google Gemma 2 is and its variants
How to convert a Hugging Face/PyTorch model to MLX
Steps to run Gemma 2 on your local machine
What is Google Gemma 2?
Gemma 2 is a family of lightweight, state-of-the-art open-source language models developed using the same technology behind Google's Gemini models. It comes in three sizes:
1. Gemma 2.6B
2. Gemma 9B
3. Gemma 27B
Each size is available in pre-trained and instruction-tuned variants.
Is Google Gemma 2 free?
Yes, Gemma 2 is completely open-source and accessible through the Hugging Face Hub.
Model Weights
Quantized for MLX: huggingface.co...
Full Precision: huggingface.co...
Additional Resources
Gemma 2 MLX Conversion Script:github.com/Bla...
Gemma 2 Transformers Implementation: github.com/hug...
Gemma 2 PyTorch Implementation: github.com/goo...
Fine-tuning Gemma Guide: unsloth.ai/blo...
Gradio App for MLX: github.com/SOS...
Connect with Me
LinkedIn: / prince-canuma
Twitter: / prince_canuma
Medium: / prince-canuma

Опубликовано:

 

15 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 7   
@MaziyarPanahi
@MaziyarPanahi 2 месяца назад
A complete walk through! Thank you. king!
@princecanuma
@princecanuma 2 месяца назад
My pleasure @MaziyarPanahi
@gokayfem
@gokayfem 2 месяца назад
lets go king!!
@princecanuma
@princecanuma 2 месяца назад
Let’s go 🚀
@skanderbegvictor6487
@skanderbegvictor6487 2 месяца назад
Subscribed, been following you on twitter, I am currently trying to write custom kernels for graph machine learning in mlx and am stuck.
@princecanuma
@princecanuma 2 месяца назад
Great to hear 👌🏽 keep up the good work
@vinaypandya7054
@vinaypandya7054 16 дней назад
@@princecanuma I was able to contribute to mlx_graphs because of this. I also created mlx-cluster for fast er random walk generations. I will keep working on it, thank you for inspiring us
Далее
Run Google Gemma 2B 7B Locally on the CPU & GPU
25:03
РЫБКА С ПИВОМ
00:39
Просмотров 626 тыс.
REALITY vs Apple’s Memory Claims | vs RTX4090m
8:53
Просмотров 176 тыс.
Running Gemma using HuggingFace Transformers or Ollama
23:51
Introducing Gemma 2 for developers and researchers
6:24
Google Gemma-2: Technical Report Deep Dive
14:04
Просмотров 3,7 тыс.
Apple M3 Max MLX beats RTX4090m
10:24
Просмотров 75 тыс.
Coding LLaMA-2 from scratch in PyTorch - Part 1
1:08:17
Просмотров 1,4 тыс.
РЫБКА С ПИВОМ
00:39
Просмотров 626 тыс.