Тёмный
Mosleh Mahamud
Mosleh Mahamud
Mosleh Mahamud
Подписаться
Subscribe for regular AI content
Fine Tune Qwen 2.5 With TRL & Unsloth
9:13
9 часов назад
Fine Tune Phi 3.5 Vision With Your Data
7:24
16 часов назад
Fine Tune Qwen 2 VL With Your Data
8:26
21 день назад
Fine Tune Phi 3.5 with Your Data
7:41
Месяц назад
Fine Tune SAM 2 With Your Data
5:19
Месяц назад
I Analysed My Finances Using Local LLMs
6:05
2 месяца назад
Why GPT4-o Mini Is Better Than Llama 3.1
7:04
2 месяца назад
Building RAG With Llama 3.1
3:36
2 месяца назад
Fine Tune Llama 3.1 with Your Data
5:58
2 месяца назад
Is It Cheaper To Deploy Local LLM?
8:05
2 месяца назад
What are AI Agents and How Can It Help Me?
12:47
2 месяца назад
How to 10x Your RAG With Open Source Models
7:10
2 месяца назад
Комментарии
@Noumaan_Ahamed
@Noumaan_Ahamed 21 час назад
What about longer videos ? Won't the transcript exceed the context size of the llm ?
@moslehmahamud
@moslehmahamud 21 час назад
Good point! I have to check what else i have to do to make it work for long videos
@encoderdecoder-f8e
@encoderdecoder-f8e 4 дня назад
The guy just makes a video of things that's available on the internet and doesn't even cite the source. You are just taking someone's code without 0 revisions and you make a video to increase views?
@moslehmahamud
@moslehmahamud 4 дня назад
Spread love instead
@raomuhammadsharjeel8029
@raomuhammadsharjeel8029 6 дней назад
retriever = SelfQueryRetriever.from_llm( llm, vectorstore, document_content_description, metadata_field_info, enable_limit=True, verbose=True, ) this enable_limit will solve your problem, that retriever is returning incorrect responses.
@mrdbourke
@mrdbourke 6 дней назад
Sensational video Mosleh! The world of VLMs looks to be exploding. Will be good to be able to fine-tune these models for specific use cases.
@ManojKumar-s7p9h
@ManojKumar-s7p9h 9 дней назад
Can you tell what ollama and llm2vec is using behind the scene to extract embeddings from a decoder only model? are they processing hidden state?
@rebhuroy3713
@rebhuroy3713 9 дней назад
You are doing on pricing gpu. On free tier not possible
@rebhuroy3713
@rebhuroy3713 9 дней назад
I am trying to fine-tuning on my custom dataset but I am having 30% data loss on fine-tuning. Can you help on that
@tecnopadre
@tecnopadre 10 дней назад
Would be great to see how to use it. Thnks
@amritsubramanian8384
@amritsubramanian8384 10 дней назад
great video :)) would love to have videos on Chain of thoughts, program of thoughts and Tool Reasoning.
@moslehmahamud
@moslehmahamud 11 дней назад
Getting extreme amounts of spam comments on this video.
@LyttonRose-b5n
@LyttonRose-b5n 11 дней назад
Alexis Avenue
@GlendaHornback-u9v
@GlendaHornback-u9v 11 дней назад
Ziemann Light
@ShawnRoach-u9h
@ShawnRoach-u9h 11 дней назад
Koss Park
@DeloresAmanda-p1x
@DeloresAmanda-p1x 12 дней назад
Lawson Parkways
@EzequielMckay-z5x
@EzequielMckay-z5x 12 дней назад
Brody Parkways
@hirotrex42
@hirotrex42 12 дней назад
Thank you
@ZimmermanMark-l6d
@ZimmermanMark-l6d 12 дней назад
Waldo Drive
@ChestertonReg-f6b
@ChestertonReg-f6b 12 дней назад
Tia Courts
@tyty-v8d
@tyty-v8d 12 дней назад
Shirley Bridge
@MarionBaird-c6i
@MarionBaird-c6i 12 дней назад
Keebler Station
@WalterMaxwell-b4s
@WalterMaxwell-b4s 12 дней назад
Kreiger Spur
@RobMorris-e3v
@RobMorris-e3v 12 дней назад
Schmeler Field
@KeenedakbeKebabs-f9l
@KeenedakbeKebabs-f9l 12 дней назад
Rusty Valley
@MelindaDejoode-m7t
@MelindaDejoode-m7t 13 дней назад
Bashirian Extensions
@GladstoneNancy-i9e
@GladstoneNancy-i9e 13 дней назад
Joey Field
@LattimoreMarjorie-u5e
@LattimoreMarjorie-u5e 13 дней назад
Jaren Square
@NinaDewitt-o2j
@NinaDewitt-o2j 13 дней назад
Green Rapid
@BartlettGreg-p9w
@BartlettGreg-p9w 14 дней назад
Herzog Stravenue
@jcchoo2973
@jcchoo2973 14 дней назад
this video is completely useless. thanks for wasting my time
@SeánCarmody-y3p
@SeánCarmody-y3p 14 дней назад
Hi, great video. How can I inference the saved model I just fine tuned?
@moslehmahamud
@moslehmahamud 14 дней назад
really good question, i've been scripting it for phi 3.5 using swift. Having a hard time myself doing inferences. I've been able to export (there is an export script for swift) the model but struggling to make inference on my colab instance. Have you tried exporting and saving it on HF hub?
@SeánCarmody-y3p
@SeánCarmody-y3p 14 дней назад
​@@moslehmahamud They hid this line in those docs that I'm trying at the moment: CUDA_VISIBLE_DEVICES=0 swift infer \ --ckpt_dir output/qwen2-vl-7b-instruct/vx-xxx/checkpoint-xxx \ --load_dataset_config true --merge_lora true I think this creates a merged model that you can load in a script then like this: # Load the model model_checkpoint = "swift/output/qwen2-vl-7b-instruct/v2-20240919-150643/checkpoint-1200-merged" model = Qwen2VLForConditionalGeneration.from_pretrained(model_checkpoint, torch_dtype=torch.bfloat16, device_map="auto") # Load processor processor = AutoProcessor.from_pretrained("Qwen/Qwen2-VL-7B-Instruct")
@moslehmahamud
@moslehmahamud 14 дней назад
@@SeánCarmody-y3p nice detective work. This should be it
@SeánCarmody-y3p
@SeánCarmody-y3p 14 дней назад
@@moslehmahamud yeah it worked!
@VasanthShankar-z8e
@VasanthShankar-z8e 16 дней назад
Good Info mate!. Can you guide me how we can serve this quantized/fine tuned model as an API endpoint? I tried with vLLM ans Litserv but in both case they are not supporting the local model or our own finetuned model.
@moslehmahamud
@moslehmahamud 16 дней назад
facinating problem! why don't you send me a mail we can take it there
@VasanthShankar-z8e
@VasanthShankar-z8e 15 дней назад
@@moslehmahamud Sure, we'll do that!
@MrRaveHaven
@MrRaveHaven 17 дней назад
There are numerous other videos on RU-vid that are 20+ minutes in length. They have their merits, but yours is nice and straight to the point - with examples! Thank you!
@moslehmahamud
@moslehmahamud 17 дней назад
@@MrRaveHaven Thank you! I value my audiences time and put extra effort to make it straight to the point
@WalidBoudabbous
@WalidBoudabbous 18 дней назад
Always up to date ;)
@AIwithAniket
@AIwithAniket 18 дней назад
hi Mosleh, great video! 💜 I am one of the maintainers of LitServe and would love to talk you. Please feel free to reach out (also I am sending an email)
@niaznafirahman344
@niaznafirahman344 18 дней назад
What is your specs? I have a 4090ti gpu. Should I feel that I am low with resoources?
@moslehmahamud
@moslehmahamud 18 дней назад
Should work for a 8B model just fine, might need to quantize though.
@Rubyboat
@Rubyboat 19 дней назад
I literally lowered my batch size to 1, and im still getting memory issues
@Hmmm0135
@Hmmm0135 19 дней назад
please give notebook access you shared in description
@darionappa9706
@darionappa9706 20 дней назад
I got the error cannot pickle 'classmethod' object when using from langchain_huggingface import HuggingFaceEmbeddings
@SurajPrasad-bf9qn
@SurajPrasad-bf9qn 20 дней назад
Hi Molesh, please give me access to the notebook, I have sent the request
@rabeyatussadia6729
@rabeyatussadia6729 21 день назад
Hello, Thanks for the video. It was very informative. I requested for the code access. Would you please share it? Thanks in Advance.
@wombodombo9005
@wombodombo9005 21 день назад
nice video!
@amedyasar9468
@amedyasar9468 23 дня назад
what about dataset structure for QA llama 3.1? Could you please guide me?
@asafdelmedigo5893
@asafdelmedigo5893 25 дней назад
Hi Mosleh, thanks for sharing this straight forward tutorial! I wonder how to use a custom dataset? I'm struggeling to find the path to latex-ocr-print, also on model scope
@moslehmahamud
@moslehmahamud 25 дней назад
Thanks! You should be able to pass in a .jsonl file with your data. You can probably find latex dataset in hf hub (if i'm not completely wrong)
@abdshomad
@abdshomad 24 дня назад
​@@moslehmahamud folow up questions: how to create the custom datasets? I heard that creating synthetic dataset is faster than labelling itself? Hope there is easy to follow tutorial on it.
@philtoa334
@philtoa334 26 дней назад
Nice .
@gileneusz
@gileneusz 28 дней назад
great vid!
@davidinawe791
@davidinawe791 29 дней назад
Thanks, very helpful and straightforward
@Max-lr6dk
@Max-lr6dk 29 дней назад
i have custom data but also with an history of a conversation and i want to fine tune qwen 2 with it. Where should i write it, in the input should i wirte the history of the conversation ? if so should i use a specifi format or should i make my own ? Maybe something like that : "User 1 : hello AI: Hello" ? i can't find any answers
@aadyapipersenia668
@aadyapipersenia668 Месяц назад
please give access to the code. have requested it.
@atultiwari88
@atultiwari88 Месяц назад
hi. thank you for this awesome video. i request you to please make videos on finetuning both of these. that would be a great help. thank you.
@moslehmahamud
@moslehmahamud Месяц назад
Working on it :)
@atultiwari88
@atultiwari88 Месяц назад
@@moslehmahamud thank you
@Samuraiizen-Studio
@Samuraiizen-Studio Месяц назад
Does it support windows installation ?
@moslehmahamud
@moslehmahamud 29 дней назад
Looked through the documentation. Unsure if vllm was built for windows