Very well done explaining the uses of this tool! I believe future version of it will change how we go about everything, and a big step will be the ability to visualize and interact with these connections. I wanted to let you (and whoever sees this) know that I've collaborated with Brian (Creator of Smart Connections) and released the official plugin companion: "Smart Connections Visualizer". For this first version, it's an obsidian graph view that shows relevant connections to your current note. I'll still be adding much much more to it, also with an ability to customize how you see things like no other! Give my channel a peak if you want to find out more If it's not too much, could you pin this comment to let people know about this tool to enhance their experience with SC?
This is good Working tutorial will help viewers to understand more deeply nuances and start applying the concepts learnt. And some of them will come back and enrich every one with their experience
Hello, I don't know the principle of the pipeline, I wonder how can I establish a didicated law's gpt, I just upload some pdf, and each is at about 10Mb,but I failed to let gpt answer my quetion even if the pdf are so precise.
Thanks! And thank you for watching! I like Langsmith but if you’re on an old version of Langchain I think it won’t be compatible with the current method usage as the API has changed a lot. Especially with old workarounds from 7 months ago. Couldn’t even get the old app building so it was unlikely I’d be able to add tooling on top. But thanks for the recommendation!
My vault is over 3million words across 2k+ files, all geared towards one project. I was excited to use the Co-Pilot plugin because it advertises "vault mode", but was quickly disappointed when its default reference amount was only 3 notes/files at a time (lol), and can only stretch to 10 but gives a warning that it will prolly screw up the responses. I want to communicate with my vault as a whole for perfect macro context, but it seems my case use is still not possible with current Ai? SmartConnections doesn't seem to be any better, or am I mistaken?
Sweet! That’s a good size! Smart Connections is similar but breaks files into blocks. You’ll pull several related references from within the context of some files which performs better but may take things out of context. Seems like it would perform better This problem is actually generally not so much about what AI is capable of today, it’s that general (solve all) solutions often don’t perfectly fit the problem space or some specific domain. Early days, and everyone is still figuring it out. Some new models can handle a lot of context but no infinite search over your notes yet but one solution could exist that might make you not care about it much
We had a well crafted GPT4 prompt with many tests covering our desired outputs. We took gpt35 and fine tuned it and now it's performing the same. Worked well for our use case!
dude try obsidian copilot can use ollama, with that can use dolphin mistral = it's uncensored AI so there is no guardrail, this way we can be anything creative with texts.
Thanks for the suggestion! I like the look of that but don’t see a lot of activity on the GitHub. Still seems nice to be able to bring in other models, especially local models Dolphin has been pretty fun to play with too
Hi! Good video! I am wondering if you are using any of the openai service essentially you are not running the model locally right? Is there a local embedding and llm available for this, such as llama2 or mistral. Thanks! Awesome video.
Thanks! Glad you enjoyed it! Local embedding are definitely available. Local models currently don’t seem supported but you can proxy requests in these plugins and that could make something like that work too
Yes! I’m ready to upload my decades of original notes and content into Obsidian and instruct my LLM to cull through them to create new material based on my own source material. Come on, pick my brain.
Aye, very insightful, and was something I also picked up from using Stable Diffusion a lot. One thing that's helped a lot with prompting ChatGPT is the '-' dash seperator. It's a good way to make sure words are often not linked together as tokens, as 'Space Dash Space' is often just 1 seperate token in and of itself. If there are two seemingly odd words that I want to seperate in a prompt, it seems to give decent indication. I would say with json formatting, this can actually be pretty decent, as it's a clean well understood set of characters with clear meanings, [1,2,3] or whatever value is probably easily interpritable.
Oh this is just what I was looking for! A really great breakdown of this, thank you bud. Didn't even realise that something like Obsidian existed... Here I am doing api calls through vscode like a pleb, definitely time to move!
I would say you are actually looking for Embeddings. You can set up a database with Embeddings based on our specific data which will be checked for similarities. The matches would then be used to create the context for the completions api. Fine tuning is more to modify the way how it answers. This was my understanding.
Can the content for training be collected from ChatGPT-4? For example, after chatting with ChatGPT-4, can the desired content be filtered and integrated into ChatGPT-3.5 for fine-tuning? Is this approach feasible and effective? Are there any considerations to keep in mind?
This struck me too, it seems not many quite noticed the gravity of the massive context. Not sure if you saw it, but there's a paper on arxiv with the title "World Model on Million-Length Video And Language With RingAttention" published a day or two before Sora and google's gemma 1.5 were announced. They show results with near perfect retrieval.
Question - as it relates to the needle in the haystack performance the research you quote is very recent work but unfortunately this damn space moves so freaking fast. Or is this even a question or maybe a statement. From what I understand in the Google blog post Gemini uses a new Mixture-of-Experts (MoE) architecture, which apparently delivers much improved processing/understanding of long context. They are quoting crazy needle in the haystack results. At the same time, GPT4 has a hard time processing moderately complex prompts (well within the token range in which it preforms well at in the needle test) so as a builder with road-rash I have a raised eyebrow! What am I missing?! :-)
You’re quite right this space moves fast! There are a lot of things we’re finding now we have access to Gemini. Specifically, the way it handles controllability and context is very interesting. I think the way these AI are measured is going to continue to evolve. The needle in a haystack is a great test but we’re going to have to see what the cost of that optimization is going to have.
i didnt get your point. You showed us a diagram with two system prompts, one before and one after the user prompts to control them. But in your code example, we didn't see that in action.
I really like the video and your style so count me in as a subscriber. But showcasing Deno inside of Jupyter notebook and trying it out for myself was a bug disappointment as Deno can't track the values between cells and therefore there is no autocomplete or intellisense at all. I see in your video that it's the same in your installation.
Thanks for the kind words! It’s definitely a mixed bag. The ability to have a consistent running environment in which to execute scripts makes iteration on AI solutions much more efficient. I found that often times I would be re-executing costly generations because I needed to return to a desired state to test some aspect of my solution. Notebooks are better at this task. But the integration still has much to be desired and I totally think the lack of type consistency, autocomplete, and other core dev features makes it difficult to appreciate or enjoy
Hi, new sub here. Do you have any video teaching how to build an small language model or using RAG to get a very simple conversation project demo going?
Thanks for the subscription! We probably won’t be doing a video creating a small language model but we may look at adding RAG videos in the future. If you’re looking for something simple, you may refer to the code I referenced in the video. That’s a good starter for RAG concepts and you can look to optimize and break down later
Very timely. I was just experimenting earlier with generating RFP responses based on past winning bids and corporate knowledge bases. The RAG is returning irrelevant docs even when there are very relevant docs. On other use cases, RAG seemed to work fairly well. I think it is to do with obscure concepts. In the RFP questions throwing off the vector distances. The buyers are after all trying to separate the good from the best. I will try some of these options as quick experiments but I’m going to have to dive deep into the vector distances, etc, and properly diagnose, build solid evaluation criteria, unit tests and monitoring KPIs for this stage of the pipeline. And try fine tuning for this use case.
Sounds like it! Glad you found value in the overview! I’d imagine RFP generation wholesale is a very tricky space due to hallucinations. RAG will definitely help you in some cases but I’d imagine it would have a hard time keeping track of a firm’s competencies in relation to winning bid competencies. I’d look to reframe towards overall strategic direction over diving into the details
Tinkering builds skills. Most people search for knowledge but its skills that they want, and skills will never be learned by watching an endless stream of social media, tutorials, lectures. It's doing vs knowing. Skills = applied knowledge = ability to do/create something
AI hype is big business. After watching a lot of AI dedicated channels for almost a year, I have learned to mainly view them with distrust. They hype everything AI to death, never honestly discussing limitations or problems, nor the solutions to the problems. Sources of good information are not mentioned, moreover they rarely have insight into what works and doesn't. Just superficial regurgitation of press announcements. This isn't true of everyone, but many.
Using chat gpt 4 its a odd mix of mid blowing capabilities, nuance, depth, subtlety to its answers but then often just really dumb predictive text where it has zero understanding
Great work! Excellent user experience :) It makes the AI image-generation process less daunting. One of the biggest challenges I encountered generating images is to generate multiple images with a consistent character. I usually ask the AI to provide the seed ID of a preferred image and generate a new one with the same seed ID. Sometimes it works, sometimes it doesn't. It would be cool to see if the Stable Canvas can solve a similar problem.
That's a really interesting idea @larsfaye292 We have a set of features in mind related to using prior images but are still exploring how best to incorporate them. Some analysis of the images that are generated to pull out styles could be a cool approach and it sounds plausible!
🎯 Key Takeaways for quick navigation: 00:00 AI *agents aren't replacing developers.* 01:29 AI *lacks real-world ambiguity.* 03:03 Development *involves people.* 04:36 AI *faces challenges in adaptation.* 08:15 Technology *shifts, but job needs persist.* Made with HARPA AI
This video and opinion is fairly incorrect in regards to fine-tuning. Especially, fine-tuning can be utilized in any deep learning hyperparameters (i.e. GenAI, Discriminative AI, BERT, NLP) with any data set. Self supervision, supervised, to reinforcement learning just to name a few use cases of algorithms to solve a problem. RU-vids fine-tuning in their algorithm made me stumble upon this video. Highly recommended re-evaluation of this video to save folks from misunderstanding.
You're right that the API is much more flexible, but bear in mind that ChatGPT is both cheaper (not charging per-token), and much easier to use multi-modally. Also, you can get better control of the context by knowing when and when not to start a new chat session.
I would like to get your advice for creating conversational chatbot. Do RAG or Finetune be suitable because we have a CourtLAW based dataset that contains 1000's of PDF which is unstructured dataset of paragraphs?
Hello I am also trying to feed pdf data as an input to openai its unstructured set of data and ai is not able to process it correctly when I ask it to list transactions in pdf that it generated garbage values and not the actual values that are in pdf I am tired of giving prompts so I am looking forward to fine tune now