Тёмный

5 Secrets for making PostgreSQL run BLAZING FAST. How to improve database performance. 

Dreams of Code
Подписаться 120 тыс.
Просмотров 89 тыс.
50% 1

There's an approach in here for everyone! PostgreSQL is one of the most versatile SQL databases but often does not provide performance out of the box.
Fortunately, there are some easy approaches to improve the speed of database queries when it comes to PostgreSQL. In this video, we look at 5 different approaches which can be used based on your requirements.
#postgresql #database #optimization
Become a better developer in 4 minutes: bit.ly/45C7a29 👈
Source code: github.com/dreamsofcode-io/po...
My socials:
Twitter: / dreamsofcode_io
Discord Server: / discord
Please consider supporting me as well!
Patreon: / dreamsofcode

Наука

Опубликовано:

 

1 июн 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 76   
@HrHaakon
@HrHaakon 11 месяцев назад
Prepared statements should not be used because they're faster. They should be used because they're much safer. The speed increase is just a free bonus.
@dreamsofcode
@dreamsofcode 11 месяцев назад
This is a great point.
@tipeon
@tipeon 7 месяцев назад
As a Java developer, I don't even remember the last time I used an unprepared statement.
@HrHaakon
@HrHaakon 7 месяцев назад
@@tipeon I have been reliably informed that Oracle have imprisoned a Night Elf in their compiler and whenever it sees a java.sql.Statement, he screams "YOU ARE NOT PREPARED", teleports to it and starts wailing. It may be true, but I haven't used a PreparedStatement since Blizzard released Heroes of the Storm, so...
@tizian_heckmann
@tizian_heckmann 6 месяцев назад
While that is correct, I am pretty sure prepared statements were initially developed for performance.
@luispalacios27
@luispalacios27 Год назад
I discovered your channel a few days ago; it's been amazing. Keep up the excellent work.
@dreamsofcode
@dreamsofcode Год назад
Thank you! I appreciate that a lot
@makkusu3866
@makkusu3866 Год назад
Really well made video. Staying here for more!
@HideBuz
@HideBuz Месяц назад
Bravo, subbed!
@SirJagerYT
@SirJagerYT Год назад
Suggestion for next sql video is "how to vectorize sql database for fast searching"
@dreamsofcode
@dreamsofcode Год назад
This is a great suggestion!
@wexwexexort
@wexwexexort 9 месяцев назад
​@@dreamsofcodeHave you done it
@heroe1486
@heroe1486 Год назад
Thanks for the video, very good content and well edited, I'd just recommend putting more dynamism in your voice to match the pacing
@dreamsofcode
@dreamsofcode Год назад
Thanks for the tips!
@n0kodoko143
@n0kodoko143 11 месяцев назад
Awesome video
@dreamsofcode
@dreamsofcode 11 месяцев назад
Thank you 🙏
@prathameshjoshi007
@prathameshjoshi007 9 месяцев назад
slight addition to COPY. you can use \copy from client if you don't have access to store input files on server. i.e. You can locally stream csv to server.
@leonardoplaza7677
@leonardoplaza7677 5 месяцев назад
Can you explain this in a better way?
@and_rotate69
@and_rotate69 Год назад
I want to be from ur first subscribers so when u reach a million in the next year i will comment i was here when he was getting started (i was here at 5k)
@dreamsofcode
@dreamsofcode Год назад
Haha that would be awesome
@xucongzhan9151
@xucongzhan9151 Год назад
Nice and informative vid. Care to share what mic you are using? Sounds very nice
@dreamsofcode
@dreamsofcode Год назад
Thank you! I appreciate the feedback. For this video I used the Electrovoice RE20. I also recorded in a soundtreated room this time as well which made a difference!
@xucongzhan9151
@xucongzhan9151 Год назад
@@dreamsofcode Ah, that explains! I am looking to upgrade my gear but RE20 is really out of my budget 😂 Thanks for the reply
@dreamsofcode
@dreamsofcode Год назад
@@xucongzhan9151 It's pricey! I think the Shure SM57 is pretty decent as well and much cheaper, I use that one whenever I travel!
@socks5proxy
@socks5proxy Месяц назад
Thanks!
@dreamsofcode
@dreamsofcode Месяц назад
Thank you so much!
@Fanaro
@Fanaro 9 месяцев назад
Any book recommendations on how to optimize PostgreSQL?
@pss_crs
@pss_crs 10 месяцев назад
I would like to see content possible and good way to implement multi tenant on postgres
@MrAtomUniverse
@MrAtomUniverse 9 месяцев назад
What tool are you using the terminal looks so good
@mishasawangwan6652
@mishasawangwan6652 6 месяцев назад
alacrity
@FelipeCampelo0
@FelipeCampelo0 13 дней назад
I have restricted my studies to data manipulation tasks. There is a lot to take a look on data definition and control yet!
@thepaulcraft957
@thepaulcraft957 Год назад
there is no link to your code in the video description. very interesting video!
@dreamsofcode
@dreamsofcode Год назад
Oh shoot! Thank you for letting me know. I'll fix that now.
@dreamsofcode
@dreamsofcode Год назад
@ThePaulCraft Fixed! Thank you again.
@budmonk2819
@budmonk2819 8 месяцев назад
I'm from the Oracle world, a lot of familiar concepts
@dungeon4971
@dungeon4971 Год назад
I am wondering how did you insert 20 million row into a table, where did you get that data from
@dreamsofcode
@dreamsofcode Год назад
I just randomly generated it using a mock data library in Go.
@rembautimes8808
@rembautimes8808 3 месяца назад
Joined as a sub , excellent content especially on read replicas
@gjermundification
@gjermundification 9 месяцев назад
Which language would I write a postgreSQL extension in? PL/SQL? ECMA? Python?
@dreamsofcode
@dreamsofcode 9 месяцев назад
SQL and C code are typically used for creating an extension. Mainlg SQL code and C if you need something more powerful!
@torvic99
@torvic99 8 месяцев назад
MOOOORE
@apinanyogaratnam
@apinanyogaratnam 11 месяцев назад
🔥any good resources to learn more?
@dreamsofcode
@dreamsofcode 11 месяцев назад
There's very little out there really on optimizing PostgreSQL. If it's something of interest I can dedicate some more videos into optimization!
@apinanyogaratnam
@apinanyogaratnam 11 месяцев назад
@@dreamsofcode yes pls!!!
@abdu4729
@abdu4729 11 месяцев назад
@@dreamsofcode YES that would be really helpful
@thegrumpydeveloper
@thegrumpydeveloper 7 месяцев назад
Nice! Now I don’t have to use web3 and store my data on crypto and pay per request and have huge latencies and non acid transactions. 😂
@ongayijohnian8787
@ongayijohnian8787 Месяц назад
Mahn, please do the dadbod plugins for NvChad
@doce3609
@doce3609 Год назад
chef
@ujjwalaggarwal7065
@ujjwalaggarwal7065 Месяц назад
you should add a timestamp for the copy statement part of the video
@DerClaudius
@DerClaudius 8 месяцев назад
Why would you use preared statements instead of stored procedures? They are automatically "prepared" and don't need to be recreated in every session
@Lightbeerer
@Lightbeerer 7 месяцев назад
Stored procedures are wonderful, but prepared statements have the advantage of being more dynamic in nature. Imagine e.g. a web page displaying a list with multiple columns each with different filters and sorting options. It would be a nightmare to implement with a stored procedure, but using a prepared statement you can dynamically build the necessary query.
@DerClaudius
@DerClaudius 7 месяцев назад
@@Lightbeerer Kinda disagree here. Especially with web pages, the connections/sessions are very short and only exists for the short time the page is rendered. And if you only execute the query once per request or paging request, preparing the statements make it slower. And you can absolutely implement dynamic filtering/sorting etc with an SP... and with a lot less SQL injection dangers...
@DerClaudius
@DerClaudius 7 месяцев назад
@@Lightbeerer of course it's all trade-offs but especially for web pages, preparing doesn't make sense if you don't call the query multiple times.
@tipeon
@tipeon 7 месяцев назад
With connection pooling, prepared statements make sense because the connections are actually long lived.
@DerClaudius
@DerClaudius 7 месяцев назад
@@tipeon That's not what Connection pooling is for and I would consider this bad design. Connection pooling is for mitigating the connection overhead, but you're not supposed to assume that the connection is pooled or in any state.. you should assume it's a new or in some sense resetted connection. So you would have to first ask the server if there's already a prepared statement in the session and if not, recreate it. That would make things slow again. But it's probably not even possible, because you couldn't reconnect to your prepared statement after you reconnect. Which API allows you to reconnect to an already prepared statement on the server once you let go of the statement object you held? I'm not aware of any. So for this to work you'd need to implement your own connection pooling and keep track of the statement.. and that's an even worse idea.
@halfsourlizard9319
@halfsourlizard9319 Месяц назад
Who considered Mongo 'fancy'!? I thought everyone had got over the NoSql silliness.
@SXsoft99
@SXsoft99 9 месяцев назад
people are not using indexes in an SQL service? also what people should learn about indexes if you create an index on a column the db will search faster after it if you have 3 where conditions, for example, then you need to create an index for those 3 colum combination for speed
@dreamsofcode
@dreamsofcode 9 месяцев назад
Definitely at some of the places I've worked at. Indexes are kind of interesting, they're not very useful for small data sizes, and there's the risk of over optimizing for them.
@mcspud
@mcspud 6 месяцев назад
>if you have 3 where conditions, for example, then you need to create an index for those 3 colum combination for speed Query planners are smart - if you have very large data sets you can do multi-column indexes and make sure the set reduction is in the correct order, but in my experience even up to a few billion records just having b-trees on each column individually is enough.
@2547techno
@2547techno Месяц назад
Your index explaination is not entirely correct. Postgres does offer hash-based indexes which are a lot closer to your explaination but the default index type (which you used in your creation example) is a B-Tree index, the data structure is very different. Paritions don't do anything meaningful to speed up writes, they would only speed up reads. Instead of scanning a whole table for a record, you only need to scan the single partition (assuming you're querying a single key) where you know your record lives in. It's the same concept as database sharding, but on one machine instead of multiple.
@dreamsofcode
@dreamsofcode Месяц назад
Thanks for your comment! I'm not sure what you're referencing but yes, btree is the default index which uses a binary tree for its lookup table. Partitions can speed up writes usually when asscosited with b tree indexes. One such factor is because of b tree balancing, which on partitioned table is usually less intensive than on the entire data set. Another increase in performance is when performing deletes associated with the partition column, as deleting the partition rather than deleting the rows of a table prevents rebalancing from taking place. This is common in time series data and dramatically improves write performance.
@garm0nb0z1a
@garm0nb0z1a 9 месяцев назад
Prepared statements STILL don't work with pgbouncer and most other db proxies. No thanks.
@LinhLinhBD
@LinhLinhBD 20 дней назад
A good database should be fast by default. if something requires deep knowledge to make it fast, it's a nerdy database.
@dreamsofcode
@dreamsofcode 20 дней назад
Which databases would you consider "fast by default"?
@NysShortCut
@NysShortCut 15 дней назад
So everything in programming is nerdy then cuz you need to learn to make things work. wait, that indeed makes some sense...
@LinhLinhBD
@LinhLinhBD 14 дней назад
@@dreamsofcode Mysql is faster, but lack feature. So they both not good. Mongodb is pretty god, unfortunately it's not SQL which is what most people need. That's why companies should make better databases. Something that plug and play and fast.
@roelhemerik5715
@roelhemerik5715 5 дней назад
It would be brilliant to have some database that is fast by default, but I’m afraid that is not possible in every use-case. Every choice in a database is a tradeoff. (Indexes for instance makes some reads a lot faster, but every write a lot slower…) I think the main selling point for PostgreSQL is that it is relatively easy to change these tradeoffs after you build your application.
@unitythemaker
@unitythemaker 20 дней назад
Thanks!
@dreamsofcode
@dreamsofcode 20 дней назад
Thank you so much for the support. It's really appreciated!!!
Далее
Never write another loop again (maybe)
10:48
Просмотров 248 тыс.
Solving one of PostgreSQL's biggest weaknesses.
17:12
Просмотров 171 тыс.
У каждого есть такой друг😂
00:31
The invisible ransomware virus epidemic
22:17
Database Indexing Explained (with PostgreSQL)
18:19
Просмотров 289 тыс.
The (almost) perfect Neovim setup for Node.js
20:37
Просмотров 67 тыс.
Tonic makes gRPC in Rust stupidly simple
19:08
Просмотров 40 тыс.
Using docker in unusual ways
12:58
Просмотров 387 тыс.
PostgreSQL CRASH COURSE - Learn PostgreSQL in 2022
50:26
Полезные программы для Windows
0:56
Куда пропал 3D Touch? #apple #iphone
0:51
Просмотров 695 тыс.
iPhone 12 socket cleaning #fixit
0:30
Просмотров 3,4 млн