5 Secrets for making PostgreSQL run BLAZING FAST. How to improve database performance.

Поділитися
Вставка
  • Опубліковано 3 жов 2024

КОМЕНТАРІ • 88

  • @luispalacios27
    @luispalacios27 Рік тому +16

    I discovered your channel a few days ago; it's been amazing. Keep up the excellent work.

  • @HrHaakon
    @HrHaakon Рік тому +83

    Prepared statements should not be used because they're faster. They should be used because they're much safer.
    The speed increase is just a free bonus.

    • @dreamsofcode
      @dreamsofcode  Рік тому +15

      This is a great point.

    • @tipeon
      @tipeon 11 місяців тому +7

      As a Java developer, I don't even remember the last time I used an unprepared statement.

    • @HrHaakon
      @HrHaakon 11 місяців тому +7

      @@tipeon I have been reliably informed that Oracle have imprisoned a Night Elf in their compiler and whenever it sees a java.sql.Statement, he screams "YOU ARE NOT PREPARED", teleports to it and starts wailing.
      It may be true, but I haven't used a PreparedStatement since Blizzard released Heroes of the Storm, so...

    • @tizian_heckmann
      @tizian_heckmann 11 місяців тому +1

      While that is correct, I am pretty sure prepared statements were initially developed for performance.

    • @cptmorgan92
      @cptmorgan92 Місяць тому

      Our source has a lots of unprepared statements ❤ I gave up any security concerns in my company 😂 at least we just have a small number of accounts. It’s not perfect if someone is targeting us to exploit data but I haven’t written that bullshit

  • @tobyzieglerrr
    @tobyzieglerrr 24 дні тому

    Great info which was valuable to me, and your channel is top notch anyway.
    And to add some humor into the mix: Beware, if you have more microservices than users, your
    system probably does not need indexing or the rest 🙂 stay health, stay fresh, and good luck out there

  • @SirJagerYT
    @SirJagerYT Рік тому +37

    Suggestion for next sql video is "how to vectorize sql database for fast searching"

  • @gajabalaji
    @gajabalaji 3 дні тому

    Great Video. Thank you!

  • @2547techno
    @2547techno 5 місяців тому +19

    Your index explaination is not entirely correct. Postgres does offer hash-based indexes which are a lot closer to your explaination but the default index type (which you used in your creation example) is a B-Tree index, the data structure is very different.
    Paritions don't do anything meaningful to speed up writes, they would only speed up reads. Instead of scanning a whole table for a record, you only need to scan the single partition (assuming you're querying a single key) where you know your record lives in. It's the same concept as database sharding, but on one machine instead of multiple.

    • @dreamsofcode
      @dreamsofcode  5 місяців тому +3

      Thanks for your comment! I'm not sure what you're referencing but yes, btree is the default index which uses a binary tree for its lookup table.
      Partitions can speed up writes usually when asscosited with b tree indexes.
      One such factor is because of b tree balancing, which on partitioned table is usually less intensive than on the entire data set. Another increase in performance is when performing deletes associated with the partition column, as deleting the partition rather than deleting the rows of a table prevents rebalancing from taking place.
      This is common in time series data and dramatically improves write performance.

    • @MATHURIN92
      @MATHURIN92 Місяць тому

      @@dreamsofcode great point, when dealing with 100 000 000+ rows, I need to try this!

  • @enossalar8837
    @enossalar8837 3 місяці тому

    Thank you for the detailed explanation

  • @tibzdankan
    @tibzdankan 3 місяці тому

    "You can lower the roof and feel the wind in your hair", I love Dreams of Code, I love PostgreSQL

  • @makkusu3866
    @makkusu3866 Рік тому

    Really well made video. Staying here for more!

  • @prathameshjoshi007
    @prathameshjoshi007 Рік тому +3

    slight addition to COPY. you can use \copy from client if you don't have access to store input files on server. i.e. You can locally stream csv to server.

  • @xetera
    @xetera 3 місяці тому +1

    Seems kinda wild that the first point is prepared statements. It's not even a drop in the bucket for performance optimizations compared to indexing

  • @and_rotate69
    @and_rotate69 Рік тому +1

    I want to be from ur first subscribers so when u reach a million in the next year i will comment i was here when he was getting started (i was here at 5k)

  • @CloudWithUgur
    @CloudWithUgur Місяць тому

    I have found the channel ! awesome thank you

  • @HideBuz
    @HideBuz 6 місяців тому

    Bravo, subbed!

  • @halfsourlizard9319
    @halfsourlizard9319 5 місяців тому +2

    Who considered Mongo 'fancy'!? I thought everyone had got over the NoSql silliness.

  • @n0kodoko143
    @n0kodoko143 Рік тому

    Awesome video

  • @budmonk2819
    @budmonk2819 Рік тому +1

    I'm from the Oracle world, a lot of familiar concepts

  • @FelipeCampelo0
    @FelipeCampelo0 4 місяці тому

    I have restricted my studies to data manipulation tasks. There is a lot to take a look on data definition and control yet!

  • @Fanaro
    @Fanaro Рік тому +3

    Any book recommendations on how to optimize PostgreSQL?

    • @quanhua92
      @quanhua92 3 місяці тому +1

      1. the art of postgresql
      2. sql antipatterns

  • @unitythemaker
    @unitythemaker 4 місяці тому

    Thanks!

    • @dreamsofcode
      @dreamsofcode  4 місяці тому +1

      Thank you so much for the support. It's really appreciated!!!

  • @xucongzhan9151
    @xucongzhan9151 Рік тому +1

    Nice and informative vid.
    Care to share what mic you are using? Sounds very nice

    • @dreamsofcode
      @dreamsofcode  Рік тому

      Thank you! I appreciate the feedback.
      For this video I used the Electrovoice RE20. I also recorded in a soundtreated room this time as well which made a difference!

    • @xucongzhan9151
      @xucongzhan9151 Рік тому

      @@dreamsofcode Ah, that explains! I am looking to upgrade my gear but RE20 is really out of my budget 😂 Thanks for the reply

    • @dreamsofcode
      @dreamsofcode  Рік тому +1

      @@xucongzhan9151 It's pricey! I think the Shure SM57 is pretty decent as well and much cheaper, I use that one whenever I travel!

  • @heroe1486
    @heroe1486 Рік тому +1

    Thanks for the video, very good content and well edited, I'd just recommend putting more dynamism in your voice to match the pacing

  • @thepaulcraft957
    @thepaulcraft957 Рік тому +3

    there is no link to your code in the video description. very interesting video!

    • @dreamsofcode
      @dreamsofcode  Рік тому +3

      Oh shoot! Thank you for letting me know. I'll fix that now.

    • @dreamsofcode
      @dreamsofcode  Рік тому +3

      @ThePaulCraft Fixed! Thank you again.

  • @pss_crs
    @pss_crs Рік тому

    I would like to see content possible and good way to implement multi tenant on postgres

  • @ujjwalaggarwal7065
    @ujjwalaggarwal7065 5 місяців тому

    you should add a timestamp for the copy statement part of the video

  • @thegrumpydeveloper
    @thegrumpydeveloper 11 місяців тому

    Nice! Now I don’t have to use web3 and store my data on crypto and pay per request and have huge latencies and non acid transactions. 😂

  • @ongayijohnian8787
    @ongayijohnian8787 5 місяців тому

    Mahn, please do the dadbod plugins for NvChad

  • @rembautimes8808
    @rembautimes8808 7 місяців тому

    Joined as a sub , excellent content especially on read replicas

  • @torvic99
    @torvic99 Рік тому

    MOOOORE

  • @apinanyogaratnam
    @apinanyogaratnam Рік тому +1

    🔥any good resources to learn more?

    • @dreamsofcode
      @dreamsofcode  Рік тому +5

      There's very little out there really on optimizing PostgreSQL. If it's something of interest I can dedicate some more videos into optimization!

    • @apinanyogaratnam
      @apinanyogaratnam Рік тому +2

      @@dreamsofcode yes pls!!!

    • @abdu4729
      @abdu4729 Рік тому

      @@dreamsofcode YES that would be really helpful

  • @DerClaudius
    @DerClaudius Рік тому +1

    Why would you use preared statements instead of stored procedures? They are automatically "prepared" and don't need to be recreated in every session

    • @Lightbeerer
      @Lightbeerer 11 місяців тому

      Stored procedures are wonderful, but prepared statements have the advantage of being more dynamic in nature. Imagine e.g. a web page displaying a list with multiple columns each with different filters and sorting options. It would be a nightmare to implement with a stored procedure, but using a prepared statement you can dynamically build the necessary query.

    • @DerClaudius
      @DerClaudius 11 місяців тому +2

      @@Lightbeerer Kinda disagree here. Especially with web pages, the connections/sessions are very short and only exists for the short time the page is rendered. And if you only execute the query once per request or paging request, preparing the statements make it slower. And you can absolutely implement dynamic filtering/sorting etc with an SP... and with a lot less SQL injection dangers...

    • @DerClaudius
      @DerClaudius 11 місяців тому

      @@Lightbeerer of course it's all trade-offs but especially for web pages, preparing doesn't make sense if you don't call the query multiple times.

    • @tipeon
      @tipeon 11 місяців тому

      With connection pooling, prepared statements make sense because the connections are actually long lived.

    • @DerClaudius
      @DerClaudius 11 місяців тому

      @@tipeon That's not what Connection pooling is for and I would consider this bad design. Connection pooling is for mitigating the connection overhead, but you're not supposed to assume that the connection is pooled or in any state.. you should assume it's a new or in some sense resetted connection. So you would have to first ask the server if there's already a prepared statement in the session and if not, recreate it. That would make things slow again. But it's probably not even possible, because you couldn't reconnect to your prepared statement after you reconnect. Which API allows you to reconnect to an already prepared statement on the server once you let go of the statement object you held? I'm not aware of any. So for this to work you'd need to implement your own connection pooling and keep track of the statement.. and that's an even worse idea.

  • @gjermundification
    @gjermundification Рік тому

    Which language would I write a postgreSQL extension in? PL/SQL? ECMA? Python?

    • @dreamsofcode
      @dreamsofcode  Рік тому +2

      SQL and C code are typically used for creating an extension. Mainlg SQL code and C if you need something more powerful!

  • @dungeon4971
    @dungeon4971 Рік тому

    I am wondering how did you insert 20 million row into a table, where did you get that data from

    • @dreamsofcode
      @dreamsofcode  Рік тому +2

      I just randomly generated it using a mock data library in Go.

  • @doce3609
    @doce3609 Рік тому +1

    chef

  • @MrAtomUniverse
    @MrAtomUniverse Рік тому

    What tool are you using the terminal looks so good

  • @SandraWantsCoke
    @SandraWantsCoke 5 днів тому

    mongodb is web scale

  • @garm0nb0z1a
    @garm0nb0z1a Рік тому +3

    Prepared statements STILL don't work with pgbouncer and most other db proxies. No thanks.

  • @SXsoft99
    @SXsoft99 Рік тому

    people are not using indexes in an SQL service?
    also what people should learn about indexes
    if you create an index on a column the db will search faster after it
    if you have 3 where conditions, for example, then you need to create an index for those 3 colum combination for speed

    • @dreamsofcode
      @dreamsofcode  Рік тому

      Definitely at some of the places I've worked at.
      Indexes are kind of interesting, they're not very useful for small data sizes, and there's the risk of over optimizing for them.

    • @mcspud
      @mcspud 11 місяців тому +1

      >if you have 3 where conditions, for example, then you need to create an index for those 3 colum combination for speed
      Query planners are smart - if you have very large data sets you can do multi-column indexes and make sure the set reduction is in the correct order, but in my experience even up to a few billion records just having b-trees on each column individually is enough.

  • @LinhLinhBD
    @LinhLinhBD 4 місяці тому +1

    A good database should be fast by default. if something requires deep knowledge to make it fast, it's a nerdy database.

    • @dreamsofcode
      @dreamsofcode  4 місяці тому +1

      Which databases would you consider "fast by default"?

    • @nyssc
      @nyssc 4 місяці тому

      So everything in programming is nerdy then cuz you need to learn to make things work.
      wait, that indeed makes some sense...

    • @LinhLinhBD
      @LinhLinhBD 4 місяці тому

      @@dreamsofcode Mysql is faster, but lack feature. So they both not good. Mongodb is pretty god, unfortunately it's not SQL which is what most people need. That's why companies should make better databases. Something that plug and play and fast.

    • @roelhemerik5715
      @roelhemerik5715 4 місяці тому

      It would be brilliant to have some database that is fast by default, but I’m afraid that is not possible in every use-case. Every choice in a database is a tradeoff. (Indexes for instance makes some reads a lot faster, but every write a lot slower…)
      I think the main selling point for PostgreSQL is that it is relatively easy to change these tradeoffs after you build your application.

  • @awksedgreep
    @awksedgreep Місяць тому +1

    MongoDB for speed? PostgreSQL is a faster document db than mongo and it’s not even the main focus.

  • @socks5proxy
    @socks5proxy 5 місяців тому

    Thanks!