• Kausta@lemm.ee
    link
    fedilink
    English
    arrow-up
    35
    ·
    3 months ago

    You havent seen anything until you need to put a 4.2gb gzipped csv into a pandas dataframe, which works without any issues I should note.

    • thisfro@slrpnk.net
      link
      fedilink
      English
      arrow-up
      13
      ·
      3 months ago

      I raise you thousands of gzipped files (total > 20GB) combined into one dataframe. Frankly, my work laptop did not like it all that much. But most basic operations still worked fine tho

      • Kausta@lemm.ee
        link
        fedilink
        English
        arrow-up
        8
        ·
        3 months ago

        Yeah, it was just a simple example. Although using just pandas (without something like dask) for loading terabytes of data at once into a single dataframe may not be the best idea, even with enough memory.

  • QuizzaciousOtter@lemm.ee
    link
    fedilink
    English
    arrow-up
    25
    ·
    3 months ago

    Is 600 MB a lot for pandas? Of course, CSV isn’t really optimal but I would’ve sworn pandas happily works with gigabytes of data.

    • MoonHawk@lemmy.world
      link
      fedilink
      English
      arrow-up
      22
      ·
      edit-2
      3 months ago

      What do you mean not optimal? This is quite literally the most popular format for any serious data handling and exchange. One byte per separator and newline is all you need. It is not compressed so allows you to stream as well. If you don’t need tree structure it is massively better than others

      • QuizzaciousOtter@lemm.ee
        link
        fedilink
        English
        arrow-up
        9
        ·
        3 months ago

        I think portability and easy parsing is the only advantage od CSV. It’s definitely good enough (maybe even the best) for small datasets but if you have a lot of data you need a compressed binary format, something like parquet.

      • elmicha@feddit.org
        link
        fedilink
        English
        arrow-up
        6
        ·
        3 months ago

        But which separator is it, and which line ending? ASCII, UTF-8, UTF-16 or something else? What about quoting separators and line endings? Yes, there is an RFC, but a million programs were made before the RFC and won’t change their ways now.

        Also you can gzip CSV and still stream them.

      • merari42@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        3 months ago

        Have you heard that there are great serialised file formats like .parquet from appache arrow, that can easily be used in typical data science packages like duckdb or polars. Perhaps it even works with pandas (although do not know it that well. I avoid pandas as much as possible as someone who comes from the R tidyverse and try to use polars more when I work in python, because it often feels more intuitive to work with for me.)

        • driving_crooner@lemmy.eco.br
          link
          fedilink
          English
          arrow-up
          1
          ·
          3 months ago

          I used to export my pandas DataFrames as pickles, but decided to test parquet and it was great. It was like 10x smaller and allowed me to had the the databases on a server directory instead of having to copy everything to the local machine.

    • tequinhu@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      3 months ago

      It really depends on the machine that is running the code. Pandas will always have the entire thing loaded in memory, and while 600Mb is not a concern for our modern laptops running a single analysis at a time, it can get really messy if the person is not thinking about hardware limitations

    • marcos@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 months ago

      Is 600 MB a lot for pandas?

      No, but it’s easy to make a program in Python that doesn’t like it.

    • mvirts@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 months ago

      It’s more likely you’ll eat up storage when you read a 600mb parquet and try to write it as CSV.

        • mvirts@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          3 months ago

          Ah I was trying to point out that CSV is the inefficient format. Reading a large amount of data from a more efficient format like parquet is more likely to cause trouble because the memory required can be more than the file size. CSV is the opposite where it will almost always use more disk space than is required to represent the data in memory.

  • fadhl3y@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    3 months ago

    No, just buy some more RAM. 64Gb is the minimum for a professional data analyst. 128Gb, is the sweet spot.

  • FiniteBanjo@lemmy.today
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 months ago

    CSV are a cool concept. Not so much any standard but rather a text doc where values are separated by commas. Sometimes banks use them and its hell to format them for Excel. Sometimes its just a list of readable words and values.

    I had to build a Twitch Bot to add banned words in a CSV to a black list a while back, wish they would just let you copy paste like YT does.

  • Buddahriffic@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 months ago

    Did taking that picture damage that gun? It doesn’t look like the barrel is parallel to the rest of the frame (or whatever it’s called).

    Or is it deliberately angled upwards to add some automatic bullet drop compensation to the sights?