Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Code Project
  1. Home
  2. General Programming
  3. C#
  4. Storing Huge volume of data for processing...

Storing Huge volume of data for processing...

Scheduled Pinned Locked Moved C#
questionhelp
6 Posts 4 Posters 0 Views 1 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • A Offline
    A Offline
    amatbrewer
    wrote on last edited by
    #1

    Ok, thanks to some help, I think I have what I need to extract the data I need from a HUGE (900Mb-2Gb) text file. Now I need to store the huge volume of data so that I can process it. Its only 5 fields, but lots of entries. I will need to group the data based upon various fields so I can bin and plot it. I will also need to do things like MIN/Max/Average of one field for groups of another field. So the question is what type of data structure do I put it into? Any suggestions?

    David Wilkes

    C E 2 Replies Last reply
    0
    • A amatbrewer

      Ok, thanks to some help, I think I have what I need to extract the data I need from a HUGE (900Mb-2Gb) text file. Now I need to store the huge volume of data so that I can process it. Its only 5 fields, but lots of entries. I will need to group the data based upon various fields so I can bin and plot it. I will also need to do things like MIN/Max/Average of one field for groups of another field. So the question is what type of data structure do I put it into? Any suggestions?

      David Wilkes

      C Offline
      C Offline
      Colin Angus Mackay
      wrote on last edited by
      #2

      amatbrewer wrote:

      So the question is what type of data structure do I put it into? Any suggestions?

      A database. It is designed specifically to hold and process large volumes of data like this.


      Upcoming events: * Glasgow Geek Dinner (5th March) * Glasgow: Tell us what you want to see in 2007 My: Website | Blog | Photos

      D 1 Reply Last reply
      0
      • C Colin Angus Mackay

        amatbrewer wrote:

        So the question is what type of data structure do I put it into? Any suggestions?

        A database. It is designed specifically to hold and process large volumes of data like this.


        Upcoming events: * Glasgow Geek Dinner (5th March) * Glasgow: Tell us what you want to see in 2007 My: Website | Blog | Photos

        D Offline
        D Offline
        Dan Neely
        wrote on last edited by
        #3

        and with several gigs of data, that DB should be something other than access. :)

        -- Rules of thumb should not be taken for the whole hand.

        1 Reply Last reply
        0
        • A amatbrewer

          Ok, thanks to some help, I think I have what I need to extract the data I need from a HUGE (900Mb-2Gb) text file. Now I need to store the huge volume of data so that I can process it. Its only 5 fields, but lots of entries. I will need to group the data based upon various fields so I can bin and plot it. I will also need to do things like MIN/Max/Average of one field for groups of another field. So the question is what type of data structure do I put it into? Any suggestions?

          David Wilkes

          E Offline
          E Offline
          Ennis Ray Lynch Jr
          wrote on last edited by
          #4

          Ie every time your app runs the data is different I wouldn't recommend a database. When parsing the national do not call registry I wrote a simple binary search to traverse the massive mulit-gig file to return entries by area code. Putting it into SQL Server took 2 days. Based on the need and my knowledge of algorithms no db was a better answer. Of course depending on how complicated you get will depend on when a db becomes a valid choice.


          File Not Found

          A 1 Reply Last reply
          0
          • E Ennis Ray Lynch Jr

            Ie every time your app runs the data is different I wouldn't recommend a database. When parsing the national do not call registry I wrote a simple binary search to traverse the massive mulit-gig file to return entries by area code. Putting it into SQL Server took 2 days. Based on the need and my knowledge of algorithms no db was a better answer. Of course depending on how complicated you get will depend on when a db becomes a valid choice.


            File Not Found

            A Offline
            A Offline
            amatbrewer
            wrote on last edited by
            #5

            Ennis Ray Lynch, Jr. wrote:

            Ie every time your app runs the data is different I wouldn't recommend a database.

            :omg: Thanks!!! I had not thought of that part. The data will be 100% new each time. I basically need short term data storage that I can build and access quickly, and that is gone once I am done. We are only talkign about 6 fields (all numbers with one being in Octial format). But the quantity of data is Huge! I am working on creating an object that will read the file, parse out the data I need, and store it so I can perform functions on the data. This is an entirely new area for me and I am open to any ideas.

            David Wilkes

            E 1 Reply Last reply
            0
            • A amatbrewer

              Ennis Ray Lynch, Jr. wrote:

              Ie every time your app runs the data is different I wouldn't recommend a database.

              :omg: Thanks!!! I had not thought of that part. The data will be 100% new each time. I basically need short term data storage that I can build and access quickly, and that is gone once I am done. We are only talkign about 6 fields (all numbers with one being in Octial format). But the quantity of data is Huge! I am working on creating an object that will read the file, parse out the data I need, and store it so I can perform functions on the data. This is an entirely new area for me and I am open to any ideas.

              David Wilkes

              E Offline
              E Offline
              Ennis Ray Lynch Jr
              wrote on last edited by
              #6

              Use streams and only persist the results to permanent storage. You may have to fiddle if you havent taken a good Data Structures and Algorithms class but it can be done reasonably fast.


              File Not Found

              1 Reply Last reply
              0
              Reply
              • Reply as topic
              Log in to reply
              • Oldest to Newest
              • Newest to Oldest
              • Most Votes


              • Login

              • Don't have an account? Register

              • Login or register to search.
              • First post
                Last post
              0
              • Categories
              • Recent
              • Tags
              • Popular
              • World
              • Users
              • Groups