ML
    • Recent
    • Categories
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Exablox testing journal

    IT Discussion
    7
    75
    27.6k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • MattSpellerM
      MattSpeller
      last edited by

      So, ran into a hiccup yesterday just before I left. Received an email from Exablox pointing to this link - one for each of the Oneblox units we have. 90 minutes later I received an email for one unit, that it had "cleared" the error (link was identical, no additional info provided). During this time I noticed no difference in it's performance or behaviour - watching it sync across the "ring" to the other unit I know it was still transferring data (of some kind). Currently I still have one unit displaying the error, the other has been cleared and it has not returned (~20 hours and counting).

      Picture of the error that was displayed on the LCD screen
      IMG_20150129_170626.jpg

      1 Reply Last reply Reply Quote 0
      • scottalanmillerS
        scottalanmiller
        last edited by

        I like that LCD screen. That is cool.

        MattSpellerM 1 Reply Last reply Reply Quote 0
        • MattSpellerM
          MattSpeller @scottalanmiller
          last edited by

          @scottalanmiller Me too - it's got just enough room for important info

          1 Reply Last reply Reply Quote 2
          • MattSpellerM
            MattSpeller
            last edited by

            When I received the error I was attempting to simulate a common experience - big data dump onto the file share. To simulate this I have the "local" oneblox unit connected into our 100mbit lan properly, the other unit is connected first to a 10mbit hub to simulate the speed and quality of our connection between sites. Also on that hub is a laptop with Wireshark so I can learn more about networking and watch to see what these things do and why they're so chatty - thats next up.

            scottalanmillerS 1 Reply Last reply Reply Quote 2
            • MattSpellerM
              MattSpeller
              last edited by MattSpeller

              Chatty little boxes! These things call home quite a bit. It's not a huge quantity of traffic (around 300mb/day/oneblox = 600mb/day total for me). Keep in mind that is without anyone actually accessing them or doing anything to them. I'm not sure if that will effect their level of communication or not. This screen cap is over 24h period.

              Capture.JPG

              SeanExabloxS 2 Replies Last reply Reply Quote 0
              • scottalanmillerS
                scottalanmiller
                last edited by

                My guess, but it is only a guess, is that that number is diagnostics and will grow slightly but be generally pretty consistent over time. Maybe @SeanExablox can shed some light on that.

                SeanExabloxS 2 Replies Last reply Reply Quote 0
                • scottalanmillerS
                  scottalanmiller @MattSpeller
                  last edited by

                  @MattSpeller said:

                  When I received the error I was attempting to simulate a common experience - big data dump onto the file share. To simulate this I have the "local" oneblox unit connected into our 100mbit lan properly, the other unit is connected first to a 10mbit hub to simulate the speed and quality of our connection between sites. Also on that hub is a laptop with Wireshark so I can learn more about networking and watch to see what these things do and why they're so chatty - thats next up.

                  Sounds like some serious testing. I like where this is all going.

                  MattSpellerM 1 Reply Last reply Reply Quote 0
                  • SeanExabloxS
                    SeanExablox @scottalanmiller
                    last edited by

                    This post is deleted!
                    1 Reply Last reply Reply Quote 0
                    • SeanExabloxS
                      SeanExablox @MattSpeller
                      last edited by

                      This post is deleted!
                      1 Reply Last reply Reply Quote 1
                      • SeanExabloxS
                        SeanExablox @scottalanmiller
                        last edited by

                        @scottalanmiller

                        @scottalanmiller said:

                        My guess, but it is only a guess, is that that number is diagnostics and will grow slightly but be generally pretty consistent over time. Maybe @SeanExablox can shed some light on that.

                        @scottalanmiller you're correct. Because our management is cloud-based the heartbeat and metadata comprise the payload sent over a 24 hour period. The larger and more complex the environment (#shares, users, events, etc) the amount traffic OneBlox and OneSystem pass back and forth will increase. Having said that it should be a very modest increase in traffic and it's fairly evenly distributed throughout the day.

                        1 Reply Last reply Reply Quote 3
                        • SeanExabloxS
                          SeanExablox @MattSpeller
                          last edited by

                          @MattSpeller

                          @MattSpeller said:

                          Chatty little boxes! These things call home quite a bit. It's not a huge quantity of traffic (around 300mb/day/oneblox = 600mb/day total for me). Keep in mind that is without anyone actually accessing them or doing anything to them. I'm not sure if that will effect their level of communication or not. This screen cap is over 24h period.

                          Capture.JPG

                          @MattSpeller user access and storage capacity won't really impact the amount of data that OneBlox and OneSystem send to each other. If you do see a significant uptick, please let me know.

                          PS. apologies for the delete/repost, neglected to quote the original post.

                          1 Reply Last reply Reply Quote 3
                          • MattSpellerM
                            MattSpeller @scottalanmiller
                            last edited by

                            @scottalanmiller said:

                            Sounds like some serious testing. I like where this is all going.

                            I'd appreciate suggestions for other things to try!! I have a ton to learn about networking and this is a great opportunity. I will post about the hardware angle as well.

                            1 Reply Last reply Reply Quote 0
                            • coliverC
                              coliver
                              last edited by

                              I will be interested to see what you learn. This looks like a cool technology.

                              1 Reply Last reply Reply Quote 2
                              • scottalanmillerS
                                scottalanmiller
                                last edited by

                                I am really interested to see this running production virtualization workloads via NFS.

                                MattSpellerM SeanExabloxS 2 Replies Last reply Reply Quote 2
                                • MattSpellerM
                                  MattSpeller
                                  last edited by

                                  Ignorance ahoy! I really didn't understand at all how these things work and the following will probably illuminate that, hopefully I'll learn more!

                                  My primary concern (curiosity? frustration?) is how the bleeping heck do they claim 100% space utilization. So this means you toss in a 1TB drive, it appears with 1TB more free space for you to use. This is basically anathema to me when they claim to have your data backed up and safe - to the point where you can just yank a drive out and it'll continue on it's merry way. As we all know, with RAID to get your data backed up it's usually a 50% space penalty (RAID1 / 10). Knowing that the answer probably isn't "witchcraft" I set about trying to figure it all out. The following is my super high level guess and assumptions and I'm probably wrong about half of it.

                                  Object Based Storage is the heart of the answer, I'm pretty sure. My one sentence understand of it is: cut data into chunks, number the chunks, hash them?, store them across a couple (3?) drives. It appears that this is very similar to RAID in that the data is duped; where I know I'm missing something key is that you don't get 1:1 utilization if you're duping data anywhere. Impossible right? Yes. My best guess is that they do take a penalty from duping data, but they MUST recoup it somewhere!

                                  Taking the units I have as the working example: put in 4x4TB drives, got 14.4TB usable storage out. That's not unreasonable to lose some capacity to file system, formatting, whatever. We do know that they use de-dupe and this is primarily how I suppose they recoup the space and claim 1:1. I have checked on the management page and our units are claiming a de-dupe ratio of 1.3. According to "My Computer" I have 2.23TB used currently. Now, here is where the slight of hand comes in - I have (I hope... I didn't actually keep careful track) actually put 2.23TB of data on the thing. AH HAH! Witchcraft this is not! It's just not displaying the space it gained by compacting all your data.

                                  So, best guess: overhead penalty of duping data to maintain redundancy is recouped through compression of original data, and probably some other OBS trickery. I'm not at all sure what that trickery is exactly, though I suspect they can somehow reconstitute blocks of data??? I'm not at all satisfied with my answer but I hope to learn more.

                                  Further suppositions: I strongly suspect that there is a fair bit of overhead from all the work the OBS has to do. How much of a penalty this is I would really like to find out, but I lack the equipment (gigabit networking) to really put the hurt on them for I/O. Given my company's use case for these units, we would likely never push their I/O limitations anyway.

                                  SeanExabloxS 1 Reply Last reply Reply Quote 0
                                  • MattSpellerM
                                    MattSpeller @scottalanmiller
                                    last edited by

                                    @scottalanmiller said:

                                    I am really interested to see this running production virtualization workloads via NFS.

                                    I'd like to see that too! We are setup as SMB though 😞

                                    1 Reply Last reply Reply Quote 0
                                    • scottalanmillerS
                                      scottalanmiller
                                      last edited by

                                      Does it support SMB 3.o and, therefore, able to back HyperV?

                                      MattSpellerM SeanExabloxS 2 Replies Last reply Reply Quote 0
                                      • MattSpellerM
                                        MattSpeller @scottalanmiller
                                        last edited by

                                        @scottalanmiller No clue, but I remember something about that being a potential future feature. I checked the control panel for them and it simply stated "SMB" no version number or anything.

                                        1 Reply Last reply Reply Quote 0
                                        • MattSpellerM
                                          MattSpeller
                                          last edited by MattSpeller

                                          So the follow up to my rant below - I need a way to generate 14TB of uncompressable data and see what happens.

                                          scottalanmillerS SeanExabloxS 2 Replies Last reply Reply Quote 0
                                          • scottalanmillerS
                                            scottalanmiller @MattSpeller
                                            last edited by

                                            @MattSpeller uncompressable data is pretty tough. Video and audio files would be the easiest thing.

                                            MattSpellerM 1 Reply Last reply Reply Quote 1
                                            • 1
                                            • 2
                                            • 3
                                            • 4
                                            • 1 / 4
                                            • First post
                                              Last post