Perl 6 - the future is here, just unevenly distributed

IRC log for #gluster, 2016-07-21

| Channels | #gluster index | Today | | Search | Google Search | Plain-Text | summary

All times shown according to UTC.

Time Nick Message
00:07 muneerse joined #gluster
00:13 JoeJulian gnulnx: is it replicated?
00:15 muneerse2 joined #gluster
00:16 JoeJulian Here's how I did it with a replicated volume: kill the brick server (glusterfsd) for server1:/ftp/bricks. gluster volume replace-brick $volname server1:/ftp/bricks server1:/storage/bricks commit force ; kill glusterfsd for server1:/storage/bricks ; rm -rf /storage/* ; Move my mount from /ftp to /storage ; gluster volume start $volname
00:24 gnulnx JoeJulian: Sorry, no - distributed
00:29 JoeJulian gnulnx: can you afford downtime for the files on that brick?
00:32 gnulnx JoeJulian: minimal - few hours.
00:33 JoeJulian It only takes seconds.
00:34 JoeJulian Do the same thing I described. You'll just have downtime for those files while glusterfsd is killed.
00:45 muneerse joined #gluster
00:45 dnunez joined #gluster
00:56 shdeng joined #gluster
01:19 Alghost joined #gluster
01:39 muneerse2 joined #gluster
01:45 Lee1092 joined #gluster
01:55 muneerse joined #gluster
02:15 RameshN joined #gluster
02:34 RameshN_ joined #gluster
02:41 harish_ joined #gluster
02:55 farhorizon joined #gluster
03:15 ashiq joined #gluster
03:31 kramdoss_ joined #gluster
03:33 magrawal joined #gluster
03:39 karnan joined #gluster
03:44 julim joined #gluster
03:49 atinm joined #gluster
04:03 mchangir joined #gluster
04:04 itisravi joined #gluster
04:07 ppai joined #gluster
04:08 RameshN_ joined #gluster
04:15 rafi joined #gluster
04:22 poornimag joined #gluster
04:24 nbalacha joined #gluster
04:24 kshlm joined #gluster
04:27 Vaelatern joined #gluster
04:30 ndarshan joined #gluster
04:31 atinm joined #gluster
04:35 gem joined #gluster
04:38 jiffin joined #gluster
04:46 nehar joined #gluster
04:49 atalur joined #gluster
04:51 farhorizon joined #gluster
04:59 aravindavk joined #gluster
05:03 karthik_ joined #gluster
05:08 hchiramm joined #gluster
05:13 raghug joined #gluster
05:15 aspandey joined #gluster
05:20 shubhendu joined #gluster
05:22 shubhendu joined #gluster
05:24 msvbhat joined #gluster
05:26 kdhananjay joined #gluster
05:29 hgowtham joined #gluster
05:30 sakshi joined #gluster
05:30 jiffin1 joined #gluster
05:31 kotreshhr joined #gluster
05:31 Muthu_ joined #gluster
05:32 ramky joined #gluster
05:33 prasanth joined #gluster
05:42 kdhananjay joined #gluster
05:42 jiffin1 joined #gluster
05:47 ashiq joined #gluster
05:47 atalur joined #gluster
05:53 cliluw joined #gluster
05:55 satya4ever joined #gluster
05:55 Manikandan joined #gluster
05:58 pur_ joined #gluster
05:59 jwd joined #gluster
06:01 ppai joined #gluster
06:04 nehar joined #gluster
06:07 mchangir joined #gluster
06:08 aravindavk joined #gluster
06:15 md2k joined #gluster
06:16 mhulsman joined #gluster
06:18 Bhaskarakiran joined #gluster
06:19 kovshenin joined #gluster
06:19 Bhaskarakiran joined #gluster
06:22 devyani7_ joined #gluster
06:22 spalai joined #gluster
06:27 itisravi joined #gluster
06:27 nehar joined #gluster
06:36 raghug joined #gluster
06:36 msvbhat joined #gluster
06:44 poornimag joined #gluster
06:45 muneerse2 joined #gluster
06:48 kramdoss_ joined #gluster
06:52 aravindavk joined #gluster
06:53 Seth_Karlo joined #gluster
06:54 itisravi joined #gluster
06:58 shdeng joined #gluster
07:00 muneerse joined #gluster
07:05 kshlm joined #gluster
07:08 muneerse2 joined #gluster
07:14 Seth_Karlo joined #gluster
07:15 muneerse joined #gluster
07:16 mhulsman joined #gluster
07:18 ivan_rossi joined #gluster
07:19 ivan_rossi left #gluster
07:23 Saravanakmr joined #gluster
07:27 Che-Anarch joined #gluster
07:27 derjohn_mob joined #gluster
07:33 PaulCuzner joined #gluster
07:34 raghug joined #gluster
07:39 anil_ joined #gluster
07:40 muneerse joined #gluster
07:40 mhulsman joined #gluster
07:42 pur_ joined #gluster
07:44 jww joined #gluster
07:45 jww Hello.
07:45 glusterbot jww: Despite the fact that friendly greetings are nice, please ask your question. Carefully identify your problem in such a way that when a volunteer has a few minutes, they can offer you a potential solution. These are volunteers, so be patient. Answers may come in a few minutes, or may take hours. If you're still in the channel, someone will eventually offer an answer.
07:45 jww I use gluster 3.7 with nfs export. i've set auth.allow on the volume to 1 ip, but everyone can mount the nfs export... did I missed something ?
07:47 mhulsman joined #gluster
07:50 Klas I'm only guessing here, but my guess would be that the NFS share works on a different auth then the FUSE client
07:50 poornimag joined #gluster
07:51 shubhendu_ joined #gluster
07:52 Klas http://www.gluster.org/community/documentation/index.php/Gluster_3.2:_Setting_Volume_Options#nfs.rpc-auth-allow
07:52 Klas seems to have options for nfs.rpc-auth-allow
07:52 Klas my guess is that is a good place to check
07:54 Klas in general, using the FUSE client is a better option (I think) since otherwhise there will be no HA (unless I'm mistaken)
07:54 mhulsman joined #gluster
07:57 jww thanks, I'm checking out this nfc.rpc-auth-allow !
07:58 muneerse joined #gluster
08:00 jww umm it's strange I can't see nfs.rpc-auth-allow or nfs.rpc-auth-reject in gluster volume info .
08:04 jww umm it's ok I could get / set it with gluster set volume ...
08:04 Klas generally, default options are not displayed
08:05 jww this seems to be the good option, but what should I use for auth.allow and auth.deny so ? I won't use gluster mount on clients.
08:05 Klas Disallow everything then I suppose
08:06 Klas if you don't want anyone to have access with it, don't give anyone access
08:06 Klas so, basically, you just want replicated data, not HA?
08:06 jww yes.
08:10 muneerse2 joined #gluster
08:10 Klas unless I was clear, I suppose auth.deny * should be a good option
08:10 Klas I don't remember which is prioritized of the two though
08:11 Klas meaning, I'm not certain what value you should put in allow
08:11 [Enrico] joined #gluster
08:12 jww I'll make a test.
08:12 xmuxs joined #gluster
08:12 Klas good luck =)
08:12 Klas I'm pretty much a newbie as well btw
08:14 ahino joined #gluster
08:15 xmuxs Hi All, We are running two node distibuted-replicated glusterfs 3.7 cluster. gluster volume heal info is showing split-brain entries.
08:16 xmuxs can you please advice how to remove this files from split-brain state and how to avoid split-brain
08:16 Slashman joined #gluster
08:17 Klas removing I never fully understood
08:17 Klas but avoiding split-brain needs quorum
08:17 jww Klas: hehe, still you helped me well, thanks.
08:18 Klas in a two-node replica, that means one of the the nodes must have a greater number of votes
08:18 cloph you need decide what is the real/valid copy and then copy that to the other peer
08:18 Klas which is handled automatically (the first one gets 2)
08:18 Klas this also means that the second node can't write while the first one is down though
08:18 Klas jww: hehe, hope it works =)
08:19 xmuxs so if i enable quorum when using 2 node, i will loose fault tolerance? because both nodes needs to be online?
08:19 Klas not both
08:20 Klas but the first will need to be online to allow writes
08:20 Klas that is the only way to allow split-brain
08:20 Klas a better way is to add an arbiter
08:20 Klas making a 2+1 replica
08:20 xmuxs an arbiter is third node right?
08:20 Klas that basically means that you have a third node only storing the metadata
08:21 Klas it hardly needs any disk
08:21 xmuxs So my options are: 1. server quorum, 2. client quorum, 3. add arbiter
08:21 Klas but it's a necessity to have HA and never get split-brains due to one server being down
08:22 arcolife joined #gluster
08:22 Klas I can never remember what server/client quorum means =P
08:22 Klas sorry!
08:22 Klas I understand the logic, but I suck at the terminology used by gluster
08:23 xmuxs Fuck i should i considered all of these before going live :)
08:23 Klas hehe
08:23 Klas story of every technicians life =)
08:24 Klas I live working for a university, we do get the time to actually plan shit
08:24 Klas *love
08:24 MikeLupe joined #gluster
08:25 xmuxs My task is first to remove the files that are in split-brain state because currently these files are not readable and writable.
08:25 xmuxs Second, i need to add quorum.
08:26 xmuxs If someone could assist with removing files from split-brain, i dont want to corrupt files :)
08:31 PaulCuzner joined #gluster
08:33 fsimonce joined #gluster
08:45 kramdoss_ joined #gluster
08:47 Klas cloph: do you do this directly on the servers?
08:47 mhulsman joined #gluster
08:47 Klas I never really grasped how to heal split-brain when I managed to create it
08:49 pdrakewe_ joined #gluster
08:51 cloph I also need to refer to docs to resolve it - it involves resetting attributes and I never can remember the syntax :-) https://gluster.readthedocs.io/en/latest/Troubleshooting/split-brain/
08:51 glusterbot Title: Split Brain (Manual) - Gluster Docs (at gluster.readthedocs.io)
08:53 jiffin jww: u can also give a try for http://gluster.readthedocs.io/en/latest/Administrator%20Guide/Export%20And%20Netgroup%20Authentication/
08:53 glusterbot Title: Export and Netgroup Authentication - Gluster Docs (at gluster.readthedocs.io)
08:53 shubhendu joined #gluster
08:54 cloph and /me sees there is now a easier method - woot , shows that I didn't have to resolve a split brain for quite a while :-)
09:02 itisravi cloph: Klas That document is old. You should use the methods described in http://gluster.readthedocs.io/en/latest/Troubleshooting/heal-info-and-split-brain-resolution/
09:02 glusterbot Title: Split Brain (Auto) - Gluster Docs (at gluster.readthedocs.io)
09:03 mhulsman joined #gluster
09:06 Bhaskarakiran joined #gluster
09:06 hackman joined #gluster
09:07 Bhaskarakiran joined #gluster
09:07 nehar joined #gluster
09:10 pdrakeweb joined #gluster
09:12 karthik_ joined #gluster
09:13 knightsamar joined #gluster
09:16 jww jiffin: I take a look ! thanks.
09:20 rafaels joined #gluster
09:21 Alghost joined #gluster
09:22 knightsamar hi, when striping a file across bricks, how is the size of a strip determined ? Also is the file broken into stripes while it the process is writing it or after the process has completely written it ?
09:22 archit_ joined #gluster
09:23 mhulsman joined #gluster
09:28 gem joined #gluster
09:29 muneerse joined #gluster
09:32 muneerse2 joined #gluster
09:33 mhulsman joined #gluster
09:38 Seth_Karlo joined #gluster
09:41 Manikandan joined #gluster
09:46 msvbhat joined #gluster
09:50 muneerse joined #gluster
10:00 itisravi joined #gluster
10:01 muneerse joined #gluster
10:04 kdhananjay joined #gluster
10:05 muneerse2 joined #gluster
10:12 Bhaskarakiran joined #gluster
10:18 mhulsman joined #gluster
10:21 Muthu_ joined #gluster
10:22 archit__ joined #gluster
10:32 kaushal_ joined #gluster
10:34 mhulsman1 joined #gluster
10:37 atalur joined #gluster
10:59 rafaels joined #gluster
11:22 archit_ joined #gluster
11:30 JesperA joined #gluster
11:35 mhulsman joined #gluster
11:38 johnmilton joined #gluster
11:42 raghug joined #gluster
11:43 atinm joined #gluster
11:43 xmuxs joined #gluster
11:46 ppai joined #gluster
11:46 armyriad joined #gluster
11:50 xmuxs Hi All, gluster vol heal <volume_name> info split-brain command is showing that 2 entries are in split-brain
11:51 xmuxs Directory in split-brain state shows when i list the direcotry:
11:51 xmuxs ??????????? ? ?    ?? ?? 34
11:51 xmuxs 34 is the name of directory in split-brain
11:53 mchangir joined #gluster
11:54 xmuxs When i run: getfattr: -d -m . -e hex <path_to_dir>
11:54 xmuxs i'm getting Input/output error
11:54 atalur xmuxs, are you trying to find out the replica xattrs on this directory?
11:55 xmuxs I'm trying to recover entries in split-brain
11:55 xmuxs When i run: gluster vol heal <volume_name> info split-brain
11:56 xmuxs it lists two entries both of them are directories
11:56 rastar joined #gluster
11:57 atalur ohh, okay. and are you running this " getfattr: -d -m . -e hex <path_to_dir>" on the bricks?
12:00 xmuxs sorry on mount point, its needs to run on brick directory right?
12:01 atalur yes, needs to be done on bricks
12:01 xmuxs Does this mean that just the directory is in split-brain or the directory plus all the content
12:02 atalur when you analyze the extended attributes it lists, you will be able to figure out the kind of split-brain the directory is in
12:02 atalur If any file contents were in split-brain, it would have shown up in heal info split-brain command's output
12:03 xmuxs ok many thanks, let me check and report back.
12:03 atalur okay.
12:03 atalur :)
12:08 raghug xmuxs: just to be sure, can you check the gfid of the direcotries in question on all the bricks?
12:08 kotreshhr joined #gluster
12:09 xmuxs will do and report back. Many thanks for your support :)
12:10 raghug getfattr -n trusted.gfid -e hex <backend-dir-path>
12:10 raghug Ideally gfid should be same
12:11 raghug Also check for is already associated with directory (%s). " "Hence, both directories will share same gfid " "and this can lead to inconsistencies."
12:11 raghug Also look for msg - "is already associated with directory (%s). Hence, both directories will share same gfid and this can lead to inconsistencies."
12:11 raghug in the brick logs on all the bricks
12:12 raghug xmuxs: ^^
12:14 xmuxs raghug: Thanks, will check now.
12:17 xmuxs getfattr -n trusted.gfid -e hex <backend-dir-path>
12:18 xmuxs the output is: # file: 13
12:18 xmuxs trusted.gfid=hex
12:18 xmuxs Does this mean there is 13 files in split-brain state
12:19 poornimag joined #gluster
12:19 xmuxs Do i need to compare gfid on all bricks?
12:19 B21956 joined #gluster
12:20 atalur xmuxs, could you paste the output of  getfattr -d -m . -e hex <path_to_dir> too?
12:20 xmuxs I have replica count 2, so i need to check the other nodes gfid
12:21 atalur and it doesn't mean that 13 files are in split-brain state
12:22 RameshN joined #gluster
12:22 xmuxs http://pastebin.com/ZtzpHsva
12:22 glusterbot Please use http://fpaste.org or http://paste.ubuntu.com/ . pb has too many ads. Say @paste in channel for info about paste utils.
12:22 atalur xmuxs, getfattr output shows the name of the file/directory after #file. Isn't 13 the name of the directory you are inspecting?
12:24 xmuxs atalur: sorry, yes thats the name of the directory
12:24 atalur xmuxs, :)
12:24 xmuxs gfid are the same on both node
12:25 unclemarc joined #gluster
12:25 atalur xmuxs, this doc should help you in resolving directory split-brain that you are seeing : https://gluster.readthedocs.io/en/latest/Troubleshooting/split-brain/
12:25 glusterbot Title: Split Brain (Manual) - Gluster Docs (at gluster.readthedocs.io)
12:25 atalur xmuxs, let us know if you have more questions
12:26 xmuxs atalur: Thanks, i will read the doc.
12:30 raghug joined #gluster
12:30 jiffin joined #gluster
12:33 B21956 left #gluster
12:35 mhulsman1 joined #gluster
12:36 jiffin joined #gluster
12:39 rastar joined #gluster
12:39 Manikandan joined #gluster
12:40 spalai left #gluster
12:41 ppai joined #gluster
12:41 ira_ joined #gluster
12:42 aravindavk joined #gluster
12:43 jiffin1 joined #gluster
12:46 atinm joined #gluster
12:51 [Enrico] joined #gluster
12:57 atalur_ joined #gluster
12:58 devyani7 joined #gluster
13:01 julim joined #gluster
13:02 jiffin joined #gluster
13:02 chirino joined #gluster
13:02 mhulsman joined #gluster
13:06 muneerse joined #gluster
13:10 Wizek__ joined #gluster
13:12 derjohn_mob joined #gluster
13:13 Wizek joined #gluster
13:25 muneerse2 joined #gluster
13:28 bsdjosh joined #gluster
13:33 kpease joined #gluster
13:35 squizzi joined #gluster
13:48 sanoj joined #gluster
13:51 shyam joined #gluster
13:51 jwd joined #gluster
13:54 skylar joined #gluster
13:55 dnunez joined #gluster
13:56 arcolife joined #gluster
14:00 dgandhi joined #gluster
14:03 kshlm joined #gluster
14:04 shyam joined #gluster
14:05 rwheeler joined #gluster
14:06 msvbhat joined #gluster
14:06 jiffin1 joined #gluster
14:06 farhorizon joined #gluster
14:06 mhulsman joined #gluster
14:09 arcolife joined #gluster
14:15 dnunez joined #gluster
14:15 bwerthmann joined #gluster
14:17 bowhunter joined #gluster
14:23 lpabon joined #gluster
14:26 jiffin joined #gluster
14:27 ira joined #gluster
14:28 hagarth joined #gluster
14:31 msvbhat joined #gluster
14:32 kkeithley @bug
14:32 glusterbot kkeithley: (bug <bug_id> [<bug_ids>]) -- Reports the details of the bugs with the listed ids to this channel. Accepts bug aliases as well as numeric ids. Your list can be separated by spaces, commas, and the word "and" if you want.
14:32 kkeithley @fileabug
14:32 glusterbot kkeithley: Please file a bug at http://goo.gl/UUuCq
14:33 jugaad joined #gluster
14:35 jugaad Hi everyone
14:39 pur_ joined #gluster
14:43 pur joined #gluster
14:45 derjohn_mob joined #gluster
14:54 shyam joined #gluster
14:55 ahino1 joined #gluster
14:57 farhoriz_ joined #gluster
15:01 atalur joined #gluster
15:03 chirino_m joined #gluster
15:03 jugaad Is there anyone around to help me with a performance issue with "ls"?
15:03 gem joined #gluster
15:05 lpabon joined #gluster
15:07 hagarth joined #gluster
15:08 shyam joined #gluster
15:17 paul98 joined #gluster
15:18 paul98 hi, does any one using the encryption on gluster? trying to generate a key using gluster volume set vol_name encryption.master-ley /root/volume-key.txt but ti's not generating it
15:18 paul98 i guess a success from gluster once command is run
15:29 farhorizon joined #gluster
15:31 ahino joined #gluster
15:32 kramdoss_ joined #gluster
15:35 malevolent joined #gluster
15:35 shubhendu joined #gluster
15:35 ndevos paul98: this should have a decent description of how to use it - https://kshlm.in/network-encryption-in-glusterfs/
15:35 glusterbot Title: Setting up network encryption in GlusterFS | kshlm's blog (at kshlm.in)
15:35 MikeLupe joined #gluster
15:37 shubhendu joined #gluster
15:40 farhoriz_ joined #gluster
15:41 paul98 thanks i'll have a read the docs on glusterfs site aint all that good
15:45 squizzi joined #gluster
15:46 ndevos I thought the docs for ssl were on http://gluster.readthedocs.io/en/latest/ too, but I never can find them there...
15:46 glusterbot Title: Gluster Docs (at gluster.readthedocs.io)
15:49 mhulsman joined #gluster
15:52 paul98 http://www.gluster.org/community/documentation/index.php/Features/disk-encryption is what i used
15:53 ndevos oh, that documentation is obsolete and should not even exist anymore... :-/
15:56 * misc propose to make a doc cleanup hackfest the day that jenkins and gerrit wil be migrated
15:57 paul98 ah
15:58 paul98 makes sense why it's so hard
16:00 ndevos misc: I've just send an email about it to the infra list, amye did some work already, not sure where that went though
16:01 misc ndevos: yeah, but I think that it cannot hurt to have more work, and have more people doing it
16:01 farhorizon joined #gluster
16:01 paul98 or even if it was just marked as old
16:02 ndevos misc: she (or, well Sean?) had a huge spreadsheet with links that were old/deprecated/migrated/todo/... having access to that would be a good start
16:02 ndevos paul98: yes, I was hoping to get a header on each page, saying something like "these docs are being deprecated, go to ..."
16:03 misc ndevos: yeah, and review content for old stuff, that requires some domain specific knowledge I guess
16:03 ndevos misc: yes, definitely, but there was some work done by interns to migrate the docs, and that started the RTD site, hchiramm was leading that iirc
16:03 paul98 always the wya tohugh isn't docs always get u changed and never updated
16:04 ndevos paul98: the new docs are in MarkDown and managed on GitHub, hopefuly more people feel comfortable contributing to them that way
16:05 misc mhh markdown for docs is not great, there is no common spec :/
16:05 jiffin1 joined #gluster
16:05 ndevos well, markdown, or ascii<something>
16:06 JoeJulian It's all markdown.
16:07 JoeJulian I'm hoping humble will get some time to complete the refactor branch. It's already much better organized.
16:10 jiffin joined #gluster
16:10 farhoriz_ joined #gluster
16:10 chirino_m joined #gluster
16:11 R4yTr4cer joined #gluster
16:13 paul98 i meant more in general docs always forget to get changed
16:13 paul98 even here you do the work no one ever updates docs again
16:13 jiffin joined #gluster
16:25 poornimag joined #gluster
16:28 jiffin joined #gluster
16:34 mhulsman joined #gluster
16:37 farhorizon joined #gluster
16:38 robb_nl joined #gluster
16:41 poornimag joined #gluster
16:53 jiffin joined #gluster
16:55 R4yTr4cer joined #gluster
17:01 farhorizon joined #gluster
17:03 poornimag joined #gluster
17:07 klaas joined #gluster
17:19 xmusaz joined #gluster
17:20 xmusaz Hi All, gluster vol heal is detecting two split-brain entries. can someone please assist me?
17:21 xmusaz Both of them are directories
17:22 JoeJulian xmusaz: Check the ,,(extended attributes) on both of them. Remove the trusted.afr.* attributes (this only works for directories).
17:22 glusterbot xmusaz: (#1) To read the extended attributes on the server: getfattr -m .  -d -e hex {filename}, or (#2) For more information on how GlusterFS uses extended attributes, see this article: http://pl.atyp.us/hekafs.org/index.php/2011/04/glusterfs-extended-attributes/
17:23 msvbhat joined #gluster
17:25 shyam joined #gluster
17:26 xmusaz JoeJulien: i'm using total of 12 bricks between two nodes. do i need to check attributes on all 12 bricks?
17:27 JoeJulian What does the self-heal report say?
17:29 xmusaz When i run: gluster vol heal vol_name info split-brain
17:30 xmusaz brick 1 no entries, brick 2 to 6 = 2 entries both directory
17:30 farhoriz_ joined #gluster
17:30 JoeJulian Then I guess you fix it on 2-6.
17:31 xmusaz JoeJulian: ok thanks
17:34 julim joined #gluster
17:44 farhorizon joined #gluster
17:47 farhorizon joined #gluster
17:49 kkeithley I've archived the old community documentation and placed a redirect to our readthedocs.io documentation
17:50 farhorizon joined #gluster
17:51 farhorizon joined #gluster
17:52 farhorizon joined #gluster
17:53 farhorizon joined #gluster
17:55 farhorizon joined #gluster
17:55 farhorizon joined #gluster
17:56 farhorizon joined #gluster
17:56 farhorizon joined #gluster
17:57 R4yTr4cer joined #gluster
17:57 farhorizon joined #gluster
18:01 squizzi joined #gluster
18:01 julim_ joined #gluster
18:03 bowhunter joined #gluster
18:03 shyam1 joined #gluster
18:15 arcolife joined #gluster
18:16 farhorizon joined #gluster
18:25 msvbhat joined #gluster
18:29 farhoriz_ joined #gluster
18:31 [o__o] joined #gluster
18:33 farhorizon joined #gluster
18:38 chirino_m joined #gluster
18:42 JoeJulian [10:40] <xmusaz> What is the different between these two command. 1. gluster vol heal vol_name info split-brain and 2. gluster vol heal vol_name info
18:43 JoeJulian The first shows a log of all the split-brain files that have been seen since the self-heal daemons were started. The latter only currently split-brain entries.
18:44 JoeJulian (and please don't DM me gluster questions)
18:45 shyam1 left #gluster
18:49 rastar joined #gluster
18:56 ashiq joined #gluster
18:56 mhulsman joined #gluster
19:02 farhoriz_ joined #gluster
19:14 PaulCuzner joined #gluster
19:22 ahino joined #gluster
19:29 mhulsman joined #gluster
19:33 chirino_m joined #gluster
19:52 derjohn_mob joined #gluster
20:02 md2k joined #gluster
20:32 farhorizon joined #gluster
20:36 shaunm joined #gluster
20:56 kpease joined #gluster
20:58 shyam joined #gluster
21:03 kpease joined #gluster
21:15 md2k joined #gluster
21:50 johnmilton joined #gluster
21:58 johnmilton joined #gluster
22:00 johnmilton joined #gluster
23:13 muneerse joined #gluster
23:35 hackman joined #gluster
23:57 Wizek joined #gluster

| Channels | #gluster index | Today | | Search | Google Search | Plain-Text | summary