[Top][All Lists]
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Gluster-devel] Re: ALU Scheduler ?
From: |
Angel |
Subject: |
[Gluster-devel] Re: ALU Scheduler ? |
Date: |
Thu, 17 Jan 2008 11:04:17 +0100 |
User-agent: |
KMail/1.9.1 |
Hia
You can refeer to
http://www.gluster.org/docs/index.php/GlusterFS_Translators_v1.3#ALU_Scheduler
for complete information,
ALU is Adaptative Least Usage a intelligent scheduler based on submodules
volume bricks
type cluster/unify
subvolumes brick1 brick2 brick3 brick4
option scheduler alu # use the ALU scheduler
option alu.limits.min-free-disk 5% # Don't create files one a volume
with less than 5% free diskspace
option alu.limits.max-open-files 10000 # Don't create files on a volume
with more than 10000 files open
# When deciding where to place a file, first look at the disk-usage, then at
# read-usage, write-usage, open files, and finally the disk-speed-usage.
option alu.order
disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
option alu.disk-usage.entry-threshold 2GB # Kick in if the discrepancy in
disk-usage between volumes is more than 2GB
option alu.disk-usage.exit-threshold 60MB # Don't stop writing to the
least-used volume until the discrepancy is 1988MB
option alu.open-files-usage.entry-threshold 1024 # Kick in if the
discrepancy in open files is 1024
option alu.open-files-usage.exit-threshold 32 # Don't stop until 992 files
have been written the least-used volume
# option alu.read-usage.entry-threshold 20% # Kick in when the read-usage
discrepancy is 20%
# option alu.read-usage.exit-threshold 4% # Don't stop until the discrepancy
has been reduced to 16% (20% - 4%)
# option alu.write-usage.entry-threshold 20% # Kick in when the write-usage
discrepancy is 20%
# option alu.write-usage.exit-threshold 4% # Don't stop until the discrepancy
has been reduced to 16%
# option alu.disk-speed-usage.entry-threshold # NEVER SET IT. SPEED IS
CONSTANT!!!
# option alu.disk-speed-usage.exit-threshold # NEVER SET IT. SPEED IS
CONSTANT!!!
option alu.stat-refresh.interval 10sec # Refresh the statistics used for
decision-making every 10 seconds
# option alu.stat-refresh.num-file-create 10 # Refresh the statistics used
for decision-making after creating 10 files
end-volume
El Jueves, 17 de Enero de 2008 03:15, An. Dinh Nhat escribió:
>
> I have a question.I don’t understand Parameter below.Please,you can explain
> this Parameter for me.Thanks you.
>
> * ALU Scheduler Volume example
>
> volume bricks
>
> type cluster/unify
>
> subvolumes brick1 brick2 brick3 brick4
>
> option scheduler alu # use the ALU scheduler
>
> option alu.limits.min-free-disk 5%
>
> option alu.limits.max-open-files 10000
>
> option alu.order
> disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
>
> option alu.disk-usage.entry-threshold 2GB
>
> option alu.disk-usage.exit-threshold 60MB
>
> option alu.open-files-usage.entry-threshold 1024
>
> option alu.open-files-usage.exit-threshold 32
>
> option alu.read-usage.entry-threshold 20%
>
> option alu.read-usage.exit-threshold 4%
>
> option alu.stat-refresh.interval 10sec
>
> option alu.stat-refresh.num-file-create 10
>
>
>
> Thanks & Best Regard,
>
> Đinh Nhật An
>
> System Engineer
>
> System Operation - Vinagame JSC
>
> Email:address@hidden - Yahoo:atuladn
>
> Vinagame JSC - 459B Nguyễn Đình Chiểu. Q3 , HCMC , VietNam
>
> Office phone: 8.328.426 Ext 310
>
>
>
>
>
> -----Original Message-----
> From: Angel [mailto:address@hidden
> Sent: Thursday, January 17, 2008 3:37 AM
> To: An. Dinh Nhat
> Cc: address@hidden
> Subject: Re: AFR Translator have problem
>
>
>
>
>
> I see, glusterfs developers have this point in mind on the roadmap:
>
>
>
> for the 1.4 release roadmap says:
>
>
>
> active self-heal - log and replay failed I/O transactions
>
> brick hot-add/remove/swap - live storage hardware maintenance
>
>
>
> so till that day, we the users have to figure out how to force lazy afrs into
> doing their job :-)
>
>
>
> One issue positive is that this way you can control how much resources are
> devoted to afr, the more you touch files the more replication occurs and
>
> in the event of high net or cpu pressure, lowering touching speed should
> lower afr requirements.
>
>
>
> Your mileage may vary. :-P
>
>
>
> Perhaps GlusterFS client (maybe servers) should talk to a housekeeping daemon
> to acomplish this tasks instead of over-engineering the code to do as much
> things as required..
>
>
>
> Let's wait what developers have to say about this issue...
>
>
>
> Regards Angel
>
>
>
> El Miércoles, 16 de Enero de 2008 An. Dinh Nhat escribió:
>
> > Thanks your answer.
>
> >
>
> > I understand when I touch file after server 3 go on when afr issue.Howerver
> > if I have 2 server.After I edit add one server in glusterfs-client.vol, and
> > Mount point have 40000 file,size: 800 Gb.How to AFR replication file to
> > server3 automatic?
>
> >
>
> > -----Original Message-----
>
> > From: Angel [mailto:address@hidden
>
> > Sent: Wednesday, January 16, 2008 11:16 PM
>
> > To: address@hidden
>
> > Cc: An. Dinh Nhat
>
> > Subject: Re: AFR Translator have problem
>
> >
>
> > I thinks AFR replication occurs on file access
>
> >
>
> > try to touch all files from the client and probably you will trigger
> > replication onto server3
>
> >
>
> > client --> creates files on AFR(server1,server2)
>
> >
>
> > server 3 goes up now we have AFR(server1,server2,server3)
>
> >
>
> > you wont see any files on server3
>
> >
>
> >
>
> > now touch files from the client, AFR will be triggered
>
> >
>
> > now you will see touched files on server3
>
> >
>
> > ive made similar test on local scenarios client --> local AFR(dir1,dir2)
>
> >
>
> > i copied a file test.pdf to my mountpoint and it got replicated on both
> > 'remote' dirs. Next i deleted one copy on the exported 'remote' directories
> > (dir1)
>
> > After that, i opened the pdf file on the mount point, it opened right and
> > i could see now dir1 was storing a new copy of test.pdf again.
>
> >
>
> > it seems for me looking at the code that things mostly occur on file
> > operations because xlator work intercepting fuse calls along the path to
> > posix modules.
>
> >
>
> > my tests showed things occurring like this...
>
> >
>
> >
>
> > Regards Angel
>
> > El Miércoles, 16 de Enero de 2008 Anand Avati escribió:
>
> > > Dinh,
>
> > > can you post your spec files, mentioning the order of events in terms of
>
> > > subvolumes?
>
> > >
>
> > > thanks,
>
> > > avati
>
> > >
>
> > > ---------- Forwarded message ----------
>
> > > From: An. Dinh Nhat <address@hidden>
>
> > > Date: 16-ene-2008 16:07
>
> > > Subject: AFR Translator have problem
>
> > > To: address@hidden
>
> > >
>
> > > Hi.
>
> > >
>
> > > I set up 3 server using
>
> > > GlusterFS<http://www.gluster.org/docs/index.php/GlusterFS>.Begin
>
> > > I start 2 server after From Client I mount
>
> > > GlusterFS<http://www.gluster.org/docs/index.php/GlusterFS>and copy 10
>
> > > file on volume
>
> > > gluster.After I start 'server 3' however I don't see any file on 'server
>
> > > 3',I think AFR Translator have problem.
>
> > >
>
> > >
>
> > >
>
> > > address@hidden examples]# glusterfs -V
>
> > >
>
> > > glusterfs 1.3.7 built on Dec 18 2007
>
> > >
>
> > >
>
> > >
>
> > >
>
> > >
>
> > > Thanks & Best Regard,[image: victory]
>
> > > Đinh Nhật An
>
> > > System Engineer
>
> > >
>
> > > System Operation - Vinagame JSC
>
> > > Email:address@hidden - Yahoo:atuladn
>
> > > *V*inagame *J*SC - 459B Nguyễn Đình Chiểu. Q3 , HCMC , VietNam
>
> > >
>
> > > Office phone: 8.328.426 Ext 310
>
> > >
>
> > >
>
> > >
>
> > >
>
> > >
>
> >
>
> >
>
> >
>
>
>
>
>
>
>
--
----------------------------
Clister UAH
----------------------------
[Prev in Thread] |
Current Thread |
[Next in Thread] |
- [Gluster-devel] Re: ALU Scheduler ?,
Angel <=