1#
2# IP Virtual Server configuration
3#
4menuconfig IP_VS
5	tristate "IP virtual server support"
6	depends on NET && INET && NETFILTER
7	depends on (NF_CONNTRACK || NF_CONNTRACK=n)
8	---help---
9	  IP Virtual Server support will let you build a high-performance
10	  virtual server based on cluster of two or more real servers. This
11	  option must be enabled for at least one of the clustered computers
12	  that will take care of intercepting incoming connections to a
13	  single IP address and scheduling them to real servers.
14
15	  Three request dispatching techniques are implemented, they are
16	  virtual server via NAT, virtual server via tunneling and virtual
17	  server via direct routing. The several scheduling algorithms can
18	  be used to choose which server the connection is directed to,
19	  thus load balancing can be achieved among the servers.  For more
20	  information and its administration program, please visit the
21	  following URL: <http://www.linuxvirtualserver.org/>.
22
23	  If you want to compile it in kernel, say Y. To compile it as a
24	  module, choose M here. If unsure, say N.
25
26if IP_VS
27
28config	IP_VS_IPV6
29	bool "IPv6 support for IPVS"
30	depends on IPV6 = y || IP_VS = IPV6
31	---help---
32	  Add IPv6 support to IPVS. This is incomplete and might be dangerous.
33
34	  See http://www.mindbasket.com/ipvs for more information.
35
36	  Say N if unsure.
37
38config	IP_VS_DEBUG
39	bool "IP virtual server debugging"
40	---help---
41	  Say Y here if you want to get additional messages useful in
42	  debugging the IP virtual server code. You can change the debug
43	  level in /proc/sys/net/ipv4/vs/debug_level
44
45config	IP_VS_TAB_BITS
46	int "IPVS connection table size (the Nth power of 2)"
47	range 8 20
48	default 12
49	---help---
50	  The IPVS connection hash table uses the chaining scheme to handle
51	  hash collisions. Using a big IPVS connection hash table will greatly
52	  reduce conflicts when there are hundreds of thousands of connections
53	  in the hash table.
54
55	  Note the table size must be power of 2. The table size will be the
56	  value of 2 to the your input number power. The number to choose is
57	  from 8 to 20, the default number is 12, which means the table size
58	  is 4096. Don't input the number too small, otherwise you will lose
59	  performance on it. You can adapt the table size yourself, according
60	  to your virtual server application. It is good to set the table size
61	  not far less than the number of connections per second multiplying
62	  average lasting time of connection in the table.  For example, your
63	  virtual server gets 200 connections per second, the connection lasts
64	  for 200 seconds in average in the connection table, the table size
65	  should be not far less than 200x200, it is good to set the table
66	  size 32768 (2**15).
67
68	  Another note that each connection occupies 128 bytes effectively and
69	  each hash entry uses 8 bytes, so you can estimate how much memory is
70	  needed for your box.
71
72	  You can overwrite this number setting conn_tab_bits module parameter
73	  or by appending ip_vs.conn_tab_bits=? to the kernel command line
74	  if IP VS was compiled built-in.
75
76comment "IPVS transport protocol load balancing support"
77
78config	IP_VS_PROTO_TCP
79	bool "TCP load balancing support"
80	---help---
81	  This option enables support for load balancing TCP transport
82	  protocol. Say Y if unsure.
83
84config	IP_VS_PROTO_UDP
85	bool "UDP load balancing support"
86	---help---
87	  This option enables support for load balancing UDP transport
88	  protocol. Say Y if unsure.
89
90config	IP_VS_PROTO_AH_ESP
91	def_bool IP_VS_PROTO_ESP || IP_VS_PROTO_AH
92
93config	IP_VS_PROTO_ESP
94	bool "ESP load balancing support"
95	---help---
96	  This option enables support for load balancing ESP (Encapsulation
97	  Security Payload) transport protocol. Say Y if unsure.
98
99config	IP_VS_PROTO_AH
100	bool "AH load balancing support"
101	---help---
102	  This option enables support for load balancing AH (Authentication
103	  Header) transport protocol. Say Y if unsure.
104
105config  IP_VS_PROTO_SCTP
106	bool "SCTP load balancing support"
107	select LIBCRC32C
108	---help---
109	  This option enables support for load balancing SCTP transport
110	  protocol. Say Y if unsure.
111
112comment "IPVS scheduler"
113
114config	IP_VS_RR
115	tristate "round-robin scheduling"
116	---help---
117	  The robin-robin scheduling algorithm simply directs network
118	  connections to different real servers in a round-robin manner.
119
120	  If you want to compile it in kernel, say Y. To compile it as a
121	  module, choose M here. If unsure, say N.
122
123config	IP_VS_WRR
124	tristate "weighted round-robin scheduling"
125	select GCD
126	---help---
127	  The weighted robin-robin scheduling algorithm directs network
128	  connections to different real servers based on server weights
129	  in a round-robin manner. Servers with higher weights receive
130	  new connections first than those with less weights, and servers
131	  with higher weights get more connections than those with less
132	  weights and servers with equal weights get equal connections.
133
134	  If you want to compile it in kernel, say Y. To compile it as a
135	  module, choose M here. If unsure, say N.
136
137config	IP_VS_LC
138        tristate "least-connection scheduling"
139	---help---
140	  The least-connection scheduling algorithm directs network
141	  connections to the server with the least number of active
142	  connections.
143
144	  If you want to compile it in kernel, say Y. To compile it as a
145	  module, choose M here. If unsure, say N.
146
147config	IP_VS_WLC
148        tristate "weighted least-connection scheduling"
149	---help---
150	  The weighted least-connection scheduling algorithm directs network
151	  connections to the server with the least active connections
152	  normalized by the server weight.
153
154	  If you want to compile it in kernel, say Y. To compile it as a
155	  module, choose M here. If unsure, say N.
156
157config	IP_VS_LBLC
158	tristate "locality-based least-connection scheduling"
159	---help---
160	  The locality-based least-connection scheduling algorithm is for
161	  destination IP load balancing. It is usually used in cache cluster.
162	  This algorithm usually directs packet destined for an IP address to
163	  its server if the server is alive and under load. If the server is
164	  overloaded (its active connection numbers is larger than its weight)
165	  and there is a server in its half load, then allocate the weighted
166	  least-connection server to this IP address.
167
168	  If you want to compile it in kernel, say Y. To compile it as a
169	  module, choose M here. If unsure, say N.
170
171config  IP_VS_LBLCR
172	tristate "locality-based least-connection with replication scheduling"
173	---help---
174	  The locality-based least-connection with replication scheduling
175	  algorithm is also for destination IP load balancing. It is
176	  usually used in cache cluster. It differs from the LBLC scheduling
177	  as follows: the load balancer maintains mappings from a target
178	  to a set of server nodes that can serve the target. Requests for
179	  a target are assigned to the least-connection node in the target's
180	  server set. If all the node in the server set are over loaded,
181	  it picks up a least-connection node in the cluster and adds it
182	  in the sever set for the target. If the server set has not been
183	  modified for the specified time, the most loaded node is removed
184	  from the server set, in order to avoid high degree of replication.
185
186	  If you want to compile it in kernel, say Y. To compile it as a
187	  module, choose M here. If unsure, say N.
188
189config	IP_VS_DH
190	tristate "destination hashing scheduling"
191	---help---
192	  The destination hashing scheduling algorithm assigns network
193	  connections to the servers through looking up a statically assigned
194	  hash table by their destination IP addresses.
195
196	  If you want to compile it in kernel, say Y. To compile it as a
197	  module, choose M here. If unsure, say N.
198
199config	IP_VS_SH
200	tristate "source hashing scheduling"
201	---help---
202	  The source hashing scheduling algorithm assigns network
203	  connections to the servers through looking up a statically assigned
204	  hash table by their source IP addresses.
205
206	  If you want to compile it in kernel, say Y. To compile it as a
207	  module, choose M here. If unsure, say N.
208
209config	IP_VS_SED
210	tristate "shortest expected delay scheduling"
211	---help---
212	  The shortest expected delay scheduling algorithm assigns network
213	  connections to the server with the shortest expected delay. The
214	  expected delay that the job will experience is (Ci + 1) / Ui if
215	  sent to the ith server, in which Ci is the number of connections
216	  on the ith server and Ui is the fixed service rate (weight)
217	  of the ith server.
218
219	  If you want to compile it in kernel, say Y. To compile it as a
220	  module, choose M here. If unsure, say N.
221
222config	IP_VS_NQ
223	tristate "never queue scheduling"
224	---help---
225	  The never queue scheduling algorithm adopts a two-speed model.
226	  When there is an idle server available, the job will be sent to
227	  the idle server, instead of waiting for a fast one. When there
228	  is no idle server available, the job will be sent to the server
229	  that minimize its expected delay (The Shortest Expected Delay
230	  scheduling algorithm).
231
232	  If you want to compile it in kernel, say Y. To compile it as a
233	  module, choose M here. If unsure, say N.
234
235comment 'IPVS application helper'
236
237config	IP_VS_FTP
238  	tristate "FTP protocol helper"
239        depends on IP_VS_PROTO_TCP && NF_CONNTRACK && NF_NAT
240	select IP_VS_NFCT
241	---help---
242	  FTP is a protocol that transfers IP address and/or port number in
243	  the payload. In the virtual server via Network Address Translation,
244	  the IP address and port number of real servers cannot be sent to
245	  clients in ftp connections directly, so FTP protocol helper is
246	  required for tracking the connection and mangling it back to that of
247	  virtual service.
248
249	  If you want to compile it in kernel, say Y. To compile it as a
250	  module, choose M here. If unsure, say N.
251
252config	IP_VS_NFCT
253	bool "Netfilter connection tracking"
254	depends on NF_CONNTRACK
255	---help---
256	  The Netfilter connection tracking support allows the IPVS
257	  connection state to be exported to the Netfilter framework
258	  for filtering purposes.
259
260config	IP_VS_PE_SIP
261	tristate "SIP persistence engine"
262        depends on IP_VS_PROTO_UDP
263	depends on NF_CONNTRACK_SIP
264	---help---
265	  Allow persistence based on the SIP Call-ID
266
267endif # IP_VS
268