redis ubuntu 开机自启
内容导读
互联网集市收集整理的这篇技术教程文章主要介绍了redis ubuntu 开机自启,小编现在分享给大家,供广大互联网技能从业者学习和参考。文章包含26599字,纯文字阅读大概需要38分钟。
内容图文
1 mv redis-server /etc/init.d/redis-server 2chmod +x /etc/init.d/redis-server 3mv redis.conf /etc/redis.conf
1 sudo mkdir -p /var/lib/redis 2sudomkdir -p /var/log/redis
1 update-rc.d redis-server defaults
1 #! /bin/sh 2### BEGIN INIT INFO 3 # Provides: redis-server 4 # Required-Start: $syslog 5 # Required-Stop: $syslog 6 # Should-Start: $local_fs 7 # Should-Stop: $local_fs 8 # Default-Start: 2345 9 # Default-Stop: 01610 # Short-Description: redis-server - Persistent key-value db 11 # Description: redis-server - Persistent key-value db 12### END INIT INFO 131415 PATH=/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin 16 DAEMON=/usr/local/bin/redis-server 17 DAEMON_ARGS=/etc/redis.conf 18 NAME=redis-server 19 DESC=redis-server 20 PIDFILE=/var/run/redis.pid 2122 test -x $DAEMON || exit 023 test -x $DAEMONBOOTSTRAP || exit 02425 set -e 2627case"$1"in28 start) 29echo -n "Starting $DESC: "30touch $PIDFILE 31if start-stop-daemon --start --quiet --umask 007 --pidfile $PIDFILE --exec $DAEMON --background -- $DAEMON_ARGS 32then33echo"$NAME."34else35echo"failed"36fi37 ;; 38 stop) 39echo -n "Stopping $DESC: "40if start-stop-daemon --stop --retry 10 --quiet --oknodo --pidfile $PIDFILE --exec $DAEMON 41then42echo"$NAME."43else44echo"failed"45fi46rm -f $PIDFILE 47 ;; 4849 restart|force-reload) 50 ${0} stop 51 ${0} start 52 ;; 53 *) 54echo"Usage: /etc/init.d/$NAME {start|stop|restart|force-reload}" >&255 exit 156 ;; 57esac5859 exit 0
1 # Redis configuration file example 2 # Based on the default redis.conf shipped with Redis 2.6.9 3 4# Note on units: when memory size is needed, it is possible to specify 5 # it in the usual form of 1k 5GB 4M and so forth: 6# 7 # 1k => 1000 bytes 8 # 1kb => 1024 bytes 9 # 1m => 1000000 bytes 10 # 1mb => 1024*1024 bytes 11 # 1g => 1000000000 bytes 12 # 1gb => 1024*1024*1024 bytes 13# 14 # units are case insensitive so 1GB 1Gb 1gB are all the same. 15 16 # By default Redis does not run as a daemon. Use ‘yes‘if you need it. 17 # Note that Redis will write a pid filein /var/run/redis.pid when daemonized. 18daemonize yes 19 20 # When running daemonized, Redis writes a pid filein /var/run/redis.pid by 21 # default. You can specify a custom pid file location here. 22 pidfile /var/run/redis.pid 23 24 # Accept connections on the specified port, default is 6379. 25 # If port 0 is specified Redis will not listen on a TCP socket. 26 port 6379 27 28 # If you want you can bind a single interface, if the bind option is not 29 # specified all the interfaces will listen for incoming connections. 30# 31 # bind 127.0.0.1 32 33 # Specify the path for the unix socket that will be used to listen for 34# incoming connections. There is no default, so Redis will not listen 35# on a unix socket when not specified. 36# 37 # unixsocket /tmp/redis.sock 38 # unixsocketperm 755 39 40 # Close the connection after a client is idle for N seconds (0 to disable) 41 timeout 0 42 43 # Set server verbosity to ‘debug‘ 44# it can be one of: 45 # debug (a lot of information, useful for development/testing) 46 # verbose (many rarely useful info, but not a mess like the debug level) 47 # notice (moderately verbose, what you want in production probably) 48 # warning (only very important / critical messages are logged) 49loglevel notice 50 51 # Specify the log file name. Also ‘stdout‘ can be used to force 52 # Redis to log on the standard output. Note that if you use standard 53 # output for logging but daemonize, logs will be sent to /dev/null 54 logfile /var/log/redis/redis.log 55 56 # To enable logging to the system logger, just set ‘syslog-enabled‘ to yes, 57# and optionally update the other syslog parameters to suit your needs. 58 # syslog-enabled no 59 60# Specify the syslog identity. 61 # syslog-ident redis 62 63 # Specify the syslog facility. Must be USER or between LOCAL0-LOCAL7. 64 # syslog-facility local0 65 66 # Set the number of databases. The default database is DB 0, you can select 67 # a different one on a per-connection basis using SELECT <dbid> where 68 # dbid is a number between 0 and ‘databases‘-1 69 databases 16 70 71################################ SNAPSHOTTING ################################# 72# 73# Save the DB on disk: 74# 75 # save <seconds> <changes> 76# 77 # Will save the DB if both the given number of seconds and the given 78 # number of write operations against the DB occurred. 79# 80# In the example below the behaviour will be to save: 81 # after 900 sec (15 min) if at least 1 key changed 82 # after 300 sec (5 min) if at least 10 keys changed 83 # after 60 sec if at least 10000 keys changed 84# 85 # Note: you can disable saving at all commenting all the "save" lines. 86# 87# It is also possible to remove all the previously configured save 88 # points by adding a save directive with a single empty string argument 89 # like in the following example: 90# 91 # save "" 92 93 save 9001 94 save 30010 95 save 6010000 96 97 # By default Redis will stop accepting writes if RDB snapshots are enabled 98# (at least one save point) and the latest background save failed. 99 # This will make the user aware (in an hard way) that data is not persisting 100# on disk properly, otherwise chances are that no one will notice and some 101# distater will happen. 102# 103# If the background saving process will start working again Redis will 104# automatically allow writes again. 105# 106 # However if you have setup your proper monitoring of the Redis server 107# and persistence, you may want to disable this feature so that Redis will 108 # continue to work as usually even if there are problems with disk, 109# permissions, and so forth. 110 stop-writes-on-bgsave-error yes 111112 # Compress string objects using LZF when dump .rdb databases? 113 # For default that‘s set to ‘yes‘ as it‘s almost always a win. 114 # If you want to save some CPU in the saving child set it to ‘no‘ but 115 # the dataset will likely be bigger if you have compressible values or keys. 116rdbcompression yes 117118 # Since verison 5 of RDB a CRC64 checksum is placed at the end of the file. 119 # This makes the format more resistant to corruption but there is a performance 120 # hit to pay (around 10%) when saving and loading RDB files, so you can disable it 121 # for maximum performances. 122# 123# RDB files created with checksum disabled have a checksum of zero that will 124# tell the loading code to skip the check. 125rdbchecksum yes 126127# The filename where to dump the DB 128dbfilename dump.rdb 129130# The working directory. 131# 132# The DB will be written inside this directory, with the filename specified 133 # above using the ‘dbfilename‘ configuration directive. 134# 135# Also the Append Only File will be created inside this directory. 136# 137 # Note that you must specify a directory here, not a file name. 138dir /var/lib/redis 139140################################# REPLICATION ################################# 141142 # Master-Slave replication. Use slaveof to make a Redis instance a copy of 143# another Redis server. Note that the configuration is local to the slave 144 # so for example it is possible to configure the slave to save the DB with a 145# different interval, or to listen to another port, and so on. 146# 147 # slaveof <masterip> <masterport> 148149 # If the master is password protected (using the "requirepass" configuration 150# directive below) it is possible to tell the slave to authenticate before 151# starting the replication synchronization process, otherwise the master will 152# refuse the slave request. 153# 154 # masterauth <master-password> 155156# When a slave lost the connection with the master, or when the replication 157 # is still in progress, the slave can act in two different ways: 158# 159 # 1) if slave-serve-stale-data is set to ‘yes‘ (the default) the slave will 160 # still reply to client requests, possibly with out of date data, or the 161 # data set may just be empty if this is the first synchronization. 162# 163 # 2) if slave-serve-stale data is set to ‘no‘ the slave will reply with 164 # an error "SYNC with master in progress" to all the kind of commands 165# but to INFO and SLAVEOF. 166# 167 slave-serve-stale-data yes 168169# You can configure a slave instance to accept writes or not. Writing against 170# a slave instance may be useful to store some ephemeral data (because data 171# written on a slave will be easily deleted after resync with the master) but 172 # may also cause problems if clients are writing to it because of a 173# misconfiguration. 174# 175 # Since Redis 2.6 by default slaves are read-only. 176# 177# Note: read only slaves are not designed to be exposed to untrusted clients 178 # on the internet. It‘s just a protection layer against misuse of the instance.179# Still a read only slave exports by default all the administrative commands 180# such as CONFIG, DEBUG, and so forth. To a limited extend you can improve 181 # security of read only slaves using ‘rename-command‘ to shadow all the 182 # administrative / dangerous commands. 183 slave-read-only yes 184185 # Slaves send PINGs to server in a predefined interval. It‘s possible to change186 # this interval with the repl_ping_slave_period option. The default value is 10187# seconds. 188# 189 # repl-ping-slave-period 10190191 # The following option sets a timeout for both Bulk transfer I/O timeout and 192 # master data or ping response timeout. The default value is 60 seconds. 193# 194 # It is important to make sure that this value is greater than the value 195 # specified for repl-ping-slave-period otherwise a timeout will be detected 196 # every time there is low traffic between the master and the slave. 197# 198 # repl-timeout 60199200 # The slave priority is an integer number published by Redis in the INFO output. 201 # It is used by Redis Sentinel in order to select a slave to promote into a 202 # master if the master is no longer working correctly. 203# 204 # A slave with a low priority number is considered better for promotion, so 205 # for instance if there are three slaves with priority 10, 100, 25 Sentinel will 206 # pick the one wtih priority 10, that is the lowest. 207# 208 # However a special priority of 0 marks the slave as not able to perform the 209 # role of master, so a slave with priority of 0 will never be selected by 210 # Redis Sentinel for promotion. 211# 212 # By default the priority is 100. 213 slave-priority 100214215################################## SECURITY ################################### 216217 # Require clients to issue AUTH <PASSWORD> before processing any other 218 # commands. This might be useful in environments inwhich you do not trust 219 # others with access to the host running redis-server. 220# 221 # This should stay commented out for backward compatibility and because most 222 # people do not need auth (e.g. they run their own servers). 223# 224# Warning: since Redis is pretty fast an outside user can try up to 225# 150k passwords per second against a good box. This means that you should 226# use a very strong password otherwise it will be very easy to break. 227# 228# requirepass foobared 229230# Command renaming. 231# 232 # It is possible to change the name of dangerous commands in a shared 233# environment. For instance the CONFIG command may be renamed into something 234 # of hard to guess so that it will be still available for internal-use 235 # tools but not available for general clients. 236# 237# Example: 238# 239 # rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52 240# 241 # It is also possible to completely kill a command renaming it into 242 # an empty string: 243# 244 # rename-command CONFIG ""245246################################### LIMITS #################################### 247248 # Set the max number of connected clients at the same time. By default 249 # this limit is set to 10000 clients, however if the Redis server is not 250 # able ot configure the process file limit to allow for the specified limit 251 # the max number of allowed clients is set to the current file limit 252 # minus 32 (as Redis reserves a few file descriptors for internal uses). 253# 254# Once the limit is reached Redis will close all the new connections sending 255 # an error ‘max number of clients reached‘. 256# 257 # maxclients 10000258259 # Don‘t use more memory than the specified amount of bytes.260# When the memory limit is reached Redis will try to remove keys 261 # accordingly to the eviction policy selected (see maxmemmory-policy). 262# 263 # If Redis can‘t remove keys according to the policy, or if the policy is264 # set to ‘noeviction‘, Redis will start to reply with errors to commands 265 # that would use more memory, like SET, LPUSH, and so on, and will continue 266 # to reply to read-only commands like GET. 267# 268# This option is usually useful when using Redis as an LRU cache, or to set 269 # an hard memory limit for an instance (using the ‘noeviction‘ policy). 270# 271# WARNING: If you have slaves attached to an instance with maxmemory on, 272# the size of the output buffers needed to feed the slaves are subtracted 273 # from the used memory count, so that network problems / resyncs will 274 # not trigger a loop where keys are evicted, and in turn the output 275# buffer of slaves is full with DELs of keys evicted triggering the deletion 276 # of more keys, and so forth until the database is completely emptied. 277# 278 # In short... if you have slaves attached it is suggested that you set a lower 279 # limit for maxmemory so that there is some free RAM on the system for slave 280 # output buffers (but this is not needed if the policy is ‘noeviction‘). 281# 282 # maxmemory <bytes> 283284 # MAXMEMORY POLICY: how Redis will select what to remove when maxmemory 285 # is reached? You can select among five behavior: 286# 287 # volatile-lru -> remove the key with an expire set using an LRU algorithm 288 # allkeys-lru -> remove any key accordingly to the LRU algorithm 289 # volatile-random -> remove a random key with an expire set 290 # allkeys-random -> remove a random key, any key 291 # volatile-ttl -> remove the key with the nearest expire time (minor TTL) 292 # noeviction -> don‘t expire at all, just return an error on write operations293# 294 # Note: with all the kind of policies, Redis will return an error on write295 # operations, when there are not suitable keys for eviction. 296# 297 # At the date of writing this commands are: set setnx setex append 298# incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd 299# sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby 300# zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby 301 # getset mset msetnx exec sort302# 303# The default is: 304# 305 # maxmemory-policy volatile-lru 306307# LRU and minimal TTL algorithms are not precise algorithms but approximated 308 # algorithms (in order to save memory), so you can select as well the sample 309 # size to check. For instance for default Redis will check three keys and 310 # pick the one that was used less recently, you can change the sample size 311# using the following configuration directive. 312# 313 # maxmemory-samples 3314315############################## APPEND ONLY MODE ############################### 316317# By default Redis asynchronously dumps the dataset on disk. This mode is 318 # good enough in many applications, but an issue with the Redis process or 319# a power outage may result into a few minutes of writes lost (depending on 320# the configured save points). 321# 322# The Append Only File is an alternative persistence mode that provides 323# much better durability. For instance using the default data fsync policy 324 # (see later in the config file) Redis can lose just one second of writes in a 325 # dramatic event like a server power outage, or a single writeif something 326# wrong with the Redis process itself happens, but the operating system is 327# still running correctly. 328# 329 # AOF and RDB persistence can be enabled at the same time without problems. 330 # If the AOF is enabled on startup Redis will load the AOF, that is the file331# with the better durability guarantees. 332# 333 # Please check http://redis.io/topics/persistence for more information.334335appendonly no 336337 # The name of the append only file (default: "appendonly.aof") 338# appendfilename appendonly.aof 339340 # The fsync() call tells the Operating System to actually write data on disk 341 # instead to waitformore data in the output buffer. Some OS will really flush 342 # data on disk, some other OS will just try to do it ASAP. 343# 344# Redis supports three different modes: 345# 346 # no: don‘t fsync, just let the OS flush the data when it wants. Faster.347 # always: fsync after every write to the append only log . Slow, Safest. 348 # everysec: fsync only one time every second. Compromise. 349# 350 # The default is "everysec" that‘s usually the right compromise between351 # speed and data safety. It‘s up to you to understand if you can relax this to352 # "no" that will let the operating system flush the output buffer when 353 # it wants, for better performances (but if you can live with the idea of 354 # some data loss consider the default persistence mode that‘s snapshotting),355 # or on the contrary, use "always" that‘s very slow but a bit safer than356# everysec. 357# 358# More details please check the following article: 359 # http://antirez.com/post/redis-persistence-demystified.html360# 361 # If unsure, use "everysec". 362363# appendfsync always 364appendfsync everysec 365# appendfsync no 366367# When the AOF fsync policy is set to always or everysec, and a background 368# saving process (a background save or AOF log background rewriting) is 369 # performing a lot of I/O against the disk, in some Linux configurations 370 # Redis may block too long on the fsync() call. Note that there is no fix for371 # this currently, as even performing fsync in a different thread will block 372 # our synchronous write(2) call. 373# 374 # In order to mitigate this problem it‘s possible to use the following option375 # that will prevent fsync() from being called in the main process while a 376 # BGSAVE or BGREWRITEAOF is in progress. 377# 378 # This means that while another child is saving the durability of Redis is 379 # the same as "appendfsync none", that in practical terms means that it is 380 # possible to lost up to 30 seconds of log in the worst scenario (with the 381# default Linux settings). 382# 383 # If you have latency problems turn this to "yes". Otherwise leave it as 384 # "no" that is the safest pick from the point of view of durability. 385 no-appendfsync-on-rewrite no 386387 # Automatic rewrite of the append only file. 388 # Redis is able to automatically rewrite the log file implicitly calling 389# BGREWRITEAOF when the AOF log size will growth by the specified percentage. 390# 391 # This is how it works: Redis remembers the size of the AOF file after the 392 # latest rewrite (or if no rewrite happened since the restart, the size of 393# the AOF at startup is used). 394# 395# This base size is compared to the current size. If the current size is 396# bigger than the specified percentage, the rewrite is triggered. Also 397 # you need to specify a minimal size for the AOF file to be rewritten, this 398 # is useful to avoid rewriting the AOF file even if the percentage increase 399# is reached but it is still pretty small. 400# 401 # Specify a percentage of zero in order to disable the automatic AOF 402# rewrite feature. 403404 auto-aof-rewrite-percentage 100405 auto-aof-rewrite-min-size 64mb 406407################################ LUA SCRIPTING ############################### 408409 # Max execution time of a Lua script in milliseconds. 410# 411 # If the maximum execution time is reached Redis will log that a script is 412 # still in execution after the maximum allowed time and will start to 413# reply to queries with an error. 414# 415 # When a long running script exceed the maximum execution time only the 416# SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be 417 # used to stop a script that did not yet called write commands. The second 418 # is the only way to shut down the server in the case a write commands was 419 # already issue by the script but the user don‘t want to wait for the natural420# termination of the script. 421# 422 # Set it to 0 or a negative value for unlimited execution without warnings. 423 lua-time-limit 5000424425################################## SLOW LOG ################################### 426427# The Redis Slow Log is a system to log queries that exceeded a specified 428 # execution time. The execution time does not include the I/O operations 429# like talking with the client, sending the reply and so forth, 430 # but just the time needed to actually execute the command (this is the only 431# stage of command execution where the thread is blocked and can not serve 432 # other requests in the meantime). 433# 434# You can configure the slow log with two parameters: one tells Redis 435 # what is the execution time, in microseconds, to exceed in order for the 436# command to get logged, and the other parameter is the length of the 437# slow log. When a new command is logged the oldest one is removed from the 438# queue of logged commands. 439440 # The following time is expressed in microseconds, so 1000000 is equivalent 441 # to one second. Note that a negative number disables the slow log, while442# a value of zero forces the logging of every command. 443 slowlog-log-slower-than 10000444445# There is no limit to this length. Just be aware that it will consume memory. 446# You can reclaim memory used by the slow log with SLOWLOG RESET. 447 slowlog-max-len 128448449############################### ADVANCED CONFIG ############################### 450451# Hashes are encoded using a memory efficient data structure when they have a 452# small number of entries, and the biggest entry does not exceed a given 453# threshold. These thresholds can be configured using the following directives. 454 hash-max-ziplist-entries 512455 hash-max-ziplist-value 64456457 # Similarly to hashes, small lists are also encoded in a special way in order 458# to save a lot of space. The special representation is only used when 459# you are under the following limits: 460 list-max-ziplist-entries 512461 list-max-ziplist-value 64462463 # Sets have a special encoding in just one case: when a set is composed 464 # of just strings that happens to be integers in radix 10in the range 465 # of 64 bit signed integers. 466 # The following configuration setting sets the limit in the size of the 467 # set in order to use this special memory saving encoding. 468 set-max-intset-entries 512469470 # Similarly to hashes and lists, sorted sets are also specially encoded in471# order to save a lot of space. This encoding is only used when the length and 472# elements of a sorted set are below the following limits: 473 zset-max-ziplist-entries 128474 zset-max-ziplist-value 64475476 # Active rehashing uses 1 millisecond every 100 milliseconds of CPU timein477 # order to help rehashing the main Redis hash table (the one mapping top-level 478# keys to values). The hash table implementation Redis uses (see dict.c) 479 # performs a lazy rehashing: the more operation you run into an hash table 480 # that is rehashing, the more rehashing "steps" are performed, so if the 481 # server is idle the rehashing is never complete and some more memory is used 482# by the hash table. 483# 484 # The default is to use this millisecond 10 times every second in order to 485# active rehashing the main dictionaries, freeing memory when possible. 486# 487# If unsure: 488 # use "activerehashing no"if you have hard latency requirements and it is 489 # not a good thing in your environment that Redis can reply form time to time490 # to queries with 2 milliseconds delay. 491# 492 # use "activerehashing yes"if you don‘t have such hard requirements but493 # want to free memory asap when possible. 494activerehashing yes 495496# The client output buffer limits can be used to force disconnection of clients 497 # that are not reading data from the server fast enough for some reason (a 498 # common reason is that a Pub/Sub client can‘t consume messages as fast as the499# publisher can produce them). 500# 501 # The limit can be set differently for the three different classes of clients: 502# 503 # normal -> normal clients 504 # slave -> slave clients and MONITOR clients 505 # pubsub -> clients subcribed to at least one pubsub channel or pattern 506# 507 # The syntax of every client-output-buffer-limit directive is the following: 508# 509 # client-output-buffer-limit <class> <hard limit> <soft limit> <soft seconds> 510# 511 # A client is immediately disconnected once the hard limit is reached, or if512 # the soft limit is reached and remains reached for the specified number of 513# seconds (continuously). 514 # So for instance if the hard limit is 32 megabytes and the soft limit is 515 # 16 megabytes / 10 seconds, the client will get disconnected immediately 516 # if the size of the output buffers reach 32 megabytes, but will also get 517 # disconnected if the client reaches 16 megabytes and continuously overcomes 518 # the limit for10 seconds. 519# 520 # By default normal clients are not limited because they don‘t receive data521 # without asking (in a push way), but just after a request, so only 522# asynchronous clients may create a scenario where data is requested faster 523# than it can read. 524# 525 # Instead there is a default limit for pubsub and slave clients, since 526 # subscribers and slaves receive data in a push fashion. 527# 528# Both the hard or the soft limit can be disabled just setting it to zero. 529 client-output-buffer-limit normal 000530 client-output-buffer-limit slave 256mb 64mb 60531 client-output-buffer-limit pubsub 32mb 8mb 60532533################################## INCLUDES ################################### 534535 # Include one or more other config files here. This is useful if you 536# have a standard template that goes to all Redis server but also need 537 # to customize a few per-server settings. Include files can include 538# other files, so use this wisely. 539# 540 # include /path/to/local.conf 541 # include /path/to/other.conf
原文:http://www.cnblogs.com/MDK-L/p/5698746.html
内容总结
以上是互联网集市为您收集整理的redis ubuntu 开机自启全部内容,希望文章能够帮你解决redis ubuntu 开机自启所遇到的程序开发问题。 如果觉得互联网集市技术教程内容还不错,欢迎将互联网集市网站推荐给程序员好友。
内容备注
版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 gblab@vip.qq.com 举报,一经查实,本站将立刻删除。
内容手机端
扫描二维码推送至手机访问。