zoukankan      html  css  js  c++  java
  • Docker从入门到动手实践

    一些理论知识,我这里就不累赘了

    docker 入门资料,参考:https://yeasy.gitbooks.io/docker_practice/content/

    Dockerfile常用命令,图片来源于网络

    Dockerfile 打包控制台应用程序

     新建一个控制台程序,控制台程序添加一个文本文件,去掉.txt 扩展名,改成Dockerfile 输入以下代码

    FROM microsoft/dotnet:sdk AS build
    WORKDIR /code
    COPY *.csproj /code
    RUN dotnet restore
    
    COPY . /code
    RUN dotnet publish -c Release -o out
    
    FROM microsoft/dotnet:runtime
    WORKDIR /app
    COPY --from=build /code/out /app
    ENTRYPOINT ["dotnet","console.dll"]

    Program.cs 中编写测试代码

     一切准备完成。就是build把项目打包成镜像了

    切换到当前项目路径下。输入:  docker build -t cn/console:v1 . 

    docker build -t :是打包固有的命令

    cn/console:v1 :

    cn:是组织名称或者说是用户名,如果你想把自己的镜像push到hub.docker 上,cn必须是你自己的用户名

    console:是镜像名称

    v1:是tag。一个标签,可以用来区分同一个镜像,不同用途。,如果不指定。默认是latest

    . :代表当前目录为上下文,dockerfile也必定是在当前目录下

    回车后,会看到一系列的执行步骤,dockerfile中。一条命令就是一个步骤

     通过 docker images 可以查看所有镜像

    通过docker images cn/console 查看相关镜像

    比如我本地有3个 cn/console镜像,但tag不同

    既然镜像有了。那么就可以根据镜像生成容器了。容器是镜像的一个实例。镜像运行起来才会有容器,就跟类和对象一样,new一个类,是实例化的操作

     输入命令:

    docker run --name myfirst cn/console:v1

    因为是占用前端线程运行容器,所有界面无法继续输入命令了。可以Ctrl+c 结束容器运行

     从上面的dockerfile。你会发现,我们是把源码打包成镜像的。也就是执行了restore,到Release操作

    其实如果你是已经Release后的文件了。dockerfile可以更简单

    FROM microsoft/dotnet
    WORKDIR /app
    COPY . /app
    CMD ["dotnet","run"]

     以上就是一个基础的程序打包成镜像,我觉得这不是重点,常用的应该是应用程序,而不是控制台程序

    后面打算把net core api打包成镜像。在讲这个之前,我们先来搭建好环境。

    Docker mysql

    因为我有个阿里云服务器(CentOS7),然后有2台笔记本,一个是Docker for Windows 环境,一个是CentOS7,所以经常会在这3个环境中来回折腾

    两种系统还是有区别的,至少我弄的时候,遇到过不少问题

    1:for Windows中默认拉起的镜像都在C盘。会导致C盘越来越大,建议迁移

    如果迁移的盘。比如我这个E盘。路径中已经存在MobyLinuxVM.vhdx 。是迁移不过的。要删除,但之前的镜像都没有了

    如果你想保存,先重命名MobyLinuxVM.vhdx,迁移后。删除之后的。之前的重命名回来即可

    2:共享盘。为了数据卷挂载用

     3:配置镜像加速(https://hlef8lmt.mirror.aliyuncs.com)

     然后可以去hub.docker上寻找需要的镜像,官方的mysql有2个镜像

     当然你通过命令也可以收索到:  docker search mysql 

     

     首先来看docker mysql

    准备需要挂载的目录和文件,上面我设置的共享盘是D盘,所以挂载的在D盘

    my.cnf配置文件,主要是设置mysql的参数

    [mysqld]
    user=mysql
    character-set-server=utf8
    [client]
    default-character-set=utf8
    [mysql]
    default-character-set=utf8

    data是空的。当run的时候,mysql会写入文件

    sql是需要在运行myslq后执行的初始化文件,比如我这里是给刚创建的用户名分配权限

    这里为了说明sql是执行成功的。我在加条。创建数据库的sql,创建数据库 docker和user表,并插入一条数据

    GRANT ALL PRIVILEGES ON *.* TO 'test'@'%' WITH GRANT OPTION;
    Create DATABASE docker;
    USE docker;
    CREATE TABLE user (ID int auto_increment primary key,name nvarchar(20),address nvarchar(50));
    insert into user(name,address)values('刘德华','香港');

    初始化后就执行的好处是。不用在run后,去手动执行,关于run后手动执行,

    可以查看我之前的docker安装mysql https://www.cnblogs.com/nsky/p/10413136.html

    全部配置完成后,开始敲命令,以下命令需要去掉注释

    docker run -d -p 3306:3306 
    --restart always #总是自动重启。比如系统重启,该容器会自动启动
    -e MYSQL_USER=test #创建用户名test
    -e MYSQL_PASSWORD=123456 #test密码
    -e MYSQL_PASSWORD_HOST=% #test 开启外部登陆
    -e MYSQL_ROOT_PASSWORD=123456  #root密码
    -e MYSQL_ROOT_HOST=% #root开启外部登陆
    -v /d/docker/mysql/my.cnf:/etc/my.cnf #配置文件
    -v /d/docker/mysql/sql:/docker-entrypoint-initdb.d #初始化的sql
    -v /d/docker/mysql/data:/var/lib/mysql  #data文件
    --name mysql #镜像名称
    mysql #基于那个镜像创建容器

    执行成功没有异常后。通过  docker ps  可以查看运行的容器,如果没有, 那就通过 docker ps -a 一定会有的

    现在可以通过Navicat连接试试

    创建了docker库。user表也有数据,能看到mysql库,说明test用户是有权限的

     当我使用mysql-server 镜像时,创建容器会无法启动

    可以看到。启动失败后。又继续重启,因为参数指定了restart always

     输入命令  docker logs mysql  查看启动日志

     最后在my.cnf中加这个,经测试,启动成功,就不一一放图了

     数据库准备好了,那么就快速的构建一个net core api 接口

    1:引入NugGet包,MySql.Data.EntityFrameworkCore

    2:创建DbContext

    using Docker.Api.Model;
    using Microsoft.EntityFrameworkCore;
    using System;
    using System.Collections.Generic;
    using System.Linq;
    using System.Threading.Tasks;
    
    namespace Docker.Api.Data
    {
        public class DbUserInfoContext : DbContext
        {
            public DbUserInfoContext(DbContextOptions<DbUserInfoContext> options) : base(options) { }
    
            public DbSet<UseInfo> userInfos { get; set; }
    
            /// <summary>
            /// 模型创建时触发
            /// </summary>
            /// <param name="modelBuilder"></param>
            protected override void OnModelCreating(ModelBuilder modelBuilder)
            {
                /*
                 修改表名和主键,user对应数据库的表,mysql默认是区分大小写的
                 查看:show variables like '%lower%';
                lower_case_table_names 为 0 区分,1 不区分
                 */
                modelBuilder.Entity<UseInfo>(b => b.ToTable("user").HasKey(u => u.id));
    
                //or
                //modelBuilder.Entity<user>()
                //    .ToTable("user")
                //    .HasKey(u => u.id);
    
                base.OnModelCreating(modelBuilder);
            }
        }
    }

    3:添加UserInfo控制器

     1 using System;
     2 using System.Collections.Generic;
     3 using System.Linq;
     4 using System.Threading.Tasks;
     5 using Docker.Api.Data;
     6 using Microsoft.AspNetCore.Http;
     7 using Microsoft.AspNetCore.Mvc;
     8 using Microsoft.EntityFrameworkCore;
     9 
    10 namespace Docker.Api.Controllers
    11 {
    12     [Route("api/[controller]")]
    13     [ApiController]
    14     public class UserInfoController : ControllerBase
    15     {
    16         private DbUserInfoContext _DbUserInfoContext;
    17         public UserInfoController(DbUserInfoContext context)
    18         {
    19             _DbUserInfoContext = context;
    20         }
    21         [HttpGet]
    22         public async Task<IActionResult> Get()
    23         {
    24             return new JsonResult(await _DbUserInfoContext.userInfos.FirstOrDefaultAsync());
    25         }
    26     }
    27 }

    4:配置sql连接字符串: server=localhost;port=3306;userid=test;password=123456;database=docker 

     

    run项目。访问能成功获取信息

    容器互连,Docker Network

    接下来我们把这个api也打包成镜像,然后基于该镜像创建容器,然后连接mysql镜像的容器。这称之为容器互连

    容器互连有3种方式

    1:Link方式。已经被docker淘汰,docker官方不推荐使用该方式

    2:Bridger,桥接的方式,单台机器用

    3:Overlay 适用于集群时候用

    Overlay 就我目前环境不适合测试,集群也不懂。就不搞了

    说说LInk和Bridger方式,具体理论知识请看docker官方文档。我这里只实践

    现在一切来回忆下

    刚上面打包控制台应用程序用的是:microsoft/dotnet 镜像

    然后后面带上tag

    比如:

    Micirosoft/dotnet:sdk

    包含了运行时和sdk命令,打包后会很大,因为包含sdk,一般用于测试环境

    Microsoft/dotnet:<version>-runtime

    包含运行时,不包含sdk,打包后就很小了,一般用于正式环境

    Microsoft/dotnet:<version>-runtime-deps

    打包的时候,会自包含runtime,也就是部署的机器有没有runtime是没有关系

    上面2种,必须机器要包含core环境

     

     

     

     

     



    修改程序port运行在80上

    编写api的Dockerfile

    我这里用的sdk,因为要用到sdk命令比如dotnet restore,dotnet publish

    如果已经publish的文件,直接用runtime会方便很多。上面也有提及

     1 #FROM mcr.microsoft.com/dotnet/core/sdk:2.2 AS build
     2 FROM microsoft/dotnet:2.2-sdk AS build
     3 WORKDIR /src
     4 WORKDIR /source
     5 #这里的后面的 . 就是/source 路径
     6 #或者 COPY *.csproj /source
     7 COPY *.csproj .
     8 RUN dotnet restore
     9 COPY . .
    10 # 发布到 /source/out11 RUN dotnet publish -c Release -o out
    12 
    13 #FROM mcr.microsoft.com/dotnet/core/runtime:2.2
    14 FROM microsoft/dotnet:2.2-aspnetcore-runtime
    15 WORKDIR /app
    16 COPY --from=build /source/out .
    17 EXPOSE 80
    18 ENTRYPOINT ["dotnet","Docker.Api.dll"]

    开始build项目 docker build -t cn/myapi . 

    可以看到。这里没有指定tag。所以默认是latest,size也不大

     成功后开始run一个容器,不过这之前要先:

    准备挂载目录。因为配置文件 appsettings 会需要动态配置,所以挂载出来

    还有,比如一个网站都有log日志,这些也需要挂载出来。便于管理。

    我这里就只挂载appsettings.json

     执行命令:

    docker run -d -p 80:80 --restart always --link mysql:mysqldb -v /d/docker/myapi/appsettings.json:/app/appsettings.json --name api cn/myapi

     分析:

    --restart always :总是重启

    -d:是在后台执行

    -p 80:80 :第一个80是暴露给外部的。第二个80是程序的。

    --link mysql:mysqldb : mysql是容器名称,mysqldb是自定义名称,可以理解为服务器

    -v /d/docker/myapi/appsettings.json:/app/appsettings.json:这里就是挂载外部的数据卷了

    也许你会问。我怎么知道这个路径的:/app/appsettings.json。从编写的dockerfile能分析出来,待会也可以进入容器看看

    最有的工作目录是 根路径下: /app

    然后通过页面访问试试

     发现依然无法访问,因为修改appsettings.json的连接方式

    记住这里是修改D:dockermyapiappsettings.json ,因为已经挂载出来

    把server改成mysqldb,然后重启容器: docker restart api 

    再次刷新页面

     我们 进入容器看看: docker exec -it api bash 可以看到根目录下存在app目录

     

    进入app目录

    个人认为link方式是最简单的。在这3种中,接下来看看Bridge方式

    1:首先创建一个网络 network,名称叫api2bridge

     docker network create -d bridge api2bridge 

    通过: docker network ls 可以查看到已经创建成功

    2:实例化容器

    为了区别于上面的80端口,这里新增一个8081

     docker run -d -p 8081:80 --restart always  -v /d/docker/myapi/appsettings.json:/app/appsettings.json --net api2bridge --name api2 cn/myapi

     创建容器的时候,自定network 这里的--net api2bridge 就是上面的bridge

    3:连接2个容器,通过: docker network connect api2bridge mysql  把api2和mysql连接起来

     4:修改appsettings.json  server=mysql 

     5 : restart 容器,如果是在创建容器前修改的配置文件。是不需要重启的,测试通过

     

     看看这两个容器是怎么连接的。通过命令:  docker inspect api2bridge 可以查看对象的元数据(容器或者网络)

     

     分别看看;

    docker inspect api2

     

     

    docker inspect mysql

     

     你会发现mysql有个"IPAddress":地址,

     上面我们在api2中的appsettings.json的server是直接些的容器名称:mysql。也可以直接些这个ip地址。比如: server=172.20.0.3 同样是可以的。

     Overlay方式就不讲了。因为我也不知道。哈哈

    docker-compose 容器编排

    通过这几个例子你会发现。2个容器要部署2个,如果项目依赖mysql,redis,MQ等等。那得部署多次,如此重复性的工作会影响效率

    所以有了docker-compose,compose

    参考:https://yeasy.gitbooks.io/docker_practice/content/compose/install.html

    安装:

    sudo curl -L https://github.com/docker/compose/releases/download/1.17.1/docker-compose-`uname -s`-`uname -m` > /usr/local/bin/docker-compose
    sudo chmod +x /usr/local/bin/docker-compose

    安装完成后,可以通过: docker-compose --version 查看版本

     通过: docker-compose --help 查看基本的命令

     不过我英文不好,就通过百度翻译了,翻译得有点生硬。仅供参考

    Commands:
      build              建立或重建服务
      bundle             从撰写文件生成Docker捆绑包
      config             验证并查看撰写文件
      create             创建服务
      down               停止并删除容器、网络、图像和卷
      events             从容器接收实时事件
      exec               在正在运行的容器中执行命令
      help               获取有关命令的帮助
      images             列表图像
      kill               杀死容器
      logs               查看容器的输出
      pause              暂停服务
      port               打印端口绑定的公共端口
      ps                 列表容器
      pull               拉取服务图像
      push               推送服务图像
      restart            重新启动服务
      rm                 移除停止的容器
      run                运行一次性命令
      scale              设置服务的容器数
      start              启动服务
      stop               停止服务
      top                显示正在运行的进程
      unpause            取消暂停服务
      up                 创建和启动容器
      version            显示Docker撰写版本信息

    目前为止已经有3个容器了,

    为了区别于之前的mysql和api和api2,这里命名要修改,编写在程序根目录下添加docker-compose.yml文件

    compose用的是yml语法。可以参考阮一峰些的文章

    http://www.ruanyifeng.com/blog/2016/07/yaml.html

    项目准备。依然在上面的api项目中添砖加瓦

    还记得上面初始化的创建docker库,user表吗。这里我们通过在代码中来实现,

    场景:创建myslq的时候,判断数据库是否有数据,否则新增一条数据

    技术栈:项目依赖mysql,redis,其实我工作中用的都是mssql,所以待会也会介绍

     1:init.sql 只保留一条sql语句

     2:新增UserInit类。用于初始化数据

    using Microsoft.AspNetCore.Builder;
    using Microsoft.EntityFrameworkCore;
    using Microsoft.Extensions.DependencyInjection;
    using Microsoft.Extensions.Logging;
    using System;
    using System.Collections.Generic;
    using System.Linq;
    using System.Threading.Tasks;
    
    namespace Docker.Api.Data
    {
        public class UserInit
        {
            private ILogger<UserInit> _logger;
    
            public UserInit(ILogger<UserInit> logger)
            {
                _logger = logger;
            }
    
            public static async Task InitData(IApplicationBuilder app, ILoggerFactory loggerFactory)
            {
    
                using (var scope = app.ApplicationServices.CreateScope())
                {
                    var context = scope.ServiceProvider.GetService<DbUserInfoContext>();
                    var logger = scope.ServiceProvider.GetService<ILogger<UserInit>>();
                    logger.LogDebug("begin mysql init");
                    context.Database.Migrate();
                    if (context.userInfos.Count() <= 0)
                    {
                        context.userInfos.Add(new Model.UseInfo
                        {
                            name = "admin",
                            address = "博客园"
                        });
                        context.SaveChanges();
                    }
                }
                await Task.CompletedTask;
            }
        }
    }

    程序启动调用:

    3:实体类

    using System;
    using System.Collections.Generic;
    using System.Linq;
    using System.Threading.Tasks;
    
    namespace Docker.Api.Model
    {
        public class UseInfo
        {
            public int id { get; set; }
            public string name { get; set; }
            public string address { get; set; }
        }
    }

    4:DbContext 上面也列出,这里就不展示了

    5:RedisHelper网络有。这里也不提了

    只准备2个接口。用于测试redis。一个读,一个写

    using System;
    using System.Collections.Generic;
    using System.Linq;
    using System.Threading.Tasks;
    using Microsoft.AspNetCore.Http;
    using Microsoft.AspNetCore.Mvc;
    
    namespace Docker.Api.Controllers
    {
        [Route("api/[controller]")]
        [ApiController]
        public class RedisController : ControllerBase
        {
            [HttpPost]
            public void Post()
            {
                RedisCommon.GetRedis().StringSet("docker", "hello", TimeSpan.FromMinutes(1));
            }
            [HttpGet]
            public string Get()
            {
                var docker = RedisCommon.GetRedis().GetStringKey("docker");
                if (docker.HasValue) return docker.ToString();
                return "empty";
            }
        }
    }

    6:根据Model生成Migration,这里简单过一下,具体参考我之前的:https://www.cnblogs.com/nsky/p/10323415.html

    调出程序包管理控制台

     输入: Add-Migration init 

     如果成功了就会这样:

     编写docker-compose.yml 文件,我这里的注释是便于理解。尽量不要写

    注:我是直接在项目中创建的文本文件,然后修改后缀名

    在网络上看到说。如果是在外部创建的记事本。要修改编码为:ASCII编码格式,我未测试

    version: '3'
    
    services:
      db:
        image: mysql
        container_name: 'mysql01'
        command: --character-set-server=utf8 --collation-server=utf8_general_ci
        restart: always
        ports:
          - '3307:3306'
        environment:
          MYSQL_USER: test
          MYSQL_PASSWORD: 123456
          MYSQL_PASSWORD_HOST: '%'
          MYSQL_ROOT_PASSWORD: 123456
          MYSQL_ROOT_HOST: '%'
        volumes:
          - /d/docker/mysql02/my.cnf:/etc/my.cnf
          - /d/docker/mysql02/data:/var/lib/mysql
          - /d/docker/mysql02/SqlInit:/docker-entrypoint-initdb.d
      redis:
        image: redis
        container_name: 'redis'
        command: redis-server /usr/local/etc/redis/redis.conf 
        restart: always
        ports:
          - '6379:6379'
        environment:
          requirepass: 123456 #redis密码
          appendonly: 'yes' #redis是否持久化
        volumes:
          - /d/docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf
          - /d/docker/redis/data:/data #这里会保存持久化数据
      web:
        build: . #会执行当前目录下面的dockerfile文件
        container_name: 'api3' #容器名称
        restart: always # web依赖于db,如果web比db启动快。就连接不上db导致web异常,web容器启动失败,restart可以不断重试,直到连接为止
        volumes:
          - /d/docker/myapi/appsettings.json:/app/appsettings.json
        ports:
          - '8082:80'
        depends_on: #依赖db容器,并不代表执行顺序
          - db
          - redis

    如果想看编写的yml文件是否正确,可以去在线的网站,验证是否正确,比如:http://nodeca.github.io/js-yaml/

     切换到当前目录输入:  docker-compose build 会开始build项目成镜像

     查看镜像:名字叫dockerapi_web

     输入命令: docker-compose up 会开始创建容器并启动

     

    输出的日志太多。这里只点几个有用的看

    EFcore插入Migration历史记录

     创建表:

     直到最后,程序阻塞。显示成功,因为这里没用用 -d 会阻塞,调试的时候不建议 -d

     然后新打开一个PowerShell,输入docker ps 查看运行的容器

     分别测试是否成功

     

     同样验证redis,用RedisDesktopManager连接

     

    从容器可以看出api3端口是8082,尝试访问下

     测试写redis,打开Postman写入Redis

     写人成功

    那么读取就不是什么大问题了

    问题汇总:

    如果你修改了代码,需要重新build。那么先删除容器: docker-compose down 会停止容器并删除

     docker-compose ps  查看容器列表

       docker-compose up -d   后端运行,不阻塞前端

      docker-compose restart  重启所有容器。

     

     自此所有容器成功运行,但我感觉还不够,因为一直都是在windos上玩。而没有上CentOS7,可我又不缺CentOS环境。所以要玩一把

    net core Api 跨平台部署

    技术栈:Jexus,mysql,mssql,redis

    关于jexus部署net core 可以参考我前面写的文章:https://www.cnblogs.com/nsky/p/10386460.html

    既然要加入新的成员。jexus 和 mssql,那么就得修改docker-compose文件

     在通过docker-compose统一打包前,我们先来单独玩玩mssql

    准备数据卷挂载目录

     data:保存数据库文件

    sql:执行的脚本。mssql没有mysql的docker-entrypoint-initdb.d 挂载,启动mysql就执行sql。这里sql文件夹

    虽然保存的是.sql文件。但要手动执行,不知道是不是我没有找到具体的方案

    sql里面放一个init.sql文件。编写sql脚本如下

     这里要注意一点,一条语句完成必须要带一个Go语句

    参考官方文档:

    https://docs.microsoft.com/zh-cn/sql/linux/quickstart-install-connect-docker?view=sql-server-2017&pivots=cs1-bash

    https://docs.microsoft.com/zh-cn/sql/linux/tutorial-restore-backup-in-sql-server-container?view=sql-server-2017

    镜像文档:https://hub.docker.com/_/microsoft-mssql-server

    //注释部分
    docker run -d -p 1433:1433 
    -e ACCEPT_EULA=Y  #确认您接受最终用户许可协议。
    -e SA_PASSWORD=DockerPwd123  #强大的系统管理员(SA)密码:至少8个字符,包括大写,小写字母,基数为10的数字和/或非字母数字符号。
    -e MSSQL_PID=Express  #版本(Developer,Express,Enterprise,EnterpriseCore)默认值:Developer  
    -v /docker/mssql:/var/opt/mssql   # 映射数据库
    v /d/docker/mssql/sql:/script #把需要执行的脚本放这里,script路径随便改,不是初始化执行,是手到执行
    --name mssql #容器名称
    mcr.microsoft.com/mssql/server #镜像

    执行成功后。数据卷挂载目录。生成了文件

    此时data也有默认的数据库了

     通过 MSSMS(  Microsoft SQL Server Management Studio )连接试试

     刚上面说了sql中文件是没有被执行的。必须手动执行。

    手动执行前,先来看看其他一些相关命令

    进入容器后: docker exec -it mssql bash 

    登陆数据库:localhost也可以用指定的ip代替,如果有端口。则带端口号即可

    /opt/mssql-tools/bin/sqlcmd -S localhost -U SA -P DockerPwd123
    
    -S 是服务器,不管端口是多少,都不用写
    -U 是用户名 
    -P 是密码

    如果出现 1> 说明的登陆成功了

    可以输入语句:select getdate() 试试,回车后,需要加Go语句,不过日期怎么不对?好像是相差8个时区

    修改时区,可以通过TZ变量

    docker run -e TZ="Asia/Shanghai" -e 'ACCEPT_EULA=Y' -e 'MSSQL_SA_PASSWORD=DockerPwd123' -e MSSQL_PID=Express -p 1433:1433 --name mssql -d mcr.microsoft.com/mssql/server

    官网文档:https://docs.microsoft.com/zh-cn/sql/linux/sql-server-linux-configure-docker?view=sql-server-2017#tz

     执行sql中的文件

    登陆容器后执行操作: /opt/mssql-tools/bin/sqlcmd -S localhost -U SA -P DockerPwd123 -i /script/init.sql 

     

    挂载目录也有,这样就算容器无法进入。数据库也存在

    由于时间问题,docker-compose 就不加入mssql,只加jexus,修改docker-compose如下

    直接上docker-compose文件

    version: '3'
    
    services:
      db:
        image: mysql
        container_name: 'mysql'
        command: --character-set-server=utf8 --collation-server=utf8_general_ci
        restart: always
        ports:
          - '3306:3306'
        environment:
          MYSQL_USER: test
          MYSQL_PASSWORD: 123456
          MYSQL_PASSWORD_HOST: '%'
          MYSQL_ROOT_PASSWORD: 123456
          MYSQL_ROOT_HOST: '%'
        volumes:
          - /docker/mysq/my.cnf:/etc/my.cnf
          - /docker/mysql/data:/var/lib/mysql
          - /docker/mysql/sql:/docker-entrypoint-initdb.d
      redis:
        image: redis
        container_name: 'redis'
        command: redis-server /usr/local/etc/redis/redis.conf 
        restart: always
        ports:
          - '6379:6379'
        environment:
          requirepass: 123456 #redis密码
          appendonly: 'yes' #redis是否持久化
        volumes:
          - /docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf
          - /docker/redis/data:/data #这里会保存持久化数据
      jexus:
        image: byniqing/jexus
        container_name: 'jexus'
        ports:
          - '80:8803'#前一个80是暴露外部的,后一个8803是jexus监听的的。也就是配置文件中port=8803,此端口必须要开通
        restart: always
        volumes:
          - /docker/jexus/www:/var/www
          - /docker/jexus/siteconf:/usr/jexus/siteconf
          - /docker/jexus/log:/usr/jexus/log
        #depends_on:
          #- web
      web:
        build: . #会执行当前目录下面的dockerfile文件
        container_name: 'api' #容器名称
        restart: always # web依赖于db,如果web比db启动快。就连接不上db导致web异常,web容器启动失败,restart可以不断重试,直到连接为止
        volumes:
          - /docker/myapi/appsettings.json:/app/appsettings.json
        ports:
          - '8802:80'#8802端口必须开启,如果是阿里云。添加入栈规则,jexus配置。代理到8802即可,浏览器直接访问这个port也能访问
        depends_on: #依赖db容器,并不代表执行顺序
          - db
          - redis

    分析:

    1:jexus我用了自己打包后的镜像。然后push到hub.docker上去了,大家也可以打包一个自己的。方便以后测试,升级,当然直接用给我这个也可以,dockerfile 如下:

    FROM debian:latest
    MAINTAINER xxx <xxx@xxx.com>
    
    RUN apt-get update 
            && apt-get -y install wget 
            && cd /usr 
            && wget https://www.linuxdot.net/down/jexus-5.8.3-x64.tar.gz 
            && tar -zxvf jexus-5.8.3-x64.tar.gz 
            && apt-get -y autoremove --purge wget 
            && rm -rf /var/lib/apt/lists/* jexus-5.8.3-x64.tar.gz
    
    EXPOSE 80
    WORKDIR /usr/jexus
    CMD /usr/jexus/jwss

    2:web中的ports暴露了8802端口,需要开通该端口

    3:jexus文件配置:reproxy=/  ip:8802(绕过容器内部,直接访问容器,所以直接在浏览器也是直接可以访问的),或者reproxy=/ web:80 (容器互联)

     

    然后你会发现通过ip:80和ip:8802都能访问,

    一个是通过jexus访问。一个是绕过了jexus直接访问了api接口,那怎么行呢,那么jexus就没有存在的用意了

    所以:我们应该不暴露web端口,即屏蔽掉ports

     然后重新修改jexus,通过容器名称访问: reproxy=/ web:80 

    因为compose打包后后是在同一个网络中:

    比如: docker-compose ps 可以查看当前服务下的所有容器

    那怎么判断当前容器使用的那些网络呢?

    可以查看当前某个容器的元数据:比如我们来看mysql的: docker inspect mysql 也可以看出Networks节点信息

     通过: docker network ls  查看当前网络,存在 dockerapi_default

    那么就可以通过: docker inspect dockerapi_default 查看下dockerapi_default网络的元数据

    可以看到Containers节点下。docker-compose中定义的4个容器,并且ip4都在192.168.48.x/20

    不知道你跟是否有同一个疑问,虽然api没有暴露接口,但api和jexus。tcp都是80

    那我直接访问ip地址怎么确定一定就是访问的jexus。而不是Api呢。那么我们来改造一下,jexus暴露8802,api内部依然是80

     从新打包,查看容器:

    然后浏览器输入:ip:8802/api/values 访问成功

    输入:ip/aip/values 访问失败

    那么这样就达到了只能通过jexus访问到我的aip了

    修改挂载文件

    使用数据卷,文件挂载到宿主机就是为了方便修改,这里拿redis为例

    我们在创建redis的时候有个挂载目录为: /docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf 

    redis.conf这个就是redis的配置文件。可以自行修改, 

    去网上下载一个对应版本的配置文件放进去即可:http://download.redis.io/releases/

    比如上面的默认密码是:123456,我们来改成7890

    1:注释掉,bind 127.0.0.1

     2:修改,requirepass  值,是不是发现requirepass 跟docker-compose 中变量是同一个

     3:重启redis

    4:测试连接成功,

    进入redis容器。查看 /usr/local/etc/redis/redis.conf  

    你会发现,redis.conf是同步更新的,这里就不截图了

    放一个redis.confi文件

       1 # Redis configuration file example
       2 
       3 # Note on units: when memory size is needed, it is possible to specify
       4 # it in the usual form of 1k 5GB 4M and so forth:
       5 #
       6 # 1k => 1000 bytes
       7 # 1kb => 1024 bytes
       8 # 1m => 1000000 bytes
       9 # 1mb => 1024*1024 bytes
      10 # 1g => 1000000000 bytes
      11 # 1gb => 1024*1024*1024 bytes
      12 #
      13 # units are case insensitive so 1GB 1Gb 1gB are all the same.
      14 
      15 ################################## INCLUDES ###################################
      16 
      17 # Include one or more other config files here.  This is useful if you
      18 # have a standard template that goes to all Redis servers but also need
      19 # to customize a few per-server settings.  Include files can include
      20 # other files, so use this wisely.
      21 #
      22 # Notice option "include" won't be rewritten by command "CONFIG REWRITE"
      23 # from admin or Redis Sentinel. Since Redis always uses the last processed
      24 # line as value of a configuration directive, you'd better put includes
      25 # at the beginning of this file to avoid overwriting config change at runtime.
      26 #
      27 # If instead you are interested in using includes to override configuration
      28 # options, it is better to use include as the last line.
      29 #
      30 # include .path	olocal.conf
      31 # include c:path	oother.conf
      32 
      33 ################################## NETWORK #####################################
      34 
      35 # By default, if no "bind" configuration directive is specified, Redis listens
      36 # for connections from all the network interfaces available on the server.
      37 # It is possible to listen to just one or multiple selected interfaces using
      38 # the "bind" configuration directive, followed by one or more IP addresses.
      39 #
      40 # Examples:
      41 #
      42 # bind 192.168.1.100 10.0.0.1
      43 # bind 127.0.0.1 ::1
      44 #
      45 # ~~~ WARNING ~~~ If the computer running Redis is directly exposed to the
      46 # internet, binding to all the interfaces is dangerous and will expose the
      47 # instance to everybody on the internet. So by default we uncomment the
      48 # following bind directive, that will force Redis to listen only into
      49 # the IPv4 lookback interface address (this means Redis will be able to
      50 # accept connections only from clients running into the same computer it
      51 # is running).
      52 #
      53 # IF YOU ARE SURE YOU WANT YOUR INSTANCE TO LISTEN TO ALL THE INTERFACES
      54 # JUST COMMENT THE FOLLOWING LINE.
      55 # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
      56 #bind 127.0.0.1
      57 
      58 # Protected mode is a layer of security protection, in order to avoid that
      59 # Redis instances left open on the internet are accessed and exploited.
      60 #
      61 # When protected mode is on and if:
      62 #
      63 # 1) The server is not binding explicitly to a set of addresses using the
      64 #    "bind" directive.
      65 # 2) No password is configured.
      66 #
      67 # The server only accepts connections from clients connecting from the
      68 # IPv4 and IPv6 loopback addresses 127.0.0.1 and ::1, and from Unix domain
      69 # sockets.
      70 #
      71 # By default protected mode is enabled. You should disable it only if
      72 # you are sure you want clients from other hosts to connect to Redis
      73 # even if no authentication is configured, nor a specific set of interfaces
      74 # are explicitly listed using the "bind" directive.
      75 protected-mode yes
      76 
      77 # Accept connections on the specified port, default is 6379 (IANA #815344).
      78 # If port 0 is specified Redis will not listen on a TCP socket.
      79 port 6379
      80 
      81 # TCP listen() backlog.
      82 #
      83 # In high requests-per-second environments you need an high backlog in order
      84 # to avoid slow clients connections issues. Note that the Linux kernel
      85 # will silently truncate it to the value of /proc/sys/net/core/somaxconn so
      86 # make sure to raise both the value of somaxconn and tcp_max_syn_backlog
      87 # in order to get the desired effect.
      88 tcp-backlog 511
      89 
      90 # Unix socket.
      91 #
      92 # Specify the path for the Unix socket that will be used to listen for
      93 # incoming connections. There is no default, so Redis will not listen
      94 # on a unix socket when not specified.
      95 #
      96 # unixsocket /tmp/redis.sock
      97 # unixsocketperm 700
      98 
      99 # Close the connection after a client is idle for N seconds (0 to disable)
     100 timeout 0
     101 
     102 # TCP keepalive.
     103 #
     104 # If non-zero, use SO_KEEPALIVE to send TCP ACKs to clients in absence
     105 # of communication. This is useful for two reasons:
     106 #
     107 # 1) Detect dead peers.
     108 # 2) Take the connection alive from the point of view of network
     109 #    equipment in the middle.
     110 #
     111 # On Linux, the specified value (in seconds) is the period used to send ACKs.
     112 # Note that to close the connection the double of the time is needed.
     113 # On other kernels the period depends on the kernel configuration.
     114 #
     115 # A reasonable value for this option is 60 seconds.
     116 tcp-keepalive 0
     117 
     118 ################################# GENERAL #####################################
     119 
     120 # By default Redis does not run as a daemon. Use 'yes' if you need it.
     121 # Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
     122 # NOT SUPPORTED ON WINDOWS daemonize no
     123 
     124 # If you run Redis from upstart or systemd, Redis can interact with your
     125 # supervision tree. Options:
     126 #   supervised no      - no supervision interaction
     127 #   supervised upstart - signal upstart by putting Redis into SIGSTOP mode
     128 #   supervised systemd - signal systemd by writing READY=1 to $NOTIFY_SOCKET
     129 #   supervised auto    - detect upstart or systemd method based on
     130 #                        UPSTART_JOB or NOTIFY_SOCKET environment variables
     131 # Note: these supervision methods only signal "process is ready."
     132 #       They do not enable continuous liveness pings back to your supervisor.
     133 # NOT SUPPORTED ON WINDOWS supervised no
     134 
     135 # If a pid file is specified, Redis writes it where specified at startup
     136 # and removes it at exit.
     137 #
     138 # When the server runs non daemonized, no pid file is created if none is
     139 # specified in the configuration. When the server is daemonized, the pid file
     140 # is used even if not specified, defaulting to "/var/run/redis.pid".
     141 #
     142 # Creating a pid file is best effort: if Redis is not able to create it
     143 # nothing bad happens, the server will start and run normally.
     144 # NOT SUPPORTED ON WINDOWS pidfile /var/run/redis.pid
     145 
     146 # Specify the server verbosity level.
     147 # This can be one of:
     148 # debug (a lot of information, useful for development/testing)
     149 # verbose (many rarely useful info, but not a mess like the debug level)
     150 # notice (moderately verbose, what you want in production probably)
     151 # warning (only very important / critical messages are logged)
     152 loglevel notice
     153 
     154 # Specify the log file name. Also 'stdout' can be used to force
     155 # Redis to log on the standard output.
     156 logfile ""
     157 
     158 # To enable logging to the Windows EventLog, just set 'syslog-enabled' to
     159 # yes, and optionally update the other syslog parameters to suit your needs.
     160 # If Redis is installed and launched as a Windows Service, this will
     161 # automatically be enabled.
     162 # syslog-enabled no
     163 
     164 # Specify the source name of the events in the Windows Application log.
     165 # syslog-ident redis
     166 
     167 # Set the number of databases. The default database is DB 0, you can select
     168 # a different one on a per-connection basis using SELECT <dbid> where
     169 # dbid is a number between 0 and 'databases'-1
     170 databases 16
     171 
     172 ################################ SNAPSHOTTING  ################################
     173 #
     174 # Save the DB on disk:
     175 #
     176 #   save <seconds> <changes>
     177 #
     178 #   Will save the DB if both the given number of seconds and the given
     179 #   number of write operations against the DB occurred.
     180 #
     181 #   In the example below the behaviour will be to save:
     182 #   after 900 sec (15 min) if at least 1 key changed
     183 #   after 300 sec (5 min) if at least 10 keys changed
     184 #   after 60 sec if at least 10000 keys changed
     185 #
     186 #   Note: you can disable saving completely by commenting out all "save" lines.
     187 #
     188 #   It is also possible to remove all the previously configured save
     189 #   points by adding a save directive with a single empty string argument
     190 #   like in the following example:
     191 #
     192 #   save ""
     193 
     194 save 900 1
     195 save 300 10
     196 save 60 10000
     197 
     198 # By default Redis will stop accepting writes if RDB snapshots are enabled
     199 # (at least one save point) and the latest background save failed.
     200 # This will make the user aware (in a hard way) that data is not persisting
     201 # on disk properly, otherwise chances are that no one will notice and some
     202 # disaster will happen.
     203 #
     204 # If the background saving process will start working again Redis will
     205 # automatically allow writes again.
     206 #
     207 # However if you have setup your proper monitoring of the Redis server
     208 # and persistence, you may want to disable this feature so that Redis will
     209 # continue to work as usual even if there are problems with disk,
     210 # permissions, and so forth.
     211 stop-writes-on-bgsave-error yes
     212 
     213 # Compress string objects using LZF when dump .rdb databases?
     214 # For default that's set to 'yes' as it's almost always a win.
     215 # If you want to save some CPU in the saving child set it to 'no' but
     216 # the dataset will likely be bigger if you have compressible values or keys.
     217 rdbcompression yes
     218 
     219 # Since version 5 of RDB a CRC64 checksum is placed at the end of the file.
     220 # This makes the format more resistant to corruption but there is a performance
     221 # hit to pay (around 10%) when saving and loading RDB files, so you can disable it
     222 # for maximum performances.
     223 #
     224 # RDB files created with checksum disabled have a checksum of zero that will
     225 # tell the loading code to skip the check.
     226 rdbchecksum yes
     227 
     228 # The filename where to dump the DB
     229 dbfilename dump.rdb
     230 
     231 # The working directory.
     232 #
     233 # The DB will be written inside this directory, with the filename specified
     234 # above using the 'dbfilename' configuration directive.
     235 #
     236 # The Append Only File will also be created inside this directory.
     237 #
     238 # Note that you must specify a directory here, not a file name.
     239 dir ./
     240 
     241 ################################# REPLICATION #################################
     242 
     243 # Master-Slave replication. Use slaveof to make a Redis instance a copy of
     244 # another Redis server. A few things to understand ASAP about Redis replication.
     245 #
     246 # 1) Redis replication is asynchronous, but you can configure a master to
     247 #    stop accepting writes if it appears to be not connected with at least
     248 #    a given number of slaves.
     249 # 2) Redis slaves are able to perform a partial resynchronization with the
     250 #    master if the replication link is lost for a relatively small amount of
     251 #    time. You may want to configure the replication backlog size (see the next
     252 #    sections of this file) with a sensible value depending on your needs.
     253 # 3) Replication is automatic and does not need user intervention. After a
     254 #    network partition slaves automatically try to reconnect to masters
     255 #    and resynchronize with them.
     256 #
     257 # slaveof <masterip> <masterport>
     258 
     259 # If the master is password protected (using the "requirepass" configuration
     260 # directive below) it is possible to tell the slave to authenticate before
     261 # starting the replication synchronization process, otherwise the master will
     262 # refuse the slave request.
     263 #
     264 # masterauth <master-password>
     265 
     266 # When a slave loses its connection with the master, or when the replication
     267 # is still in progress, the slave can act in two different ways:
     268 #
     269 # 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
     270 #    still reply to client requests, possibly with out of date data, or the
     271 #    data set may just be empty if this is the first synchronization.
     272 #
     273 # 2) if slave-serve-stale-data is set to 'no' the slave will reply with
     274 #    an error "SYNC with master in progress" to all the kind of commands
     275 #    but to INFO and SLAVEOF.
     276 #
     277 slave-serve-stale-data yes
     278 
     279 # You can configure a slave instance to accept writes or not. Writing against
     280 # a slave instance may be useful to store some ephemeral data (because data
     281 # written on a slave will be easily deleted after resync with the master) but
     282 # may also cause problems if clients are writing to it because of a
     283 # misconfiguration.
     284 #
     285 # Since Redis 2.6 by default slaves are read-only.
     286 #
     287 # Note: read only slaves are not designed to be exposed to untrusted clients
     288 # on the internet. It's just a protection layer against misuse of the instance.
     289 # Still a read only slave exports by default all the administrative commands
     290 # such as CONFIG, DEBUG, and so forth. To a limited extent you can improve
     291 # security of read only slaves using 'rename-command' to shadow all the
     292 # administrative / dangerous commands.
     293 slave-read-only yes
     294 
     295 # Replication SYNC strategy: disk or socket.
     296 #
     297 # -------------------------------------------------------
     298 # WARNING: DISKLESS REPLICATION IS EXPERIMENTAL CURRENTLY
     299 # -------------------------------------------------------
     300 #
     301 # New slaves and reconnecting slaves that are not able to continue the replication
     302 # process just receiving differences, need to do what is called a "full
     303 # synchronization". An RDB file is transmitted from the master to the slaves.
     304 # The transmission can happen in two different ways:
     305 #
     306 # 1) Disk-backed: The Redis master creates a new process that writes the RDB
     307 #                 file on disk. Later the file is transferred by the parent
     308 #                 process to the slaves incrementally.
     309 # 2) Diskless: The Redis master creates a new process that directly writes the
     310 #              RDB file to slave sockets, without touching the disk at all.
     311 #
     312 # With disk-backed replication, while the RDB file is generated, more slaves
     313 # can be queued and served with the RDB file as soon as the current child producing
     314 # the RDB file finishes its work. With diskless replication instead once
     315 # the transfer starts, new slaves arriving will be queued and a new transfer
     316 # will start when the current one terminates.
     317 #
     318 # When diskless replication is used, the master waits a configurable amount of
     319 # time (in seconds) before starting the transfer in the hope that multiple slaves
     320 # will arrive and the transfer can be parallelized.
     321 #
     322 # With slow disks and fast (large bandwidth) networks, diskless replication
     323 # works better.
     324 repl-diskless-sync no
     325 
     326 # When diskless replication is enabled, it is possible to configure the delay
     327 # the server waits in order to spawn the child that transfers the RDB via socket
     328 # to the slaves.
     329 #
     330 # This is important since once the transfer starts, it is not possible to serve
     331 # new slaves arriving, that will be queued for the next RDB transfer, so the server
     332 # waits a delay in order to let more slaves arrive.
     333 #
     334 # The delay is specified in seconds, and by default is 5 seconds. To disable
     335 # it entirely just set it to 0 seconds and the transfer will start ASAP.
     336 repl-diskless-sync-delay 5
     337 
     338 # Slaves send PINGs to server in a predefined interval. It's possible to change
     339 # this interval with the repl_ping_slave_period option. The default value is 10
     340 # seconds.
     341 #
     342 # repl-ping-slave-period 10
     343 
     344 # The following option sets the replication timeout for:
     345 #
     346 # 1) Bulk transfer I/O during SYNC, from the point of view of slave.
     347 # 2) Master timeout from the point of view of slaves (data, pings).
     348 # 3) Slave timeout from the point of view of masters (REPLCONF ACK pings).
     349 #
     350 # It is important to make sure that this value is greater than the value
     351 # specified for repl-ping-slave-period otherwise a timeout will be detected
     352 # every time there is low traffic between the master and the slave.
     353 #
     354 # repl-timeout 60
     355 
     356 # Disable TCP_NODELAY on the slave socket after SYNC?
     357 #
     358 # If you select "yes" Redis will use a smaller number of TCP packets and
     359 # less bandwidth to send data to slaves. But this can add a delay for
     360 # the data to appear on the slave side, up to 40 milliseconds with
     361 # Linux kernels using a default configuration.
     362 #
     363 # If you select "no" the delay for data to appear on the slave side will
     364 # be reduced but more bandwidth will be used for replication.
     365 #
     366 # By default we optimize for low latency, but in very high traffic conditions
     367 # or when the master and slaves are many hops away, turning this to "yes" may
     368 # be a good idea.
     369 repl-disable-tcp-nodelay no
     370 
     371 # Set the replication backlog size. The backlog is a buffer that accumulates
     372 # slave data when slaves are disconnected for some time, so that when a slave
     373 # wants to reconnect again, often a full resync is not needed, but a partial
     374 # resync is enough, just passing the portion of data the slave missed while
     375 # disconnected.
     376 #
     377 # The bigger the replication backlog, the longer the time the slave can be
     378 # disconnected and later be able to perform a partial resynchronization.
     379 #
     380 # The backlog is only allocated once there is at least a slave connected.
     381 #
     382 # repl-backlog-size 1mb
     383 
     384 # After a master has no longer connected slaves for some time, the backlog
     385 # will be freed. The following option configures the amount of seconds that
     386 # need to elapse, starting from the time the last slave disconnected, for
     387 # the backlog buffer to be freed.
     388 #
     389 # A value of 0 means to never release the backlog.
     390 #
     391 # repl-backlog-ttl 3600
     392 
     393 # The slave priority is an integer number published by Redis in the INFO output.
     394 # It is used by Redis Sentinel in order to select a slave to promote into a
     395 # master if the master is no longer working correctly.
     396 #
     397 # A slave with a low priority number is considered better for promotion, so
     398 # for instance if there are three slaves with priority 10, 100, 25 Sentinel will
     399 # pick the one with priority 10, that is the lowest.
     400 #
     401 # However a special priority of 0 marks the slave as not able to perform the
     402 # role of master, so a slave with priority of 0 will never be selected by
     403 # Redis Sentinel for promotion.
     404 #
     405 # By default the priority is 100.
     406 slave-priority 100
     407 
     408 # It is possible for a master to stop accepting writes if there are less than
     409 # N slaves connected, having a lag less or equal than M seconds.
     410 #
     411 # The N slaves need to be in "online" state.
     412 #
     413 # The lag in seconds, that must be <= the specified value, is calculated from
     414 # the last ping received from the slave, that is usually sent every second.
     415 #
     416 # This option does not GUARANTEE that N replicas will accept the write, but
     417 # will limit the window of exposure for lost writes in case not enough slaves
     418 # are available, to the specified number of seconds.
     419 #
     420 # For example to require at least 3 slaves with a lag <= 10 seconds use:
     421 #
     422 # min-slaves-to-write 3
     423 # min-slaves-max-lag 10
     424 #
     425 # Setting one or the other to 0 disables the feature.
     426 #
     427 # By default min-slaves-to-write is set to 0 (feature disabled) and
     428 # min-slaves-max-lag is set to 10.
     429 
     430 ################################## SECURITY ###################################
     431 
     432 # Require clients to issue AUTH <PASSWORD> before processing any other
     433 # commands.  This might be useful in environments in which you do not trust
     434 # others with access to the host running redis-server.
     435 #
     436 # This should stay commented out for backward compatibility and because most
     437 # people do not need auth (e.g. they run their own servers).
     438 #
     439 # Warning: since Redis is pretty fast an outside user can try up to
     440 # 150k passwords per second against a good box. This means that you should
     441 # use a very strong password otherwise it will be very easy to break.
     442 #
     443 # requirepass foobared
     444 requirepass 7890
     445 # Command renaming.
     446 #
     447 # It is possible to change the name of dangerous commands in a shared
     448 # environment. For instance the CONFIG command may be renamed into something
     449 # hard to guess so that it will still be available for internal-use tools
     450 # but not available for general clients.
     451 #
     452 # Example:
     453 #
     454 # rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52
     455 #
     456 # It is also possible to completely kill a command by renaming it into
     457 # an empty string:
     458 #
     459 # rename-command CONFIG ""
     460 #
     461 # Please note that changing the name of commands that are logged into the
     462 # AOF file or transmitted to slaves may cause problems.
     463 
     464 ################################### LIMITS ####################################
     465 
     466 # Set the max number of connected clients at the same time. By default
     467 # this limit is set to 10000 clients, however if the Redis server is not
     468 # able to configure the process file limit to allow for the specified limit
     469 # the max number of allowed clients is set to the current file limit
     470 # minus 32 (as Redis reserves a few file descriptors for internal uses).
     471 #
     472 # Once the limit is reached Redis will close all the new connections sending
     473 # an error 'max number of clients reached'.
     474 #
     475 # maxclients 10000
     476 
     477 # If Redis is to be used as an in-memory-only cache without any kind of
     478 # persistence, then the fork() mechanism used by the background AOF/RDB
     479 # persistence is unnecessary. As an optimization, all persistence can be
     480 # turned off in the Windows version of Redis. This will redirect heap
     481 # allocations to the system heap allocator, and disable commands that would
     482 # otherwise cause fork() operations: BGSAVE and BGREWRITEAOF.
     483 # This flag may not be combined with any of the other flags that configure
     484 # AOF and RDB operations.
     485 # persistence-available [(yes)|no]
     486 
     487 # Don't use more memory than the specified amount of bytes.
     488 # When the memory limit is reached Redis will try to remove keys
     489 # according to the eviction policy selected (see maxmemory-policy).
     490 #
     491 # If Redis can't remove keys according to the policy, or if the policy is
     492 # set to 'noeviction', Redis will start to reply with errors to commands
     493 # that would use more memory, like SET, LPUSH, and so on, and will continue
     494 # to reply to read-only commands like GET.
     495 #
     496 # This option is usually useful when using Redis as an LRU cache, or to set
     497 # a hard memory limit for an instance (using the 'noeviction' policy).
     498 #
     499 # WARNING: If you have slaves attached to an instance with maxmemory on,
     500 # the size of the output buffers needed to feed the slaves are subtracted
     501 # from the used memory count, so that network problems / resyncs will
     502 # not trigger a loop where keys are evicted, and in turn the output
     503 # buffer of slaves is full with DELs of keys evicted triggering the deletion
     504 # of more keys, and so forth until the database is completely emptied.
     505 #
     506 # In short... if you have slaves attached it is suggested that you set a lower
     507 # limit for maxmemory so that there is some free RAM on the system for slave
     508 # output buffers (but this is not needed if the policy is 'noeviction').
     509 #
     510 # WARNING: not setting maxmemory will cause Redis to terminate with an
     511 # out-of-memory exception if the heap limit is reached.
     512 #
     513 # NOTE: since Redis uses the system paging file to allocate the heap memory,
     514 # the Working Set memory usage showed by the Windows Task Manager or by other
     515 # tools such as ProcessExplorer will not always be accurate. For example, right
     516 # after a background save of the RDB or the AOF files, the working set value
     517 # may drop significantly. In order to check the correct amount of memory used
     518 # by the redis-server to store the data, use the INFO client command. The INFO
     519 # command shows only the memory used to store the redis data, not the extra
     520 # memory used by the Windows process for its own requirements. Th3 extra amount
     521 # of memory not reported by the INFO command can be calculated subtracting the
     522 # Peak Working Set reported by the Windows Task Manager and the used_memory_peak
     523 # reported by the INFO command.
     524 #
     525 # maxmemory <bytes>
     526 
     527 # MAXMEMORY POLICY: how Redis will select what to remove when maxmemory
     528 # is reached. You can select among five behaviors:
     529 #
     530 # volatile-lru -> remove the key with an expire set using an LRU algorithm
     531 # allkeys-lru -> remove any key according to the LRU algorithm
     532 # volatile-random -> remove a random key with an expire set
     533 # allkeys-random -> remove a random key, any key
     534 # volatile-ttl -> remove the key with the nearest expire time (minor TTL)
     535 # noeviction -> don't expire at all, just return an error on write operations
     536 #
     537 # Note: with any of the above policies, Redis will return an error on write
     538 #       operations, when there are no suitable keys for eviction.
     539 #
     540 #       At the date of writing these commands are: set setnx setex append
     541 #       incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd
     542 #       sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby
     543 #       zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby
     544 #       getset mset msetnx exec sort
     545 #
     546 # The default is:
     547 #
     548 # maxmemory-policy noeviction
     549 
     550 # LRU and minimal TTL algorithms are not precise algorithms but approximated
     551 # algorithms (in order to save memory), so you can tune it for speed or
     552 # accuracy. For default Redis will check five keys and pick the one that was
     553 # used less recently, you can change the sample size using the following
     554 # configuration directive.
     555 #
     556 # The default of 5 produces good enough results. 10 Approximates very closely
     557 # true LRU but costs a bit more CPU. 3 is very fast but not very accurate.
     558 #
     559 # maxmemory-samples 5
     560 
     561 ############################## APPEND ONLY MODE ###############################
     562 
     563 # By default Redis asynchronously dumps the dataset on disk. This mode is
     564 # good enough in many applications, but an issue with the Redis process or
     565 # a power outage may result into a few minutes of writes lost (depending on
     566 # the configured save points).
     567 #
     568 # The Append Only File is an alternative persistence mode that provides
     569 # much better durability. For instance using the default data fsync policy
     570 # (see later in the config file) Redis can lose just one second of writes in a
     571 # dramatic event like a server power outage, or a single write if something
     572 # wrong with the Redis process itself happens, but the operating system is
     573 # still running correctly.
     574 #
     575 # AOF and RDB persistence can be enabled at the same time without problems.
     576 # If the AOF is enabled on startup Redis will load the AOF, that is the file
     577 # with the better durability guarantees.
     578 #
     579 # Please check http://redis.io/topics/persistence for more information.
     580 
     581 appendonly no
     582 
     583 # The name of the append only file (default: "appendonly.aof")
     584 appendfilename "appendonly.aof"
     585 
     586 # The fsync() call tells the Operating System to actually write data on disk
     587 # instead of waiting for more data in the output buffer. Some OS will really flush
     588 # data on disk, some other OS will just try to do it ASAP.
     589 #
     590 # Redis supports three different modes:
     591 #
     592 # no: don't fsync, just let the OS flush the data when it wants. Faster.
     593 # always: fsync after every write to the append only log. Slow, Safest.
     594 # everysec: fsync only one time every second. Compromise.
     595 #
     596 # The default is "everysec", as that's usually the right compromise between
     597 # speed and data safety. It's up to you to understand if you can relax this to
     598 # "no" that will let the operating system flush the output buffer when
     599 # it wants, for better performances (but if you can live with the idea of
     600 # some data loss consider the default persistence mode that's snapshotting),
     601 # or on the contrary, use "always" that's very slow but a bit safer than
     602 # everysec.
     603 #
     604 # More details please check the following article:
     605 # http://antirez.com/post/redis-persistence-demystified.html
     606 #
     607 # If unsure, use "everysec".
     608 
     609 # appendfsync always
     610 appendfsync everysec
     611 # appendfsync no
     612 
     613 # When the AOF fsync policy is set to always or everysec, and a background
     614 # saving process (a background save or AOF log background rewriting) is
     615 # performing a lot of I/O against the disk, in some Linux configurations
     616 # Redis may block too long on the fsync() call. Note that there is no fix for
     617 # this currently, as even performing fsync in a different thread will block
     618 # our synchronous write(2) call.
     619 #
     620 # In order to mitigate this problem it's possible to use the following option
     621 # that will prevent fsync() from being called in the main process while a
     622 # BGSAVE or BGREWRITEAOF is in progress.
     623 #
     624 # This means that while another child is saving, the durability of Redis is
     625 # the same as "appendfsync none". In practical terms, this means that it is
     626 # possible to lose up to 30 seconds of log in the worst scenario (with the
     627 # default Linux settings).
     628 #
     629 # If you have latency problems turn this to "yes". Otherwise leave it as
     630 # "no" that is the safest pick from the point of view of durability.
     631 no-appendfsync-on-rewrite no
     632 
     633 # Automatic rewrite of the append only file.
     634 # Redis is able to automatically rewrite the log file implicitly calling
     635 # BGREWRITEAOF when the AOF log size grows by the specified percentage.
     636 #
     637 # This is how it works: Redis remembers the size of the AOF file after the
     638 # latest rewrite (if no rewrite has happened since the restart, the size of
     639 # the AOF at startup is used).
     640 #
     641 # This base size is compared to the current size. If the current size is
     642 # bigger than the specified percentage, the rewrite is triggered. Also
     643 # you need to specify a minimal size for the AOF file to be rewritten, this
     644 # is useful to avoid rewriting the AOF file even if the percentage increase
     645 # is reached but it is still pretty small.
     646 #
     647 # Specify a percentage of zero in order to disable the automatic AOF
     648 # rewrite feature.
     649 
     650 auto-aof-rewrite-percentage 100
     651 auto-aof-rewrite-min-size 64mb
     652 
     653 # An AOF file may be found to be truncated at the end during the Redis
     654 # startup process, when the AOF data gets loaded back into memory.
     655 # This may happen when the system where Redis is running
     656 # crashes, especially when an ext4 filesystem is mounted without the
     657 # data=ordered option (however this can't happen when Redis itself
     658 # crashes or aborts but the operating system still works correctly).
     659 #
     660 # Redis can either exit with an error when this happens, or load as much
     661 # data as possible (the default now) and start if the AOF file is found
     662 # to be truncated at the end. The following option controls this behavior.
     663 #
     664 # If aof-load-truncated is set to yes, a truncated AOF file is loaded and
     665 # the Redis server starts emitting a log to inform the user of the event.
     666 # Otherwise if the option is set to no, the server aborts with an error
     667 # and refuses to start. When the option is set to no, the user requires
     668 # to fix the AOF file using the "redis-check-aof" utility before to restart
     669 # the server.
     670 #
     671 # Note that if the AOF file will be found to be corrupted in the middle
     672 # the server will still exit with an error. This option only applies when
     673 # Redis will try to read more data from the AOF file but not enough bytes
     674 # will be found.
     675 aof-load-truncated yes
     676 
     677 ################################ LUA SCRIPTING  ###############################
     678 
     679 # Max execution time of a Lua script in milliseconds.
     680 #
     681 # If the maximum execution time is reached Redis will log that a script is
     682 # still in execution after the maximum allowed time and will start to
     683 # reply to queries with an error.
     684 #
     685 # When a long running script exceeds the maximum execution time only the
     686 # SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be
     687 # used to stop a script that did not yet called write commands. The second
     688 # is the only way to shut down the server in the case a write command was
     689 # already issued by the script but the user doesn't want to wait for the natural
     690 # termination of the script.
     691 #
     692 # Set it to 0 or a negative value for unlimited execution without warnings.
     693 lua-time-limit 5000
     694 
     695 ################################ REDIS CLUSTER  ###############################
     696 #
     697 # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
     698 # WARNING EXPERIMENTAL: Redis Cluster is considered to be stable code, however
     699 # in order to mark it as "mature" we need to wait for a non trivial percentage
     700 # of users to deploy it in production.
     701 # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
     702 #
     703 # Normal Redis instances can't be part of a Redis Cluster; only nodes that are
     704 # started as cluster nodes can. In order to start a Redis instance as a
     705 # cluster node enable the cluster support uncommenting the following:
     706 #
     707 # cluster-enabled yes
     708 
     709 # Every cluster node has a cluster configuration file. This file is not
     710 # intended to be edited by hand. It is created and updated by Redis nodes.
     711 # Every Redis Cluster node requires a different cluster configuration file.
     712 # Make sure that instances running in the same system do not have
     713 # overlapping cluster configuration file names.
     714 #
     715 # cluster-config-file nodes-6379.conf
     716 
     717 # Cluster node timeout is the amount of milliseconds a node must be unreachable
     718 # for it to be considered in failure state.
     719 # Most other internal time limits are multiple of the node timeout.
     720 #
     721 # cluster-node-timeout 15000
     722 
     723 # A slave of a failing master will avoid to start a failover if its data
     724 # looks too old.
     725 #
     726 # There is no simple way for a slave to actually have a exact measure of
     727 # its "data age", so the following two checks are performed:
     728 #
     729 # 1) If there are multiple slaves able to failover, they exchange messages
     730 #    in order to try to give an advantage to the slave with the best
     731 #    replication offset (more data from the master processed).
     732 #    Slaves will try to get their rank by offset, and apply to the start
     733 #    of the failover a delay proportional to their rank.
     734 #
     735 # 2) Every single slave computes the time of the last interaction with
     736 #    its master. This can be the last ping or command received (if the master
     737 #    is still in the "connected" state), or the time that elapsed since the
     738 #    disconnection with the master (if the replication link is currently down).
     739 #    If the last interaction is too old, the slave will not try to failover
     740 #    at all.
     741 #
     742 # The point "2" can be tuned by user. Specifically a slave will not perform
     743 # the failover if, since the last interaction with the master, the time
     744 # elapsed is greater than:
     745 #
     746 #   (node-timeout * slave-validity-factor) + repl-ping-slave-period
     747 #
     748 # So for example if node-timeout is 30 seconds, and the slave-validity-factor
     749 # is 10, and assuming a default repl-ping-slave-period of 10 seconds, the
     750 # slave will not try to failover if it was not able to talk with the master
     751 # for longer than 310 seconds.
     752 #
     753 # A large slave-validity-factor may allow slaves with too old data to failover
     754 # a master, while a too small value may prevent the cluster from being able to
     755 # elect a slave at all.
     756 #
     757 # For maximum availability, it is possible to set the slave-validity-factor
     758 # to a value of 0, which means, that slaves will always try to failover the
     759 # master regardless of the last time they interacted with the master.
     760 # (However they'll always try to apply a delay proportional to their
     761 # offset rank).
     762 #
     763 # Zero is the only value able to guarantee that when all the partitions heal
     764 # the cluster will always be able to continue.
     765 #
     766 # cluster-slave-validity-factor 10
     767 
     768 # Cluster slaves are able to migrate to orphaned masters, that are masters
     769 # that are left without working slaves. This improves the cluster ability
     770 # to resist to failures as otherwise an orphaned master can't be failed over
     771 # in case of failure if it has no working slaves.
     772 #
     773 # Slaves migrate to orphaned masters only if there are still at least a
     774 # given number of other working slaves for their old master. This number
     775 # is the "migration barrier". A migration barrier of 1 means that a slave
     776 # will migrate only if there is at least 1 other working slave for its master
     777 # and so forth. It usually reflects the number of slaves you want for every
     778 # master in your cluster.
     779 #
     780 # Default is 1 (slaves migrate only if their masters remain with at least
     781 # one slave). To disable migration just set it to a very large value.
     782 # A value of 0 can be set but is useful only for debugging and dangerous
     783 # in production.
     784 #
     785 # cluster-migration-barrier 1
     786 
     787 # By default Redis Cluster nodes stop accepting queries if they detect there
     788 # is at least an hash slot uncovered (no available node is serving it).
     789 # This way if the cluster is partially down (for example a range of hash slots
     790 # are no longer covered) all the cluster becomes, eventually, unavailable.
     791 # It automatically returns available as soon as all the slots are covered again.
     792 #
     793 # However sometimes you want the subset of the cluster which is working,
     794 # to continue to accept queries for the part of the key space that is still
     795 # covered. In order to do so, just set the cluster-require-full-coverage
     796 # option to no.
     797 #
     798 # cluster-require-full-coverage yes
     799 
     800 # In order to setup your cluster make sure to read the documentation
     801 # available at http://redis.io web site.
     802 
     803 ################################## SLOW LOG ###################################
     804 
     805 # The Redis Slow Log is a system to log queries that exceeded a specified
     806 # execution time. The execution time does not include the I/O operations
     807 # like talking with the client, sending the reply and so forth,
     808 # but just the time needed to actually execute the command (this is the only
     809 # stage of command execution where the thread is blocked and can not serve
     810 # other requests in the meantime).
     811 #
     812 # You can configure the slow log with two parameters: one tells Redis
     813 # what is the execution time, in microseconds, to exceed in order for the
     814 # command to get logged, and the other parameter is the length of the
     815 # slow log. When a new command is logged the oldest one is removed from the
     816 # queue of logged commands.
     817 
     818 # The following time is expressed in microseconds, so 1000000 is equivalent
     819 # to one second. Note that a negative number disables the slow log, while
     820 # a value of zero forces the logging of every command.
     821 slowlog-log-slower-than 10000
     822 
     823 # There is no limit to this length. Just be aware that it will consume memory.
     824 # You can reclaim memory used by the slow log with SLOWLOG RESET.
     825 slowlog-max-len 128
     826 
     827 ################################ LATENCY MONITOR ##############################
     828 
     829 # The Redis latency monitoring subsystem samples different operations
     830 # at runtime in order to collect data related to possible sources of
     831 # latency of a Redis instance.
     832 #
     833 # Via the LATENCY command this information is available to the user that can
     834 # print graphs and obtain reports.
     835 #
     836 # The system only logs operations that were performed in a time equal or
     837 # greater than the amount of milliseconds specified via the
     838 # latency-monitor-threshold configuration directive. When its value is set
     839 # to zero, the latency monitor is turned off.
     840 #
     841 # By default latency monitoring is disabled since it is mostly not needed
     842 # if you don't have latency issues, and collecting data has a performance
     843 # impact, that while very small, can be measured under big load. Latency
     844 # monitoring can easily be enabled at runtime using the command
     845 # "CONFIG SET latency-monitor-threshold <milliseconds>" if needed.
     846 latency-monitor-threshold 0
     847 
     848 ############################# EVENT NOTIFICATION ##############################
     849 
     850 # Redis can notify Pub/Sub clients about events happening in the key space.
     851 # This feature is documented at http://redis.io/topics/notifications
     852 #
     853 # For instance if keyspace events notification is enabled, and a client
     854 # performs a DEL operation on key "foo" stored in the Database 0, two
     855 # messages will be published via Pub/Sub:
     856 #
     857 # PUBLISH __keyspace@0__:foo del
     858 # PUBLISH __keyevent@0__:del foo
     859 #
     860 # It is possible to select the events that Redis will notify among a set
     861 # of classes. Every class is identified by a single character:
     862 #
     863 #  K     Keyspace events, published with __keyspace@<db>__ prefix.
     864 #  E     Keyevent events, published with __keyevent@<db>__ prefix.
     865 #  g     Generic commands (non-type specific) like DEL, EXPIRE, RENAME, ...
     866 #  $     String commands
     867 #  l     List commands
     868 #  s     Set commands
     869 #  h     Hash commands
     870 #  z     Sorted set commands
     871 #  x     Expired events (events generated every time a key expires)
     872 #  e     Evicted events (events generated when a key is evicted for maxmemory)
     873 #  A     Alias for g$lshzxe, so that the "AKE" string means all the events.
     874 #
     875 #  The "notify-keyspace-events" takes as argument a string that is composed
     876 #  of zero or multiple characters. The empty string means that notifications
     877 #  are disabled.
     878 #
     879 #  Example: to enable list and generic events, from the point of view of the
     880 #           event name, use:
     881 #
     882 #  notify-keyspace-events Elg
     883 #
     884 #  Example 2: to get the stream of the expired keys subscribing to channel
     885 #             name __keyevent@0__:expired use:
     886 #
     887 #  notify-keyspace-events Ex
     888 #
     889 #  By default all notifications are disabled because most users don't need
     890 #  this feature and the feature has some overhead. Note that if you don't
     891 #  specify at least one of K or E, no events will be delivered.
     892 notify-keyspace-events ""
     893 
     894 ############################### ADVANCED CONFIG ###############################
     895 
     896 # Hashes are encoded using a memory efficient data structure when they have a
     897 # small number of entries, and the biggest entry does not exceed a given
     898 # threshold. These thresholds can be configured using the following directives.
     899 hash-max-ziplist-entries 512
     900 hash-max-ziplist-value 64
     901 
     902 # Lists are also encoded in a special way to save a lot of space.
     903 # The number of entries allowed per internal list node can be specified
     904 # as a fixed maximum size or a maximum number of elements.
     905 # For a fixed maximum size, use -5 through -1, meaning:
     906 # -5: max size: 64 Kb  <-- not recommended for normal workloads
     907 # -4: max size: 32 Kb  <-- not recommended
     908 # -3: max size: 16 Kb  <-- probably not recommended
     909 # -2: max size: 8 Kb   <-- good
     910 # -1: max size: 4 Kb   <-- good
     911 # Positive numbers mean store up to _exactly_ that number of elements
     912 # per list node.
     913 # The highest performing option is usually -2 (8 Kb size) or -1 (4 Kb size),
     914 # but if your use case is unique, adjust the settings as necessary.
     915 list-max-ziplist-size -2
     916 
     917 # Lists may also be compressed.
     918 # Compress depth is the number of quicklist ziplist nodes from *each* side of
     919 # the list to *exclude* from compression.  The head and tail of the list
     920 # are always uncompressed for fast push/pop operations.  Settings are:
     921 # 0: disable all list compression
     922 # 1: depth 1 means "don't start compressing until after 1 node into the list,
     923 #    going from either the head or tail"
     924 #    So: [head]->node->node->...->node->[tail]
     925 #    [head], [tail] will always be uncompressed; inner nodes will compress.
     926 # 2: [head]->[next]->node->node->...->node->[prev]->[tail]
     927 #    2 here means: don't compress head or head->next or tail->prev or tail,
     928 #    but compress all nodes between them.
     929 # 3: [head]->[next]->[next]->node->node->...->node->[prev]->[prev]->[tail]
     930 # etc.
     931 list-compress-depth 0
     932 
     933 # Sets have a special encoding in just one case: when a set is composed
     934 # of just strings that happen to be integers in radix 10 in the range
     935 # of 64 bit signed integers.
     936 # The following configuration setting sets the limit in the size of the
     937 # set in order to use this special memory saving encoding.
     938 set-max-intset-entries 512
     939 
     940 # Similarly to hashes and lists, sorted sets are also specially encoded in
     941 # order to save a lot of space. This encoding is only used when the length and
     942 # elements of a sorted set are below the following limits:
     943 zset-max-ziplist-entries 128
     944 zset-max-ziplist-value 64
     945 
     946 # HyperLogLog sparse representation bytes limit. The limit includes the
     947 # 16 bytes header. When an HyperLogLog using the sparse representation crosses
     948 # this limit, it is converted into the dense representation.
     949 #
     950 # A value greater than 16000 is totally useless, since at that point the
     951 # dense representation is more memory efficient.
     952 #
     953 # The suggested value is ~ 3000 in order to have the benefits of
     954 # the space efficient encoding without slowing down too much PFADD,
     955 # which is O(N) with the sparse encoding. The value can be raised to
     956 # ~ 10000 when CPU is not a concern, but space is, and the data set is
     957 # composed of many HyperLogLogs with cardinality in the 0 - 15000 range.
     958 hll-sparse-max-bytes 3000
     959 
     960 # Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
     961 # order to help rehashing the main Redis hash table (the one mapping top-level
     962 # keys to values). The hash table implementation Redis uses (see dict.c)
     963 # performs a lazy rehashing: the more operation you run into a hash table
     964 # that is rehashing, the more rehashing "steps" are performed, so if the
     965 # server is idle the rehashing is never complete and some more memory is used
     966 # by the hash table.
     967 #
     968 # The default is to use this millisecond 10 times every second in order to
     969 # actively rehash the main dictionaries, freeing memory when possible.
     970 #
     971 # If unsure:
     972 # use "activerehashing no" if you have hard latency requirements and it is
     973 # not a good thing in your environment that Redis can reply from time to time
     974 # to queries with 2 milliseconds delay.
     975 #
     976 # use "activerehashing yes" if you don't have such hard requirements but
     977 # want to free memory asap when possible.
     978 activerehashing yes
     979 
     980 # The client output buffer limits can be used to force disconnection of clients
     981 # that are not reading data from the server fast enough for some reason (a
     982 # common reason is that a Pub/Sub client can't consume messages as fast as the
     983 # publisher can produce them).
     984 #
     985 # The limit can be set differently for the three different classes of clients:
     986 #
     987 # normal -> normal clients including MONITOR clients
     988 # slave  -> slave clients
     989 # pubsub -> clients subscribed to at least one pubsub channel or pattern
     990 #
     991 # The syntax of every client-output-buffer-limit directive is the following:
     992 #
     993 # client-output-buffer-limit <class> <hard limit> <soft limit> <soft seconds>
     994 #
     995 # A client is immediately disconnected once the hard limit is reached, or if
     996 # the soft limit is reached and remains reached for the specified number of
     997 # seconds (continuously).
     998 # So for instance if the hard limit is 32 megabytes and the soft limit is
     999 # 16 megabytes / 10 seconds, the client will get disconnected immediately
    1000 # if the size of the output buffers reach 32 megabytes, but will also get
    1001 # disconnected if the client reaches 16 megabytes and continuously overcomes
    1002 # the limit for 10 seconds.
    1003 #
    1004 # By default normal clients are not limited because they don't receive data
    1005 # without asking (in a push way), but just after a request, so only
    1006 # asynchronous clients may create a scenario where data is requested faster
    1007 # than it can read.
    1008 #
    1009 # Instead there is a default limit for pubsub and slave clients, since
    1010 # subscribers and slaves receive data in a push fashion.
    1011 #
    1012 # Both the hard or the soft limit can be disabled by setting them to zero.
    1013 client-output-buffer-limit normal 0 0 0
    1014 client-output-buffer-limit slave 256mb 64mb 60
    1015 client-output-buffer-limit pubsub 32mb 8mb 60
    1016 
    1017 # Redis calls an internal function to perform many background tasks, like
    1018 # closing connections of clients in timeot, purging expired keys that are
    1019 # never requested, and so forth.
    1020 #
    1021 # Not all tasks are perforemd with the same frequency, but Redis checks for
    1022 # tasks to perform according to the specified "hz" value.
    1023 #
    1024 # By default "hz" is set to 10. Raising the value will use more CPU when
    1025 # Redis is idle, but at the same time will make Redis more responsive when
    1026 # there are many keys expiring at the same time, and timeouts may be
    1027 # handled with more precision.
    1028 #
    1029 # The range is between 1 and 500, however a value over 100 is usually not
    1030 # a good idea. Most users should use the default of 10 and raise this up to
    1031 # 100 only in environments where very low latency is required.
    1032 hz 10
    1033 
    1034 # When a child rewrites the AOF file, if the following option is enabled
    1035 # the file will be fsync-ed every 32 MB of data generated. This is useful
    1036 # in order to commit the file to the disk more incrementally and avoid
    1037 # big latency spikes.
    1038 aof-rewrite-incremental-fsync yes
    1039 
    1040 ################################## INCLUDES ###################################
    1041 
    1042 # Include one or more other config files here.  This is useful if you
    1043 # have a standard template that goes to all Redis server but also need
    1044 # to customize a few per-server settings.  Include files can include
    1045 # other files, so use this wisely.
    1046 #
    1047 # include /path/to/local.conf
    1048 # include /path/to/other.conf
    View Code

    如果单独用:

     docker run -d -p 6379:6379 --name redis --restart always -v /config/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf -v /config/redis/data:/data redis --appendonly yes --requirepass 9090 

    windows redis 配置

    https://github.com/microsoftarchive/redis/releases下载版本

    下载解压后的文件列表

    可以配置密码:打开redis.windows.conf文件,找到:

    放开注释,配置密码,

    然后把redis注册未服务

    cmd切换到redis目录

    输入命令:

    redis-server.exe --service-install redis.windows.conf

     

    参考:https://www.cnblogs.com/GuoJunwen/p/9238624.html

    上传镜像到hub.docker仓库

    通过push镜像,其他地方只需要pull即可,

    1:hub.docker上创建账号

    2:docker login 登陆,会提示输入用户名和密码

    注意,密码是盲打的,看到 Login Succeeded 说明登陆成功

    3:通过 docker push 镜像名 就可以上传到自己的镜像仓库了

    但这里要注意几点,镜像是有命名规范的  ,比如我我想把镜像 cn/api 上传

    我会 这样:docker push cn/api

    会提示资源拒绝访问

     是因为镜像命名规范: 组织名称/镜像名称 我这里的组织是个人。所以cn必须是自己的用户名

    所以必须是:byniqing/api  那怎么办呢?可以通过命名修改镜像

     docker tag cn/api:v1 byniqing/api:pro 

    查看镜像。已经成功修改,修改前后的IMAGE ID是不变的,可以看看

    修改前:

     

     修改后:

     

     再次push试试

    等待上传成功,hub上就有了

     

     然后你 就可以直接pull了

    如果一个镜像经常升级,就会出现很多悬空镜像,这些悬空镜像是可以删除的。

    下面 这些none的镜像。就是悬空镜像

    删除命令:

    docker rmi $(docker images -f "dangling=true" -q)

    # 或者

    docker image prune -a -f

    Portainer管理镜像

     未完待续

    上传镜像到hub.docker

    源码:https://github.com/byniqing/docker-compose

  • 相关阅读:
    dpkg 删除 百度网盘 程序
    ubuntu 安装go
    解决 swap file “*.swp”already exists!问题
    ROS Topic 常用指令
    正交概念
    vim 永久显示行号 & 临时显示行号
    awk、grep、sed
    Keil中使用Astyel进行C语言的格式化
    红黑树学习
    802.11 对于multicast 和 broadcast的处理
  • 原文地址:https://www.cnblogs.com/nsky/p/10853194.html
Copyright © 2011-2022 走看看