zoukankan      html  css  js  c++  java
  • Hadoop实例之Java代码实现利用MapReduce求π值

    需求:假如有一个边长为1的正方形。以正方形的一个端点为圆心,以1为半径,画一个圆弧,于是在正方形内就有了一个直角扇形。在正方形里随机生成若干的点,则有些点是在扇形内,有些点是在扇形外。正方形的面积是1,扇形的面积是0.25*Pi。设点的数量一共是n,扇形内的点数量是nc,在点足够多足够密集的情况下,会近似有nc/n的比值约等于扇形面积与正方形面积的比值,也就是nc/n= 0.25*Pi/1,即Pi = 4*nc/n

    首先是随机生成点的问题,利用Halton序列算法随机生成的样本点十分均匀,计算精度较高,效果比较好。

    下面是网上找到的一个利用Halton序列算法随机生成的样本点的代码:

    public class Pi {
        static int digit = 40;
        private int[] bases= new int[2];
        private double[] baseDigit = new double[2];
        private double[][] background = new double[2][digit];
        private long index;
        
        Pi(int[] base) {
            bases = base.clone();
            index = 0;
     
            for(int i=0; i<bases.length; i++) {
                double b = 1.0/bases[i];
                baseDigit[i] = b;
                for(int j=0; j<digit; j++) {
                    background[i][j] = j == 0 ? b : background[i][j-1]*b;
                }
            }
        }
        
        double[] getNext() {
            index++;
            
            double[] result = {0,0};
     
            for(int i=0; i<bases.length; i++) {
                long num = index;
                int j = 0;
                while(num != 0) {
                    result[i] += num % bases[i] * background[i][j++];
                    num /= bases[i];
                }
            }
            
            return result;
        }
        
        public static void main(String[] args) {
            int[] base = {2,5};
            Pi test = new Pi(base);
            for(int x = 0; x < 100; x++){
                double[] t = test.getNext();
                System.out.println(t[0] + "	" + t[1]);
            }
            
        }
    
    
    }

    下面是计算π值的代码:

    package mapreduce;
    
    import java.io.IOException;
    
    
    import org.apache.hadoop.conf.Configuration;
    import org.apache.hadoop.fs.Path;
    import org.apache.hadoop.io.DoubleWritable;
    import org.apache.hadoop.io.IntWritable;
    import org.apache.hadoop.io.Text;
    import org.apache.hadoop.mapreduce.Job;
    import org.apache.hadoop.mapreduce.Mapper;
    import org.apache.hadoop.mapreduce.Reducer;
    import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
    import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
    import mapreduce.Pi;//下面生成随机数的时候需要这个类,该类即上面那部分代码
    
    /**
     * 
     * @author sakura
     * 2019.9.3
     * 利用MapReduce计算π值
     *
     */
    public class CalPI {
        public static class PiMapper extends Mapper<Object, Text, Text, IntWritable>{
    
            int number=0; //定义一个变量,用来存放一共生成的点数
            
            //读取文件,每一行都是一个map 本程序读取的文件为十行,每行都是100000
            public void map(Object key, Text value, Context context) throws IOException, InterruptedException {
                int pointNum = Integer.parseInt(value.toString());//将读取到的那一行赋值给pointNum
                number=number+pointNum;//将总点数赋值给number
                 int[] base = {2,5};//生成随机点所用
                Pi test = new Pi(base);//生成随机点所用
                for(int x = 0; x < number; x++){ //循环生成随机点
                    double[] t = test.getNext();//随机生成点,并将坐标存入数组
                    System.out.println(t[0] + "	" + t[1]);//控制台输出随机点的坐标
                    IntWritable result = new IntWritable(0); //定义输出值
                    if((t[0]*t[0]+t[1]*t[1])<=1)//判断生成的点是否在扇形面积内
                    {
                        result = new IntWritable(1);//如果在,将输出值赋值为1
                    }
                    value.set(String.valueOf(number));//定义输出键,输出键为当前生成点的总数
                    context.write(value, result);//写入
                }
            }
        }
    
        public static class PiReducer extends Reducer<Text,IntWritable,Text,DoubleWritable> {
            private DoubleWritable result = new DoubleWritable();//声明输出值
    
            public void reduce(Text key, Iterable<IntWritable> values,Context context) throws IOException, InterruptedException {
    
                double pointNum =Double.parseDouble(key.toString());//获取输入的键
                double sum = 0;//定义总数
                for (IntWritable val : values) {//循环从values里取值,累加和赋值给sum
                    sum += val.get();
                }
                result.set(sum/pointNum*4);//将计算得到的π值赋值给result
                
                context.write(key, result);//将键值,即生成点总数,和result,即计算得到的π值作为一个键值对写入context
            }
        }
    
        public static void main(String[] args) throws Exception {
            Configuration conf = new Configuration();
            Job job = Job.getInstance(conf,"calculate pi");
            job.setJarByClass(CalPI.class);
            job.setMapperClass(PiMapper.class);
            job.setReducerClass(PiReducer.class);
            job.setMapOutputKeyClass(Text.class);
            job.setMapOutputValueClass(IntWritable.class);
            job.setOutputKeyClass(Text.class);
            job.setOutputValueClass(DoubleWritable.class);
    
            Path  in  =  new  Path("hdfs://192.168.68.130:9000/user/hadoop/nai.txt");  //读入文件地址
            Path  out  = new Path("hdfs://192.168.68.130:9000/user/hadoop/output4");  //输出文件地址,output4不能存在
            FileInputFormat.addInputPath(job, in);
            FileOutputFormat.setOutputPath(job, out);
            System.exit(job.waitForCompletion(true) ? 0  :  1);  
      
        }
    
    
    }
  • 相关阅读:
    Ansible中文权威指南学习
    gitlab
    Python-Day01-task
    Python-Day01
    CentOs6.7 python2.6升级到2.7.11
    网站访问慢排查方法及解决方案
    LAMP环境下zabbix安装配置
    监控知识点概述
    Java抽象类、接口和内部类
    Java 访问控制
  • 原文地址:https://www.cnblogs.com/sakura--/p/11455467.html
Copyright © 2011-2022 走看看