import pandas as pd
from sklearn.cross_validation import train_test_split
from sklearn.linear_model import LinearRegression #数据1
tem16_1 = [3113,3122,3131,3137,3146,3149,3157,3166,3172,3178,3172,3151,3137,3131,3128,3107,3095,3081,3070,3049,3038,3023,3011,2997,2988,2985,2974,2957,2942,2928,2916,2901,2887,2884,2879,2870,2856,2848,2836,2828,2810,2802,2793,2785,2776,2768,2762,2756,2762,2765,2762,2771,2748,2731,2708,2697,2674,2697,2702,2725,2737,2759,2771,2787,2802,2819,2833,2845,2856,2859,2873,2879,2896,2898,2910,2922,2933,2945,2957,2968,2980,2985,3000,3011,3023,3038,3046,3058,3070,3072,3084,3090,3098,3107,3119,3122,3131,3134,3140,3143,3149,3154,3151,3166,3189,3178,3181,3169,3166,3169,3154,3137,3116,3107,3090,3093,3087,3075,3052,3038,3026,3017,3008,3002,2994,2991,2988,2977,2957,2951,2936,2931,2936,2957,2980,2962,2948,2928,2913,2898,2879,2867,2865,2856,2853,2848,2842,2833,2830,2819,2813,2807,2799,2796,2787,2785,2779,2773]
#数据2
tem16_2 = [4185,4209,4233,4256,4279,4303,4326,4349,4372,4398,4421,4414,4387,4383,4360,4349,4326,4303,4279,4264,4233,4209,4185,4161,4113,4088,4064,4043,4019,4023,4002,3978,3954,3933,3912,3891,3870,3845,3824,3802,3781,3759,3715,3693,3671,3648,3626,3603,3580,3534,3507,3496,3490,3467,3451,3429,3419,3432,3454,3464,3486,3496,3518,3542,3565,3588,3611,3633,3656,3678,3700,3722,3744,3766,3788,3809,3831,3873,3895,3919,3940,3964,3988,4010,4039,4060,4084,4108,4133,4157,4181,4205,4228,4252,4276,4299,4322,4345,4368,4391,4414,4436,4458,4455,4470,4485,4462,4458,4451,4447,4425,4402,4387,4364,4368,4349,4326,4307,4287,4264,4240,4221,4225,4197,4177,4169,4149,4133,4113,4088,4064,4043,4019,3995,3985,3968,3947,3926,3905,3884,3863,3842,3820,3799,3777,3773,3759,3737,3715,3693,3671,3648,3626,3603,3576,3553,3530,3507]
#回归数据
result = [2364,2356,2356,2353,2353,2353,2353,2367,2367,2367,2383,2362,2435,2460,2427,2427,2451,2446,2424,2430,2407,2399,2386,2367,2372,2375,2364,2342,2326,2309,2285,2283,2288,2277,2285,2307,2329,2351,2372,2396,2421,2492,2312,2283,2280,2266,2253,2239,2234,2231,2247,2247,2242,2220,2198,2098,2034,2031,2053,2031,2093,2071,2093,2077,2142,2163,2185,2206,2215,2215,2215,2215,2217,2217,2217,2217,2217,2239,2239,2239,2256,2256,2256,2256,2274,2274,2274,2274,2293,2309,2293,2309,2331,2331,2331,2331,2331,2331,2331,2331,2331,2331,2356,2367,2380,2383,2386,2370,2378,2372,2351,2337,2320,2299,2293,2312,2334,2329,2323,2315,2304,2307,2301,2304,2290,2272,2256,2234,2237,2245,2266,2258,2261,2253,2266,2245,2223,2212,2215,2204,2198,2201,2223,2245,2261,2277,2269,2288,2290,2315,2261,2253,2266,2245,2223,2212,2215,2215]
#以DataFrame存储数据
data = pd.DataFrame([tem16_1,tem16_2,result],index=['tem16_1','tem16_2','result'])
data = data.T
X = data[['tem16_1','tem16_2']] y = data['result']
#测试集和训练集
X_train, X_test, y_train, y_test = train_test_split(X, y, random_state=1)
#训练
linreg = LinearRegression()
linreg.fit(X_train, y_train)
#结果
print linreg.intercept_
print linreg.coef_
print zip(['tem16_1','tem16_2'], linreg.coef_)

多元线性回归(pandas/scikit-learn)的更多相关文章

  1. (原创)(三)机器学习笔记之Scikit Learn的线性回归模型初探

    一.Scikit Learn中使用estimator三部曲 1. 构造estimator 2. 训练模型:fit 3. 利用模型进行预测:predict 二.模型评价 模型训练好后,度量模型拟合效果的 ...

  2. (原创)(四)机器学习笔记之Scikit Learn的Logistic回归初探

    目录 5.3 使用LogisticRegressionCV进行正则化的 Logistic Regression 参数调优 一.Scikit Learn中有关logistics回归函数的介绍 1. 交叉 ...

  3. [机器学习Lesson4]多元线性回归

    1. 多元线性回归定义 多元线性回归也被称为多元线性回归. 我们现在介绍方程的符号,我们可以有任意数量的输入变量. 这些多个特征的假设函数的多变量形式如下: hθ(x)=θ0+θ1x1+θ2x2+θ3 ...

  4. 利用TensorFlow实现多元线性回归

    利用TensorFlow实现多元线性回归,代码如下: # -*- coding:utf-8 -*- import tensorflow as tf import numpy as np from sk ...

  5. Scikit Learn: 在python中机器学习

    转自:http://my.oschina.net/u/175377/blog/84420#OSC_h2_23 Scikit Learn: 在python中机器学习 Warning 警告:有些没能理解的 ...

  6. 机器学习——Day 3 多元线性回归

    写在开头 由于某些原因开始了机器学习,为了更好的理解和深入的思考(记录)所以开始写博客. 学习教程来源于github的Avik-Jain的100-Days-Of-MLCode 英文版:https:// ...

  7. Tensorflow之多元线性回归问题(以波士顿房价预测为例)

    一.根据波士顿房价信息进行预测,多元线性回归+特征数据归一化 #读取数据 %matplotlib notebook import tensorflow as tf import matplotlib. ...

  8. 100天搞定机器学习|Day3多元线性回归

    前情回顾 [第二天100天搞定机器学习|Day2简单线性回归分析][1],我们学习了简单线性回归分析,这个模型非常简单,很容易理解.实现方式是sklearn中的LinearRegression,我们也 ...

  9. R语言解读多元线性回归模型

    转载:http://blog.fens.me/r-multi-linear-regression/ 前言 本文接上一篇R语言解读一元线性回归模型.在许多生活和工作的实际问题中,影响因变量的因素可能不止 ...

随机推荐

  1. python爬虫之路——初识基本页面构造原理

    通过chrome浏览器的使用简单介绍网页构成 360浏览器使用右键审查元素,Chrome浏览器使用右键检查,都可查看网页代码. 网页代码有两部分:HTML文件和CSS样式.其中有<script& ...

  2. x86,x64,i386,i686

    x64其实就是64位, x86其实就是32位. 1. i386 适用于intel和AMD所有32位的cpu.以及via采用X86架构的32的cpu. intel平台包括8086,80286,80386 ...

  3. 01_4_Struts路径问题

    01_4_Struts路径问题 1. Struts路径问题说明 struts2中的路径问题是根据action的路径而不是jsp路径来确定,所有尽量不要使用相对路径. 虽然可以使用redirect方式解 ...

  4. 解决cocos2dx 3.x 导入cocostudio的ui界面出现错位问题

    笔者今天发现导入cocostudio的ui界面时,会有部分控件出现错位的现象,后来我看了一下源码,发现是部分控件是没有继承 Layout类,导致不能设置控件位置造成,原因可以看看cocos2dx 源码 ...

  5. APP上线碰到的问题:Non-public API usage

    ①.Non-public API usage:The app references non-public symbols in XXXX: _UICreateCGImageFromIOSurface ...

  6. nodejs 静态资源服务与接口代理跨域

    首先需要 npm install express 和 npm install request 代码如下: const express = require('express'); const path ...

  7. VNC远程登录端使用经验之一

    1.vnc/xmanager都是经常用的远程登录软件.vnc有个缺点就是他的进程不会自动退出比如如果开了PID1再去开PID2...PIDn.那么前面的PIDn-1个进程就会一直运行如果不手动kill ...

  8. list 方法总结整理

    #!/usr/bin/env python #Python 3.7.0 列表常用方法 __author__ = "lrtao2010" #创建列表 # a = [] # b = [ ...

  9. navicat12.0.24破解方法,简单易操作,亲测可行

    navicat12.0.24 32bit 链接:https://pan.baidu.com/s/1dakPje0AzwE86p6ZRHfnsQ 密码:f1ve 破解文件 链接:https://pan. ...

  10. Applied Nonparametric Statistics-lec1

    参考网址: https://onlinecourses.science.psu.edu/stat464/node/2 Binomial Distribution Normal Distribution ...