日韩av黄I国产麻豆传媒I国产91av视频在线观看I日韩一区二区三区在线看I美女国产在线I麻豆视频国产在线观看I成人黄色短片

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 >

Python数据分析_第06课:数据清洗与初步分析_笔记

發布時間:2023/12/8 43 豆豆
生活随笔 收集整理的這篇文章主要介紹了 Python数据分析_第06课:数据清洗与初步分析_笔记 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

文章目錄

  • 缺失值處理——拉格朗日插值法
  • dataframe合并
  • 索引上的合并
  • 軸向連接
  • 合并重疊數據
  • 重塑層次化索引
  • 長寬格式的轉換
  • 移除重復數據
  • 利用函數或映射進行數據轉換
  • 數據標準化
    • 最小-最大規范化
    • 零-均值規范化
    • 小數定標規范化
  • 替換值
  • 重命名軸索引
  • 離散化與面元劃分
  • 檢測和過濾異常值
  • 排列與隨機采樣
  • 計算指標與啞變量
  • 屬性構造
  • 字符串對象方法
  • 正則表達式
  • pandas中矢量化的字符串函數
  • 示例:USDA食品數據庫

GitHub: https://github.com/RealEmperor/Python-for-Data-Analysis

缺失值處理——拉格朗日插值法

import matplotlib.pyplot as plt import numpy as np import pandas as pd from pandas import Series, DataFrame from scipy.interpolate import lagrange # 導入拉格朗日插值函數np.random.seed(12345) plt.rc('figure', figsize=(10, 6))inputfile = 'data/catering_sale.xls' # 銷量數據路徑 outputfile = 'data/sales.xls' # 輸出數據路徑data = pd.read_excel(inputfile) # 讀入數據# 過濾異常值,將其變為空值 """ data[u'銷量'][(data[u'銷量'] < 400) | (data[u'銷量'] > 5000)] = None 上面這樣寫會有警告: SettingWithCopyWarning: A value is trying to be set on a copy of a slice from a DataFrame如果要更改原始數據,請使用單一賦值操作(loc): data.loc[(data[u'銷量'] < 400) | (data[u'銷量'] > 5000), u'銷量'] = None如果想要一個副本,請確保強制讓 Pandas 創建副本: error_data = data.copy() error_data.loc[(error_data[u'銷量'] < 400) | (error_data[u'銷量'] > 5000), u'銷量'] = None參考:https://www.jianshu.com/p/72274ccb647a """ data.loc[(data[u'銷量'] < 400) | (data[u'銷量'] > 5000), u'銷量'] = None# 自定義列向量插值函數 # s為列向量,n為被插值的位置,k為取前后的數據個數,默認為5 def ployinterp_column(s, n, k=5):y = s[list(range(n - k, n)) + list(range(n + 1, n + 1 + k))] # 取數y = y[y.notnull()] # 剔除空值return lagrange(y.index, list(y))(n) # 插值并返回插值結果# 逐個元素判斷是否需要插值 for i in data.columns:for j in range(len(data)):if (data[i].isnull())[j]: # 如果為空即插值。"""data[i][j] = ployinterp_column(data[i], j) 這樣寫會有警告:SettingWithCopyWarning: A value is trying to be set on a copy of a slice from a DataFrame"""data.loc[j, i] = ployinterp_column(data[i], j)data.to_excel(outputfile) # 輸出結果,寫入文件

dataframe合并

#dataframe合并 # 1 df1 = DataFrame({'key': ['b', 'b', 'a', 'c', 'a', 'a', 'b'],'data1': range(7)}) df2 = DataFrame({'key': ['a', 'b', 'd'],'data2': range(3)}) print(df1) print(df2) data1 key 0 0 b 1 1 b 2 2 a 3 3 c 4 4 a 5 5 a 6 6 bdata2 key 0 0 a 1 1 b 2 2 d pd.merge(df1, df2) data1keydata2012345
0b1
1b1
6b1
2a0
4a0
5a0
pd.merge(df1, df2, on='key') data1keydata2012345
0b1
1b1
6b1
2a0
4a0
5a0
# 2 df3 = DataFrame({'lkey': ['b', 'b', 'a', 'c', 'a', 'a', 'b'],'data1': range(7)}) df4 = DataFrame({'rkey': ['a', 'b', 'd'],'data2': range(3)}) pd.merge(df3, df4, left_on='lkey', right_on='rkey') data1lkeydata2rkey012345
0b1b
1b1b
6b1b
2a0a
4a0a
5a0a
pd.merge(df1, df2, how='outer') data1keydata201234567
0.0b1.0
1.0b1.0
6.0b1.0
2.0a0.0
4.0a0.0
5.0a0.0
3.0cNaN
NaNd2.0
# 3 df1 = DataFrame({'key': ['b', 'b', 'a', 'c', 'a', 'b'],'data1': range(6)}) df2 = DataFrame({'key': ['a', 'b', 'a', 'b', 'd'],'data2': range(5)}) print(df1) print(df2) data1 key 0 0 b 1 1 b 2 2 a 3 3 c 4 4 a 5 5 bdata2 key 0 0 a 1 1 b 2 2 a 3 3 b 4 4 d pd.merge(df1, df2, on='key', how='left') data1keydata2012345678910
0b1.0
0b3.0
1b1.0
1b3.0
2a0.0
2a2.0
3cNaN
4a0.0
4a2.0
5b1.0
5b3.0
pd.merge(df1, df2, how='inner') data1keydata20123456789
0b1
0b3
1b1
1b3
5b1
5b3
2a0
2a2
4a0
4a2
# 4 left = DataFrame({'key1': ['foo', 'foo', 'bar'],'key2': ['one', 'two', 'one'],'lval': [1, 2, 3]}) right = DataFrame({'key1': ['foo', 'foo', 'bar', 'bar'],'key2': ['one', 'one', 'one', 'two'],'rval': [4, 5, 6, 7]}) pd.merge(left, right, on=['key1', 'key2'], how='outer') key1key2lvalrval01234
fooone1.04.0
fooone1.05.0
footwo2.0NaN
barone3.06.0
bartwoNaN7.0
# 5 pd.merge(left, right, on='key1') key1key2_xlvalkey2_yrval012345
fooone1one4
fooone1one5
footwo2one4
footwo2one5
barone3one6
barone3two7
pd.merge(left, right, on='key1', suffixes=('_left', '_right')) key1key2_leftlvalkey2_rightrval012345
fooone1one4
fooone1one5
footwo2one4
footwo2one5
barone3one6
barone3two7

索引上的合并

# 1 left1 = DataFrame({'key': ['a', 'b', 'a', 'a', 'b', 'c'], 'value': range(6)}) right1 = DataFrame({'group_val': [3.5, 7]}, index=['a', 'b']) print(left1) print(right1) key value 0 a 0 1 b 1 2 a 2 3 a 3 4 b 4 5 c 5group_val a 3.5 b 7.0 pd.merge(left1, right1, left_on='key', right_index=True) keyvaluegroup_val02314
a03.5
a23.5
a33.5
b17.0
b47.0
pd.merge(left1, right1, left_on='key', right_index=True, how='outer') keyvaluegroup_val023145
a03.5
a23.5
a33.5
b17.0
b47.0
c5NaN
# 2 lefth = DataFrame({'key1': ['Ohio', 'Ohio', 'Ohio', 'Nevada', 'Nevada'],'key2': [2000, 2001, 2002, 2001, 2002],'data': np.arange(5.)}) righth = DataFrame(np.arange(12).reshape((6, 2)),index=[['Nevada', 'Nevada', 'Ohio', 'Ohio', 'Ohio', 'Ohio'],[2001, 2000, 2000, 2000, 2001, 2002]],columns=['event1', 'event2']) print(lefth) print(righth) data key1 key2 0 0.0 Ohio 2000 1 1.0 Ohio 2001 2 2.0 Ohio 2002 3 3.0 Nevada 2001 4 4.0 Nevada 2002event1 event2 Nevada 2001 0 12000 2 3 Ohio 2000 4 52000 6 72001 8 92002 10 11 pd.merge(lefth, righth, left_on=['key1', 'key2'], right_index=True) datakey1key2event1event200123
0.0Ohio200045
0.0Ohio200067
1.0Ohio200189
2.0Ohio20021011
3.0Nevada200101
pd.merge(lefth, righth, left_on=['key1', 'key2'],right_index=True, how='outer') datakey1key2event1event20012344
0.0Ohio20004.05.0
0.0Ohio20006.07.0
1.0Ohio20018.09.0
2.0Ohio200210.011.0
3.0Nevada20010.01.0
4.0Nevada2002NaNNaN
NaNNevada20002.03.0
left2 = DataFrame([[1., 2.], [3., 4.], [5., 6.]], index=['a', 'c', 'e'],columns=['Ohio', 'Nevada']) right2 = DataFrame([[7., 8.], [9., 10.], [11., 12.], [13, 14]],index=['b', 'c', 'd', 'e'], columns=['Missouri', 'Alabama']) print(left2) print(right2) Ohio Nevada a 1.0 2.0 c 3.0 4.0 e 5.0 6.0Missouri Alabama b 7.0 8.0 c 9.0 10.0 d 11.0 12.0 e 13.0 14.0 pd.merge(left2, right2, how='outer', left_index=True, right_index=True) OhioNevadaMissouriAlabamaabcde
1.02.0NaNNaN
NaNNaN7.08.0
3.04.09.010.0
NaNNaN11.012.0
5.06.013.014.0
# 3 left2.join(right2, how='outer') OhioNevadaMissouriAlabamaabcde
1.02.0NaNNaN
NaNNaN7.08.0
3.04.09.010.0
NaNNaN11.012.0
5.06.013.014.0
left1.join(right1, on='key') keyvaluegroup_val012345
a03.5
b17.0
a23.5
a33.5
b47.0
c5NaN
# 4 another = DataFrame([[7., 8.], [9., 10.], [11., 12.], [16., 17.]],index=['a', 'c', 'e', 'f'], columns=['New York', 'Oregon']) left2.join([right2, another]) OhioNevadaMissouriAlabamaNew YorkOregonace
1.02.0NaNNaN7.08.0
3.04.09.010.09.010.0
5.06.013.014.011.012.0
left2.join([right2, another], how='outer') OhioNevadaMissouriAlabamaNew YorkOregonabcdef
1.02.0NaNNaN7.08.0
NaNNaN7.08.0NaNNaN
3.04.09.010.09.010.0
NaNNaN11.012.0NaNNaN
5.06.013.014.011.012.0
NaNNaNNaNNaN16.017.0

軸向連接

# 1 arr = np.arange(12).reshape((3, 4)) print(arr)np.concatenate([arr, arr], axis=1) [[ 0 1 2 3][ 4 5 6 7][ 8 9 10 11]]array([[ 0, 1, 2, 3, 0, 1, 2, 3],[ 4, 5, 6, 7, 4, 5, 6, 7],[ 8, 9, 10, 11, 8, 9, 10, 11]]) # 2 s1 = Series([0, 1], index=['a', 'b']) s2 = Series([2, 3, 4], index=['c', 'd', 'e']) s3 = Series([5, 6], index=['f', 'g'])pd.concat([s1, s2, s3]) a 0 b 1 c 2 d 3 e 4 f 5 g 6 dtype: int64 pd.concat([s1, s2, s3], axis=0) a 0 b 1 c 2 d 3 e 4 f 5 g 6 dtype: int64 pd.concat([s1, s2, s3], axis=1) 012abcdefg
0.0NaNNaN
1.0NaNNaN
NaN2.0NaN
NaN3.0NaN
NaN4.0NaN
NaNNaN5.0
NaNNaN6.0
s4 = pd.concat([s1 * 5, s3]) print(s4) a 0 b 5 f 5 g 6 dtype: int64 pd.concat([s1, s4], axis=1) 01abfg
0.00
1.05
NaN5
NaN6
pd.concat([s1, s4], axis=1, join='inner') 01ab
00
15
pd.concat([s1, s4], axis=1, join_axes=[['a', 'c', 'b', 'e']]) 01acbe
0.00.0
NaNNaN
1.05.0
NaNNaN
# 3 result = pd.concat([s1, s1, s3], keys=['one', 'two', 'three']) print(result) one a 0b 1 two a 0b 1 three f 5g 6 dtype: int64 result.unstack() abfgonetwothree
0.01.0NaNNaN
0.01.0NaNNaN
NaNNaN5.06.0
# 4 pd.concat([s1, s2, s3], axis=1, keys=['one', 'two', 'three']) onetwothreeabcdefg
0.0NaNNaN
1.0NaNNaN
NaN2.0NaN
NaN3.0NaN
NaN4.0NaN
NaNNaN5.0
NaNNaN6.0
df1 = DataFrame(np.arange(6).reshape(3, 2), index=['a', 'b', 'c'],columns=['one', 'two']) df2 = DataFrame(5 + np.arange(4).reshape(2, 2), index=['a', 'c'],columns=['three', 'four']) pd.concat([df1, df2], axis=1, keys=['level1', 'level2']) level1level2onetwothreefourabc
015.06.0
23NaNNaN
457.08.0
pd.concat({'level1': df1, 'level2': df2}, axis=1) level1level2onetwothreefourabc
015.06.0
23NaNNaN
457.08.0
pd.concat([df1, df2], axis=1, keys=['level1', 'level2'],names=['upper', 'lower']) upperlevel1level2loweronetwothreefourabc
015.06.0
23NaNNaN
457.08.0
# 5 df1 = DataFrame(np.random.randn(3, 4), columns=['a', 'b', 'c', 'd']) df2 = DataFrame(np.random.randn(2, 3), columns=['b', 'd', 'a']) print(df1) print(df2) a b c d 0 -0.204708 0.478943 -0.519439 -0.555730 1 1.965781 1.393406 0.092908 0.281746 2 0.769023 1.246435 1.007189 -1.296221b d a 0 0.274992 0.228913 1.352917 1 0.886429 -2.001637 -0.371843 pd.concat([df1, df2], ignore_index=True) abcd01234
-0.2047080.478943-0.519439-0.555730
1.9657811.3934060.0929080.281746
0.7690231.2464351.007189-1.296221
1.3529170.274992NaN0.228913
-0.3718430.886429NaN-2.001637

合并重疊數據

# 1 a = Series([np.nan, 2.5, np.nan, 3.5, 4.5, np.nan],index=['f', 'e', 'd', 'c', 'b', 'a']) b = Series(np.arange(len(a), dtype=np.float64),index=['f', 'e', 'd', 'c', 'b', 'a']) b[-1] = np.nanprint(a) print(b)np.where(pd.isnull(a), b, a) f NaN e 2.5 d NaN c 3.5 b 4.5 a NaN dtype: float64 f 0.0 e 1.0 d 2.0 c 3.0 b 4.0 a NaN dtype: float64array([ 0. , 2.5, 2. , 3.5, 4.5, nan]) # 2 b[:-2].combine_first(a[2:]) a NaN b 4.5 c 3.0 d 2.0 e 1.0 f 0.0 dtype: float64 # 3 df1 = DataFrame({'a': [1., np.nan, 5., np.nan],'b': [np.nan, 2., np.nan, 6.],'c': range(2, 18, 4)}) df2 = DataFrame({'a': [5., 4., np.nan, 3., 7.],'b': [np.nan, 3., 4., 6., 8.]}) df1.combine_first(df2) abc01234
1.0NaN2.0
4.02.06.0
5.04.010.0
3.06.014.0
7.08.0NaN

重塑層次化索引

# 1 data = DataFrame(np.arange(6).reshape((2, 3)),index=pd.Index(['Ohio', 'Colorado'], name='state'),columns=pd.Index(['one', 'two', 'three'], name='number')) print(data) number one two three state Ohio 0 1 2 Colorado 3 4 5 # stack 把 column 轉成 index result = data.stack() print(result) state number Ohio one 0two 1three 2 Colorado one 3two 4three 5 dtype: int32

# unstack 把 index 轉成 column result.unstack() numberonetwothreestateOhioColorado
012
345

result.unstack(0) stateOhioColoradonumberonetwothree
03
14
25
result.unstack('state') stateOhioColoradonumberonetwothree
03
14
25
# 2 s1 = Series([0, 1, 2, 3], index=['a', 'b', 'c', 'd']) s2 = Series([4, 5, 6], index=['c', 'd', 'e']) data2 = pd.concat([s1, s2], keys=['one', 'two']) print(data2) one a 0b 1c 2d 3 two c 4d 5e 6 dtype: int64 # data2.stack() # 沒有column 執行會報錯 data2.unstack() abcdeonetwo
0.01.02.03.0NaN
NaNNaN4.05.06.0

data2.unstack().stack() one a 0.0b 1.0c 2.0d 3.0 two c 4.0d 5.0e 6.0 dtype: float64

data2.unstack().stack(dropna=False) one a 0.0b 1.0c 2.0d 3.0e NaN two a NaNb NaNc 4.0d 5.0e 6.0 dtype: float64 # 3 df = DataFrame({'left': result, 'right': result + 5},columns=pd.Index(['left', 'right'], name='side')) print(df) side left right state number Ohio one 0 5two 1 6three 2 7 Colorado one 3 8two 4 9three 5 10 df.unstack('state') sideleftrightstateOhioColoradoOhioColoradonumberonetwothree
0358
1469
25710
df.unstack('state').stack('side') stateColoradoOhionumbersideoneleftrighttwoleftrightthreeleftright
30
85
41
96
52
107

長寬格式的轉換

data = pd.read_csv('data/macrodata.csv') print(data) year quarter realgdp realcons realinv realgovt realdpi \ 0 1959.0 1.0 2710.349 1707.4 286.898 470.045 1886.9 1 1959.0 2.0 2778.801 1733.7 310.859 481.301 1919.7 2 1959.0 3.0 2775.488 1751.8 289.226 491.260 1916.4 3 1959.0 4.0 2785.204 1753.7 299.356 484.052 1931.3 4 1960.0 1.0 2847.699 1770.5 331.722 462.199 1955.5 5 1960.0 2.0 2834.390 1792.9 298.152 460.400 1966.1 6 1960.0 3.0 2839.022 1785.8 296.375 474.676 1967.8 7 1960.0 4.0 2802.616 1788.2 259.764 476.434 1966.6 8 1961.0 1.0 2819.264 1787.7 266.405 475.854 1984.5 9 1961.0 2.0 2872.005 1814.3 286.246 480.328 2014.4 10 1961.0 3.0 2918.419 1823.1 310.227 493.828 2041.9 11 1961.0 4.0 2977.830 1859.6 315.463 502.521 2082.0 12 1962.0 1.0 3031.241 1879.4 334.271 520.960 2101.7 13 1962.0 2.0 3064.709 1902.5 331.039 523.066 2125.2 14 1962.0 3.0 3093.047 1917.9 336.962 538.838 2137.0 15 1962.0 4.0 3100.563 1945.1 325.650 535.912 2154.6 16 1963.0 1.0 3141.087 1958.2 343.721 522.917 2172.5 17 1963.0 2.0 3180.447 1976.9 348.730 518.108 2193.1 18 1963.0 3.0 3240.332 2003.8 360.102 546.893 2217.9 19 1963.0 4.0 3264.967 2020.6 364.534 532.383 2254.6 20 1964.0 1.0 3338.246 2060.5 379.523 529.686 2299.6 21 1964.0 2.0 3376.587 2096.7 377.778 526.175 2362.1 22 1964.0 3.0 3422.469 2135.2 386.754 522.008 2392.7 23 1964.0 4.0 3431.957 2141.2 389.910 514.603 2420.4 24 1965.0 1.0 3516.251 2188.8 429.145 508.006 2447.4 25 1965.0 2.0 3563.960 2213.0 429.119 508.931 2474.5 26 1965.0 3.0 3636.285 2251.0 444.444 529.446 2542.6 27 1965.0 4.0 3724.014 2314.3 446.493 544.121 2594.1 28 1966.0 1.0 3815.423 2348.5 484.244 556.593 2618.4 29 1966.0 2.0 3828.124 2354.5 475.408 571.371 2624.7 .. ... ... ... ... ... ... ... 173 2002.0 2.0 11538.770 7997.8 1810.779 774.408 8658.9 174 2002.0 3.0 11596.430 8052.0 1814.531 786.673 8629.2 175 2002.0 4.0 11598.824 8080.6 1813.219 799.967 8649.6 176 2003.0 1.0 11645.819 8122.3 1813.141 800.196 8681.3 177 2003.0 2.0 11738.706 8197.8 1823.698 838.775 8812.5 178 2003.0 3.0 11935.461 8312.1 1889.883 839.598 8935.4 179 2003.0 4.0 12042.817 8358.0 1959.783 845.722 8986.4 180 2004.0 1.0 12127.623 8437.6 1970.015 856.570 9025.9 181 2004.0 2.0 12213.818 8483.2 2055.580 861.440 9115.0 182 2004.0 3.0 12303.533 8555.8 2082.231 876.385 9175.9 183 2004.0 4.0 12410.282 8654.2 2125.152 865.596 9303.4 184 2005.0 1.0 12534.113 8719.0 2170.299 869.204 9189.6 185 2005.0 2.0 12587.535 8802.9 2131.468 870.044 9253.0 186 2005.0 3.0 12683.153 8865.6 2154.949 890.394 9308.0 187 2005.0 4.0 12748.699 8888.5 2232.193 875.557 9358.7 188 2006.0 1.0 12915.938 8986.6 2264.721 900.511 9533.8 189 2006.0 2.0 12962.462 9035.0 2261.247 892.839 9617.3 190 2006.0 3.0 12965.916 9090.7 2229.636 892.002 9662.5 191 2006.0 4.0 13060.679 9181.6 2165.966 894.404 9788.8 192 2007.0 1.0 13099.901 9265.1 2132.609 882.766 9830.2 193 2007.0 2.0 13203.977 9291.5 2162.214 898.713 9842.7 194 2007.0 3.0 13321.109 9335.6 2166.491 918.983 9883.9 195 2007.0 4.0 13391.249 9363.6 2123.426 925.110 9886.2 196 2008.0 1.0 13366.865 9349.6 2082.886 943.372 9826.8 197 2008.0 2.0 13415.266 9351.0 2026.518 961.280 10059.0 198 2008.0 3.0 13324.600 9267.7 1990.693 991.551 9838.3 199 2008.0 4.0 13141.920 9195.3 1857.661 1007.273 9920.4 200 2009.0 1.0 12925.410 9209.2 1558.494 996.287 9926.4 201 2009.0 2.0 12901.504 9189.0 1456.678 1023.528 10077.5 202 2009.0 3.0 12990.341 9256.0 1486.398 1044.088 10040.6 cpi m1 tbilrate unemp pop infl realint 0 28.980 139.7 2.82 5.8 177.146 0.00 0.00 1 29.150 141.7 3.08 5.1 177.830 2.34 0.74 2 29.350 140.5 3.82 5.3 178.657 2.74 1.09 3 29.370 140.0 4.33 5.6 179.386 0.27 4.06 4 29.540 139.6 3.50 5.2 180.007 2.31 1.19 5 29.550 140.2 2.68 5.2 180.671 0.14 2.55 6 29.750 140.9 2.36 5.6 181.528 2.70 -0.34 7 29.840 141.1 2.29 6.3 182.287 1.21 1.08 8 29.810 142.1 2.37 6.8 182.992 -0.40 2.77 9 29.920 142.9 2.29 7.0 183.691 1.47 0.81 10 29.980 144.1 2.32 6.8 184.524 0.80 1.52 11 30.040 145.2 2.60 6.2 185.242 0.80 1.80 12 30.210 146.4 2.73 5.6 185.874 2.26 0.47 13 30.220 146.5 2.78 5.5 186.538 0.13 2.65 14 30.380 146.7 2.78 5.6 187.323 2.11 0.67 15 30.440 148.3 2.87 5.5 188.013 0.79 2.08 16 30.480 149.7 2.90 5.8 188.580 0.53 2.38 17 30.690 151.3 3.03 5.7 189.242 2.75 0.29 18 30.750 152.6 3.38 5.5 190.028 0.78 2.60 19 30.940 153.7 3.52 5.6 190.668 2.46 1.06 20 30.950 154.8 3.51 5.5 191.245 0.13 3.38 21 31.020 156.8 3.47 5.2 191.889 0.90 2.57 22 31.120 159.2 3.53 5.0 192.631 1.29 2.25 23 31.280 160.7 3.76 5.0 193.223 2.05 1.71 24 31.380 162.0 3.93 4.9 193.709 1.28 2.65 25 31.580 163.1 3.84 4.7 194.303 2.54 1.30 26 31.650 166.0 3.93 4.4 194.997 0.89 3.04 27 31.880 169.1 4.35 4.1 195.539 2.90 1.46 28 32.280 171.8 4.62 3.9 195.999 4.99 -0.37 29 32.450 170.3 4.65 3.8 196.560 2.10 2.55 .. ... ... ... ... ... ... ... 173 180.000 1199.5 1.70 5.8 288.028 1.56 0.14 174 181.200 1204.0 1.61 5.7 288.783 2.66 -1.05 175 182.600 1226.8 1.20 5.8 289.421 3.08 -1.88 176 183.200 1248.4 1.14 5.9 290.019 1.31 -0.17 177 183.700 1287.9 0.96 6.2 290.704 1.09 -0.13 178 184.900 1297.3 0.94 6.1 291.449 2.60 -1.67 179 186.300 1306.1 0.90 5.8 292.057 3.02 -2.11 180 187.400 1332.1 0.94 5.7 292.635 2.35 -1.42 181 189.100 1340.5 1.21 5.6 293.310 3.61 -2.41 182 190.800 1361.0 1.63 5.4 294.066 3.58 -1.95 183 191.800 1366.6 2.20 5.4 294.741 2.09 0.11 184 193.800 1357.8 2.69 5.3 295.308 4.15 -1.46 185 194.700 1366.6 3.01 5.1 295.994 1.85 1.16 186 199.200 1375.0 3.52 5.0 296.770 9.14 -5.62 187 199.400 1380.6 4.00 4.9 297.435 0.40 3.60 188 200.700 1380.5 4.51 4.7 298.061 2.60 1.91 189 202.700 1369.2 4.82 4.7 298.766 3.97 0.85 190 201.900 1369.4 4.90 4.7 299.593 -1.58 6.48 191 203.574 1373.6 4.92 4.4 300.320 3.30 1.62 192 205.920 1379.7 4.95 4.5 300.977 4.58 0.36 193 207.338 1370.0 4.72 4.5 301.714 2.75 1.97 194 209.133 1379.2 4.00 4.7 302.509 3.45 0.55 195 212.495 1377.4 3.01 4.8 303.204 6.38 -3.37 196 213.997 1384.0 1.56 4.9 303.803 2.82 -1.26 197 218.610 1409.3 1.74 5.4 304.483 8.53 -6.79 198 216.889 1474.7 1.17 6.0 305.270 -3.16 4.33 199 212.174 1576.5 0.12 6.9 305.952 -8.79 8.91 200 212.671 1592.8 0.22 8.1 306.547 0.94 -0.71 201 214.469 1653.6 0.18 9.2 307.226 3.37 -3.19 202 216.385 1673.9 0.12 9.6 308.013 3.56 -3.44 [203 rows x 14 columns] periods = pd.PeriodIndex(year=data.year, quarter=data.quarter, name='date') print(periods) PeriodIndex(['1959Q1', '1959Q2', '1959Q3', '1959Q4', '1960Q1', '1960Q2','1960Q3', '1960Q4', '1961Q1', '1961Q2',...'2007Q2', '2007Q3', '2007Q4', '2008Q1', '2008Q2', '2008Q3','2008Q4', '2009Q1', '2009Q2', '2009Q3'],dtype='period[Q-DEC]', name='date', length=203, freq='Q-DEC') data = DataFrame(data.to_records(),columns=pd.Index(['realgdp', 'infl', 'unemp'], name='item'),index=periods.to_timestamp('D', 'end')) print(data) item realgdp infl unemp date 1959-03-31 2710.349 0.00 5.8 1959-06-30 2778.801 2.34 5.1 1959-09-30 2775.488 2.74 5.3 1959-12-31 2785.204 0.27 5.6 1960-03-31 2847.699 2.31 5.2 1960-06-30 2834.390 0.14 5.2 1960-09-30 2839.022 2.70 5.6 1960-12-31 2802.616 1.21 6.3 1961-03-31 2819.264 -0.40 6.8 1961-06-30 2872.005 1.47 7.0 1961-09-30 2918.419 0.80 6.8 1961-12-31 2977.830 0.80 6.2 1962-03-31 3031.241 2.26 5.6 1962-06-30 3064.709 0.13 5.5 1962-09-30 3093.047 2.11 5.6 1962-12-31 3100.563 0.79 5.5 1963-03-31 3141.087 0.53 5.8 1963-06-30 3180.447 2.75 5.7 1963-09-30 3240.332 0.78 5.5 1963-12-31 3264.967 2.46 5.6 1964-03-31 3338.246 0.13 5.5 1964-06-30 3376.587 0.90 5.2 1964-09-30 3422.469 1.29 5.0 1964-12-31 3431.957 2.05 5.0 1965-03-31 3516.251 1.28 4.9 1965-06-30 3563.960 2.54 4.7 1965-09-30 3636.285 0.89 4.4 1965-12-31 3724.014 2.90 4.1 1966-03-31 3815.423 4.99 3.9 1966-06-30 3828.124 2.10 3.8 ... ... ... ... 2002-06-30 11538.770 1.56 5.8 2002-09-30 11596.430 2.66 5.7 2002-12-31 11598.824 3.08 5.8 2003-03-31 11645.819 1.31 5.9 2003-06-30 11738.706 1.09 6.2 2003-09-30 11935.461 2.60 6.1 2003-12-31 12042.817 3.02 5.8 2004-03-31 12127.623 2.35 5.7 2004-06-30 12213.818 3.61 5.6 2004-09-30 12303.533 3.58 5.4 2004-12-31 12410.282 2.09 5.4 2005-03-31 12534.113 4.15 5.3 2005-06-30 12587.535 1.85 5.1 2005-09-30 12683.153 9.14 5.0 2005-12-31 12748.699 0.40 4.9 2006-03-31 12915.938 2.60 4.7 2006-06-30 12962.462 3.97 4.7 2006-09-30 12965.916 -1.58 4.7 2006-12-31 13060.679 3.30 4.4 2007-03-31 13099.901 4.58 4.5 2007-06-30 13203.977 2.75 4.5 2007-09-30 13321.109 3.45 4.7 2007-12-31 13391.249 6.38 4.8 2008-03-31 13366.865 2.82 4.9 2008-06-30 13415.266 8.53 5.4 2008-09-30 13324.600 -3.16 6.0 2008-12-31 13141.920 -8.79 6.9 2009-03-31 12925.410 0.94 8.1 2009-06-30 12901.504 3.37 9.2 2009-09-30 12990.341 3.56 9.6[203 rows x 3 columns] ldata = data.stack().reset_index().rename(columns={0: 'value'}) print(ldata) date item value 0 1959-03-31 realgdp 2710.349 1 1959-03-31 infl 0.000 2 1959-03-31 unemp 5.800 3 1959-06-30 realgdp 2778.801 4 1959-06-30 infl 2.340 5 1959-06-30 unemp 5.100 6 1959-09-30 realgdp 2775.488 7 1959-09-30 infl 2.740 8 1959-09-30 unemp 5.300 9 1959-12-31 realgdp 2785.204 10 1959-12-31 infl 0.270 11 1959-12-31 unemp 5.600 12 1960-03-31 realgdp 2847.699 13 1960-03-31 infl 2.310 14 1960-03-31 unemp 5.200 15 1960-06-30 realgdp 2834.390 16 1960-06-30 infl 0.140 17 1960-06-30 unemp 5.200 18 1960-09-30 realgdp 2839.022 19 1960-09-30 infl 2.700 20 1960-09-30 unemp 5.600 21 1960-12-31 realgdp 2802.616 22 1960-12-31 infl 1.210 23 1960-12-31 unemp 6.300 24 1961-03-31 realgdp 2819.264 25 1961-03-31 infl -0.400 26 1961-03-31 unemp 6.800 27 1961-06-30 realgdp 2872.005 28 1961-06-30 infl 1.470 29 1961-06-30 unemp 7.000 .. ... ... ... 579 2007-06-30 realgdp 13203.977 580 2007-06-30 infl 2.750 581 2007-06-30 unemp 4.500 582 2007-09-30 realgdp 13321.109 583 2007-09-30 infl 3.450 584 2007-09-30 unemp 4.700 585 2007-12-31 realgdp 13391.249 586 2007-12-31 infl 6.380 587 2007-12-31 unemp 4.800 588 2008-03-31 realgdp 13366.865 589 2008-03-31 infl 2.820 590 2008-03-31 unemp 4.900 591 2008-06-30 realgdp 13415.266 592 2008-06-30 infl 8.530 593 2008-06-30 unemp 5.400 594 2008-09-30 realgdp 13324.600 595 2008-09-30 infl -3.160 596 2008-09-30 unemp 6.000 597 2008-12-31 realgdp 13141.920 598 2008-12-31 infl -8.790 599 2008-12-31 unemp 6.900 600 2009-03-31 realgdp 12925.410 601 2009-03-31 infl 0.940 602 2009-03-31 unemp 8.100 603 2009-06-30 realgdp 12901.504 604 2009-06-30 infl 3.370 605 2009-06-30 unemp 9.200 606 2009-09-30 realgdp 12990.341 607 2009-09-30 infl 3.560 608 2009-09-30 unemp 9.600[609 rows x 3 columns] wdata = ldata.pivot(index='date',columns= 'item',values= 'value') print(wdata) item infl realgdp unemp date 1959-03-31 0.00 2710.349 5.8 1959-06-30 2.34 2778.801 5.1 1959-09-30 2.74 2775.488 5.3 1959-12-31 0.27 2785.204 5.6 1960-03-31 2.31 2847.699 5.2 1960-06-30 0.14 2834.390 5.2 1960-09-30 2.70 2839.022 5.6 1960-12-31 1.21 2802.616 6.3 1961-03-31 -0.40 2819.264 6.8 1961-06-30 1.47 2872.005 7.0 1961-09-30 0.80 2918.419 6.8 1961-12-31 0.80 2977.830 6.2 1962-03-31 2.26 3031.241 5.6 1962-06-30 0.13 3064.709 5.5 1962-09-30 2.11 3093.047 5.6 1962-12-31 0.79 3100.563 5.5 1963-03-31 0.53 3141.087 5.8 1963-06-30 2.75 3180.447 5.7 1963-09-30 0.78 3240.332 5.5 1963-12-31 2.46 3264.967 5.6 1964-03-31 0.13 3338.246 5.5 1964-06-30 0.90 3376.587 5.2 1964-09-30 1.29 3422.469 5.0 1964-12-31 2.05 3431.957 5.0 1965-03-31 1.28 3516.251 4.9 1965-06-30 2.54 3563.960 4.7 1965-09-30 0.89 3636.285 4.4 1965-12-31 2.90 3724.014 4.1 1966-03-31 4.99 3815.423 3.9 1966-06-30 2.10 3828.124 3.8 ... ... ... ... 2002-06-30 1.56 11538.770 5.8 2002-09-30 2.66 11596.430 5.7 2002-12-31 3.08 11598.824 5.8 2003-03-31 1.31 11645.819 5.9 2003-06-30 1.09 11738.706 6.2 2003-09-30 2.60 11935.461 6.1 2003-12-31 3.02 12042.817 5.8 2004-03-31 2.35 12127.623 5.7 2004-06-30 3.61 12213.818 5.6 2004-09-30 3.58 12303.533 5.4 2004-12-31 2.09 12410.282 5.4 2005-03-31 4.15 12534.113 5.3 2005-06-30 1.85 12587.535 5.1 2005-09-30 9.14 12683.153 5.0 2005-12-31 0.40 12748.699 4.9 2006-03-31 2.60 12915.938 4.7 2006-06-30 3.97 12962.462 4.7 2006-09-30 -1.58 12965.916 4.7 2006-12-31 3.30 13060.679 4.4 2007-03-31 4.58 13099.901 4.5 2007-06-30 2.75 13203.977 4.5 2007-09-30 3.45 13321.109 4.7 2007-12-31 6.38 13391.249 4.8 2008-03-31 2.82 13366.865 4.9 2008-06-30 8.53 13415.266 5.4 2008-09-30 -3.16 13324.600 6.0 2008-12-31 -8.79 13141.920 6.9 2009-03-31 0.94 12925.410 8.1 2009-06-30 3.37 12901.504 9.2 2009-09-30 3.56 12990.341 9.6[203 rows x 3 columns] # 2 ldata[:10] dateitemvalue0123456789
1959-03-31realgdp2710.349
1959-03-31infl0.000
1959-03-31unemp5.800
1959-06-30realgdp2778.801
1959-06-30infl2.340
1959-06-30unemp5.100
1959-09-30realgdp2775.488
1959-09-30infl2.740
1959-09-30unemp5.300
1959-12-31realgdp2785.204
pivoted = ldata.pivot('date', 'item', 'value') pivoted.head() iteminflrealgdpunempdate1959-03-311959-06-301959-09-301959-12-311960-03-31
0.002710.3495.8
2.342778.8015.1
2.742775.4885.3
0.272785.2045.6
2.312847.6995.2
ldata['value2'] = np.random.randn(len(ldata)) ldata[:10] dateitemvaluevalue20123456789
1959-03-31realgdp2710.3491.669025
1959-03-31infl0.000-0.438570
1959-03-31unemp5.800-0.539741
1959-06-30realgdp2778.8010.476985
1959-06-30infl2.3403.248944
1959-06-30unemp5.100-1.021228
1959-09-30realgdp2775.488-0.577087
1959-09-30infl2.7400.124121
1959-09-30unemp5.3000.302614
1959-12-31realgdp2785.2040.523772
pivoted = ldata.pivot('date', 'item') pivoted[:5] valuevalue2iteminflrealgdpunempinflrealgdpunempdate1959-03-311959-06-301959-09-301959-12-311960-03-31
0.002710.3495.8-0.4385701.669025-0.539741
2.342778.8015.13.2489440.476985-1.021228
2.742775.4885.30.124121-0.5770870.302614
0.272785.2045.60.0009400.5237721.343810
2.312847.6995.2-0.831154-0.713544-2.370232
pivoted['value'][:5] iteminflrealgdpunempdate1959-03-311959-06-301959-09-301959-12-311960-03-31
0.002710.3495.8
2.342778.8015.1
2.742775.4885.3
0.272785.2045.6
2.312847.6995.2
unstacked = ldata.set_index(['date', 'item']).unstack('item') unstacked[:7] valuevalue2iteminflrealgdpunempinflrealgdpunempdate1959-03-311959-06-301959-09-301959-12-311960-03-311960-06-301960-09-30
0.002710.3495.8-0.4385701.669025-0.539741
2.342778.8015.13.2489440.476985-1.021228
2.742775.4885.30.124121-0.5770870.302614
0.272785.2045.60.0009400.5237721.343810
2.312847.6995.2-0.831154-0.713544-2.370232
0.142834.3905.2-0.860757-1.8607610.560145
2.702839.0225.60.119827-1.265934-1.063512

移除重復數據

data = DataFrame({'k1': ['one'] * 3 + ['two'] * 4,'k2': [1, 1, 2, 3, 3, 4, 4]}) print(data) k1 k2 0 one 1 1 one 1 2 one 2 3 two 3 4 two 3 5 two 4 6 two 4 data.duplicated() 0 False 1 True 2 False 3 False 4 True 5 False 6 True dtype: bool data.drop_duplicates() k1k20235
one1
one2
two3
two4
data['v1'] = range(7) print(data) k1 k2 v1 0 one 1 0 1 one 1 1 2 one 2 2 3 two 3 3 4 two 3 4 5 two 4 5 6 two 4 6 data.drop_duplicates(['k1']) k1k2v103
one10
two33
data.drop_duplicates(['k1', 'k2'], keep='last') k1k2v11246
one11
one22
two34
two46

利用函數或映射進行數據轉換

# 1 data = DataFrame({'food': ['bacon', 'pulled pork', 'bacon', 'Pastrami','corned beef', 'Bacon', 'pastrami', 'honey ham','nova lox'],'ounces': [4, 3, 12, 6, 7.5, 8, 3, 5, 6]}) print(data) food ounces 0 bacon 4.0 1 pulled pork 3.0 2 bacon 12.0 3 Pastrami 6.0 4 corned beef 7.5 5 Bacon 8.0 6 pastrami 3.0 7 honey ham 5.0 8 nova lox 6.0 meat_to_animal = {'bacon': 'pig','pulled pork': 'pig','pastrami': 'cow','corned beef': 'cow','honey ham': 'pig','nova lox': 'salmon' }data['animal'] = data['food'].map(str.lower).map(meat_to_animal) print(data) food ounces animal 0 bacon 4.0 pig 1 pulled pork 3.0 pig 2 bacon 12.0 pig 3 Pastrami 6.0 cow 4 corned beef 7.5 cow 5 Bacon 8.0 pig 6 pastrami 3.0 cow 7 honey ham 5.0 pig 8 nova lox 6.0 salmon data['food'].map(lambda x: meat_to_animal[x.lower()]) 0 pig 1 pig 2 pig 3 cow 4 cow 5 pig 6 cow 7 pig 8 salmon Name: food, dtype: object

數據標準化

datafile = 'data/normalization_data.xls' # 參數初始化 data = pd.read_excel(datafile, header=None) # 讀取數據 print(data) 0 1 2 3 0 78 521 602 2863 1 144 -600 -521 2245 2 95 -457 468 -1283 3 69 596 695 1054 4 190 527 691 2051 5 101 403 470 2487 6 146 413 435 2571

最小-最大規范化

(data - data.min()) / (data.max() - data.min()) # 最小-最大規范化 01230123456
0.0743800.9372910.9235201.000000
0.6198350.0000000.0000000.850941
0.2148760.1195650.8133220.000000
0.0000001.0000001.0000000.563676
1.0000000.9423080.9967110.804149
0.2644630.8386290.8149670.909310
0.6363640.8469900.7861840.929571

零-均值規范化

(data - data.mean()) / data.std() # 零-均值規范化 01230123456
-0.9053830.6358630.4645310.798149
0.604678-1.587675-2.1931670.369390
-0.516428-1.3040300.147406-2.078279
-1.1113010.7846280.684625-0.456906
1.6571460.6477650.6751590.234796
-0.3791500.4018070.1521390.537286
0.6504380.4216420.0693080.595564

小數定標規范化

data / 10 ** np.ceil(np.log10(data.abs().max())) # 小數定標規范化 01230123456
0.0780.5210.6020.2863
0.144-0.600-0.5210.2245
0.095-0.4570.468-0.1283
0.0690.5960.6950.1054
0.1900.5270.6910.2051
0.1010.4030.4700.2487
0.1460.4130.4350.2571

替換值

data = Series([1., -999., 2., -999., -1000., 3.]) print(data) 0 1.0 1 -999.0 2 2.0 3 -999.0 4 -1000.0 5 3.0 dtype: float64 data.replace(-999, np.nan) 0 1.0 1 NaN 2 2.0 3 NaN 4 -1000.0 5 3.0 dtype: float64 data.replace([-999, -1000], np.nan) 0 1.0 1 NaN 2 2.0 3 NaN 4 NaN 5 3.0 dtype: float64 data.replace([-999, -1000], [np.nan, 0]) 0 1.0 1 NaN 2 2.0 3 NaN 4 0.0 5 3.0 dtype: float64 data.replace({-999: np.nan, -1000: 0}) 0 1.0 1 NaN 2 2.0 3 NaN 4 0.0 5 3.0 dtype: float64

重命名軸索引

data = DataFrame(np.arange(12).reshape((3, 4)),index=['Ohio', 'Colorado', 'New York'],columns=['one', 'two', 'three', 'four']) print(data) one two three four Ohio 0 1 2 3 Colorado 4 5 6 7 New York 8 9 10 11 data.index.map(str.upper) Index(['OHIO', 'COLORADO', 'NEW YORK'], dtype='object') data.index = data.index.map(str.upper) print(data) one two three four OHIO 0 1 2 3 COLORADO 4 5 6 7 NEW YORK 8 9 10 11 data.rename(index=str.title, columns=str.upper) ONETWOTHREEFOUROhioColoradoNew York
0123
4567
891011
data.rename(index={'OHIO': 'INDIANA'},columns={'three': 'peekaboo'}) onetwopeekaboofourINDIANACOLORADONEW YORK
0123
4567
891011
# 總是返回DataFrame的引用 data.rename(index={'OHIO': 'INDIANA'}, inplace=True) print(data) one two three four INDIANA 0 1 2 3 COLORADO 4 5 6 7 NEW YORK 8 9 10 11

離散化與面元劃分

# 1 ages = [20, 22, 25, 27, 21, 23, 37, 31, 61, 45, 41, 32]bins = [18, 25, 35, 60, 100] cats = pd.cut(ages, bins) print(cats) [(18, 25], (18, 25], (18, 25], (25, 35], (18, 25], ..., (25, 35], (60, 100], (35, 60], (35, 60], (25, 35]] Length: 12 Categories (4, interval[int64]): [(18, 25] < (25, 35] < (35, 60] < (60, 100]] cats.codes array([0, 0, 0, 1, 0, 0, 2, 1, 3, 2, 2, 1], dtype=int8) cats.categories IntervalIndex([(18, 25], (25, 35], (35, 60], (60, 100]]closed='right',dtype='interval[int64]') pd.value_counts(cats) (18, 25] 5 (35, 60] 3 (25, 35] 3 (60, 100] 1 dtype: int64 pd.cut(ages, [18, 26, 36, 61, 100], right=False) [[18, 26), [18, 26), [18, 26), [26, 36), [18, 26), ..., [26, 36), [61, 100), [36, 61), [36, 61), [26, 36)] Length: 12 Categories (4, interval[int64]): [[18, 26) < [26, 36) < [36, 61) < [61, 100)] group_names = ['Youth', 'YoungAdult', 'MiddleAged', 'Senior'] pd.cut(ages, bins, labels=group_names) [Youth, Youth, Youth, YoungAdult, Youth, ..., YoungAdult, Senior, MiddleAged, MiddleAged, YoungAdult] Length: 12 Categories (4, object): [MiddleAged < Senior < YoungAdult < Youth] data = np.random.rand(20) pd.cut(data, 4, precision=2) [(0.45, 0.67], (0.23, 0.45], (0.0037, 0.23], (0.45, 0.67], (0.67, 0.9], ..., (0.67, 0.9], (0.0037, 0.23], (0.0037, 0.23], (0.23, 0.45], (0.23, 0.45]] Length: 20 Categories (4, interval[float64]): [(0.0037, 0.23] < (0.23, 0.45] < (0.45, 0.67] < (0.67, 0.9]] # 2 data = np.random.randn(1000) # Normally distributed cats = pd.qcut(data, 4) # Cut into quartiles print(cats) [(-0.022, 0.641], (-3.746, -0.635], (0.641, 3.26], (-3.746, -0.635], (-0.022, 0.641], ..., (-0.022, 0.641], (0.641, 3.26], (-0.635, -0.022], (0.641, 3.26], (-0.635, -0.022]] Length: 1000 Categories (4, interval[float64]): [(-3.746, -0.635] < (-0.635, -0.022] < (-0.022, 0.641] < (0.641, 3.26]] pd.value_counts(cats) (0.641, 3.26] 250 (-0.022, 0.641] 250 (-0.635, -0.022] 250 (-3.746, -0.635] 250 dtype: int64 pd.qcut(data, [0, 0.1, 0.5, 0.9, 1.]) [(-0.022, 1.302], (-1.266, -0.022], (-0.022, 1.302], (-3.746, -1.266], (-0.022, 1.302], ..., (-0.022, 1.302], (-0.022, 1.302], (-1.266, -0.022], (-0.022, 1.302], (-1.266, -0.022]] Length: 1000 Categories (4, interval[float64]): [(-3.746, -1.266] < (-1.266, -0.022] < (-0.022, 1.302] < (1.302, 3.26]]

檢測和過濾異常值

# 1 np.random.seed(12345) data = DataFrame(np.random.randn(1000, 4)) data.describe() 0123countmeanstdmin25%50%75%max
1000.0000001000.0000001000.0000001000.000000
-0.0676840.0679240.025598-0.002298
0.9980350.9921061.0068350.996794
-3.428254-3.548824-3.184377-3.745356
-0.774890-0.591841-0.641675-0.644144
-0.1164010.1011430.002073-0.013611
0.6163660.7802820.6803910.654328
3.3666262.6536563.2603833.927528
col = data[3] col[np.abs(col) > 3] 97 3.927528 305 -3.399312 400 -3.745356 Name: 3, dtype: float64 data[(np.abs(data) > 3).any(1)] 0123597102305324400499523586808900
-0.5397410.4769853.248944-1.021228
-0.7743630.5529360.1060613.927528
-0.655054-0.5652303.1768730.959533
-2.3155550.457246-0.025907-3.399312
0.0501881.9513123.2603830.963301
0.1463260.508391-0.196713-3.745356
-0.293333-0.242459-3.0569901.918403
-3.428254-0.296336-0.439938-0.867165
0.2751441.179227-3.1843771.369891
-0.362528-3.5488241.553205-2.186301
3.366626-2.3722140.8510101.332846
# 2 data[np.abs(data) > 3] = np.sign(data) * 3 data.describe() 0123countmeanstdmin25%50%75%max
1000.0000001000.0000001000.0000001000.000000
-0.0676230.0684730.025153-0.002081
0.9954850.9902531.0039770.989736
-3.000000-3.000000-3.000000-3.000000
-0.774890-0.591841-0.641675-0.644144
-0.1164010.1011430.002073-0.013611
0.6163660.7802820.6803910.654328
3.0000002.6536563.0000003.000000

排列與隨機采樣

df = DataFrame(np.arange(5 * 4).reshape((5, 4))) print(df) 0 1 2 3 0 0 1 2 3 1 4 5 6 7 2 8 9 10 11 3 12 13 14 15 4 16 17 18 19 sampler = np.random.permutation(5) print(sampler) [1 0 2 3 4] df.take(sampler) 012310234
4567
0123
891011
12131415
16171819
# 2 df.take(np.random.permutation(len(df))[:3]) 0123134
4567
12131415
16171819
# 3 bag = np.array([5, 7, -1, 6, 4]) sampler = np.random.randint(0, len(bag), size=10) print(sampler) [4 4 2 2 2 0 3 0 4 1] draws = bag.take(sampler) print(draws) [ 4 4 -1 -1 -1 5 6 5 4 7]

計算指標與啞變量

# 1 df = DataFrame({'key': ['b', 'b', 'a', 'c', 'a', 'b'],'data1': range(6)}) print(df) data1 key 0 0 b 1 1 b 2 2 a 3 3 c 4 4 a 5 5 b pd.get_dummies(df['key']) abc012345
010
010
100
001
100
010
dummies = pd.get_dummies(df['key'], prefix='key') print(dummies) key_a key_b key_c 0 0 1 0 1 0 1 0 2 1 0 0 3 0 0 1 4 1 0 0 5 0 1 0 df_with_dummy = df[['data1']].join(dummies) print(df_with_dummy) data1 key_a key_b key_c 0 0 0 1 0 1 1 0 1 0 2 2 1 0 0 3 3 0 0 1 4 4 1 0 0 5 5 0 1 0 # 2 mnames = ['movie_id', 'title', 'genres'] movies = pd.read_table('data/movies.dat', sep='::', header=None,names=mnames) movies[:10] C:\ProgramData\Anaconda3\lib\site-packages\ipykernel_launcher.py:4: ParserWarning: Falling back to the 'python' engine because the 'c' engine does not support regex separators (separators > 1 char and different from '\s+' are interpreted as regex); you can avoid this warning by specifying engine='python'.after removing the cwd from sys.path. movie_idtitlegenres0123456789
1Toy Story (1995)Animation|Children's|Comedy
2Jumanji (1995)Adventure|Children's|Fantasy
3Grumpier Old Men (1995)Comedy|Romance
4Waiting to Exhale (1995)Comedy|Drama
5Father of the Bride Part II (1995)Comedy
6Heat (1995)Action|Crime|Thriller
7Sabrina (1995)Comedy|Romance
8Tom and Huck (1995)Adventure|Children's
9Sudden Death (1995)Action
10GoldenEye (1995)Action|Adventure|Thriller
genre_iter = (set(x.split('|')) for x in movies.genres) genres = sorted(set.union(*genre_iter)) print(genres) ['Action', 'Adventure', 'Animation', "Children's", 'Comedy', 'Crime', 'Documentary', 'Drama', 'Fantasy', 'Film-Noir', 'Horror', 'Musical', 'Mystery', 'Romance', 'Sci-Fi', 'Thriller', 'War', 'Western'] dummies = DataFrame(np.zeros((len(movies), len(genres))), columns=genres) print(dummies) Action Adventure Animation Children's Comedy Crime Documentary \ 0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3 0.0 0.0 0.0 0.0 0.0 0.0 0.0 4 0.0 0.0 0.0 0.0 0.0 0.0 0.0 5 0.0 0.0 0.0 0.0 0.0 0.0 0.0 6 0.0 0.0 0.0 0.0 0.0 0.0 0.0 7 0.0 0.0 0.0 0.0 0.0 0.0 0.0 8 0.0 0.0 0.0 0.0 0.0 0.0 0.0 9 0.0 0.0 0.0 0.0 0.0 0.0 0.0 10 0.0 0.0 0.0 0.0 0.0 0.0 0.0 11 0.0 0.0 0.0 0.0 0.0 0.0 0.0 12 0.0 0.0 0.0 0.0 0.0 0.0 0.0 13 0.0 0.0 0.0 0.0 0.0 0.0 0.0 14 0.0 0.0 0.0 0.0 0.0 0.0 0.0 15 0.0 0.0 0.0 0.0 0.0 0.0 0.0 16 0.0 0.0 0.0 0.0 0.0 0.0 0.0 17 0.0 0.0 0.0 0.0 0.0 0.0 0.0 18 0.0 0.0 0.0 0.0 0.0 0.0 0.0 19 0.0 0.0 0.0 0.0 0.0 0.0 0.0 20 0.0 0.0 0.0 0.0 0.0 0.0 0.0 21 0.0 0.0 0.0 0.0 0.0 0.0 0.0 22 0.0 0.0 0.0 0.0 0.0 0.0 0.0 23 0.0 0.0 0.0 0.0 0.0 0.0 0.0 24 0.0 0.0 0.0 0.0 0.0 0.0 0.0 25 0.0 0.0 0.0 0.0 0.0 0.0 0.0 26 0.0 0.0 0.0 0.0 0.0 0.0 0.0 27 0.0 0.0 0.0 0.0 0.0 0.0 0.0 28 0.0 0.0 0.0 0.0 0.0 0.0 0.0 29 0.0 0.0 0.0 0.0 0.0 0.0 0.0 ... ... ... ... ... ... ... ... 3853 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3854 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3855 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3856 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3857 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3858 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3859 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3860 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3861 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3862 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3863 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3864 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3865 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3866 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3867 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3868 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3869 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3870 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3871 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3872 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3873 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3874 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3875 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3876 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3877 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3878 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3879 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3880 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3881 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3882 0.0 0.0 0.0 0.0 0.0 0.0 0.0 Drama Fantasy Film-Noir Horror Musical Mystery Romance Sci-Fi \ 0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 4 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 5 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 6 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 7 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 8 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 9 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 10 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 11 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 12 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 13 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 14 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 15 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 16 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 17 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 18 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 19 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 20 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 21 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 22 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 23 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 24 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 25 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 26 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 27 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 28 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 29 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 ... ... ... ... ... ... ... ... ... 3853 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3854 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3855 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3856 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3857 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3858 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3859 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3860 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3861 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3862 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3863 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3864 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3865 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3866 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3867 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3868 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3869 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3870 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3871 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3872 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3873 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3874 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3875 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3876 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3877 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3878 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3879 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3880 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3881 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3882 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 Thriller War Western 0 0.0 0.0 0.0 1 0.0 0.0 0.0 2 0.0 0.0 0.0 3 0.0 0.0 0.0 4 0.0 0.0 0.0 5 0.0 0.0 0.0 6 0.0 0.0 0.0 7 0.0 0.0 0.0 8 0.0 0.0 0.0 9 0.0 0.0 0.0 10 0.0 0.0 0.0 11 0.0 0.0 0.0 12 0.0 0.0 0.0 13 0.0 0.0 0.0 14 0.0 0.0 0.0 15 0.0 0.0 0.0 16 0.0 0.0 0.0 17 0.0 0.0 0.0 18 0.0 0.0 0.0 19 0.0 0.0 0.0 20 0.0 0.0 0.0 21 0.0 0.0 0.0 22 0.0 0.0 0.0 23 0.0 0.0 0.0 24 0.0 0.0 0.0 25 0.0 0.0 0.0 26 0.0 0.0 0.0 27 0.0 0.0 0.0 28 0.0 0.0 0.0 29 0.0 0.0 0.0 ... ... ... ... 3853 0.0 0.0 0.0 3854 0.0 0.0 0.0 3855 0.0 0.0 0.0 3856 0.0 0.0 0.0 3857 0.0 0.0 0.0 3858 0.0 0.0 0.0 3859 0.0 0.0 0.0 3860 0.0 0.0 0.0 3861 0.0 0.0 0.0 3862 0.0 0.0 0.0 3863 0.0 0.0 0.0 3864 0.0 0.0 0.0 3865 0.0 0.0 0.0 3866 0.0 0.0 0.0 3867 0.0 0.0 0.0 3868 0.0 0.0 0.0 3869 0.0 0.0 0.0 3870 0.0 0.0 0.0 3871 0.0 0.0 0.0 3872 0.0 0.0 0.0 3873 0.0 0.0 0.0 3874 0.0 0.0 0.0 3875 0.0 0.0 0.0 3876 0.0 0.0 0.0 3877 0.0 0.0 0.0 3878 0.0 0.0 0.0 3879 0.0 0.0 0.0 3880 0.0 0.0 0.0 3881 0.0 0.0 0.0 3882 0.0 0.0 0.0 [3883 rows x 18 columns] for i, gen in enumerate(movies.genres):dummies.loc[i, gen.split('|')] = 1movies_windic = movies.join(dummies.add_prefix('Genre_')) print(movies_windic.loc[0]) movie_id 1 title Toy Story (1995) genres Animation|Children's|Comedy Genre_Action 0 Genre_Adventure 0 Genre_Animation 1 Genre_Children's 1 Genre_Comedy 1 Genre_Crime 0 Genre_Documentary 0 Genre_Drama 0 Genre_Fantasy 0 Genre_Film-Noir 0 Genre_Horror 0 Genre_Musical 0 Genre_Mystery 0 Genre_Romance 0 Genre_Sci-Fi 0 Genre_Thriller 0 Genre_War 0 Genre_Western 0 Name: 0, dtype: object # 3 np.random.seed(12345) values = np.random.rand(10) print(values) [ 0.92961609 0.31637555 0.18391881 0.20456028 0.56772503 0.59554470.96451452 0.6531771 0.74890664 0.65356987] bins = [0, 0.2, 0.4, 0.6, 0.8, 1] pd.get_dummies(pd.cut(values, bins)) (0.0, 0.2](0.2, 0.4](0.4, 0.6](0.6, 0.8](0.8, 1.0]0123456789
00001
01000
10000
01000
00100
00100
00001
00010
00010
00010

屬性構造

# 參數初始化 inputfile = 'data/electricity_data.xls' # 供入供出電量數據 outputfile = 'data/electricity_data.xls' # 屬性構造后數據文件data = pd.read_excel(inputfile) # 讀入數據 data[u'線損率'] = (data[u'供入電量'] - data[u'供出電量']) / data[u'供入電量']data.to_excel(outputfile, index=False) # 保存結果

字符串對象方法

val = 'a,b, guido' val.split(',') ['a', 'b', ' guido'] pieces = [x.strip() for x in val.split(',')] print(pieces) ['a', 'b', 'guido'] first, second, third = pieces first + '::' + second + '::' + third 'a::b::guido' '::'.join(pieces) 'a::b::guido' 'guido' in val True val.index(',') 1 val.find(':') -1 val.index(':') ---------------------------------------------------------------------------ValueError Traceback (most recent call last)<ipython-input-138-280f8b2856ce> in <module>() ----> 1 val.index(':')ValueError: substring not found val.count('a') 1 val.replace(',', '::') 'a::b:: guido' val.replace(',', '') 'ab guido'

正則表達式

# 1 import retext = "foo bar\t baz \tqux" re.split('\s+', text) ['foo', 'bar', 'baz', 'qux'] regex = re.compile('\s+') regex.split(text) ['foo', 'bar', 'baz', 'qux'] regex.findall(text) [' ', '\t ', ' \t'] # 2 text = """Dave dave@google.com Steve steve@gmail.com Rob rob@gmail.com Ryan ryan@yahoo.com """ pattern = r'[A-Z0-9._%+-]+@[A-Z0-9.-]+\.[A-Z]{2,4}'# re.IGNORECASE 的作用是使正則表達式對大小寫不敏感 regex = re.compile(pattern, flags=re.IGNORECASE)regex.findall(text) ['dave@google.com', 'steve@gmail.com', 'rob@gmail.com', 'ryan@yahoo.com'] m = regex.search(text) print(m) <_sre.SRE_Match object; span=(5, 20), match='dave@google.com'> text[m.start():m.end()] 'dave@google.com' print(regex.match(text)) None print(regex.sub('REDACTED', text)) Dave REDACTED Steve REDACTED Rob REDACTED Ryan REDACTED # 3 pattern = r'([A-Z0-9._%+-]+)@([A-Z0-9.-]+)\.([A-Z]{2,4})' regex = re.compile(pattern, flags=re.IGNORECASE)m = regex.match('wesm@bright.net') m.groups() ('wesm', 'bright', 'net') regex.findall(text) [('dave', 'google', 'com'),('steve', 'gmail', 'com'),('rob', 'gmail', 'com'),('ryan', 'yahoo', 'com')] print(regex.sub(r'Username: \1, Domain: \2, Suffix: \3', text)) Dave Username: dave, Domain: google, Suffix: com Steve Username: steve, Domain: gmail, Suffix: com Rob Username: rob, Domain: gmail, Suffix: com Ryan Username: ryan, Domain: yahoo, Suffix: com # 4 regex = re.compile(r"""(?P<username>[A-Z0-9._%+-]+)@(?P<domain>[A-Z0-9.-]+)\.(?P<suffix>[A-Z]{2,4})""", flags=re.IGNORECASE | re.VERBOSE)m = regex.match('wesm@bright.net') m.groupdict() {'domain': 'bright', 'suffix': 'net', 'username': 'wesm'}

pandas中矢量化的字符串函數

data = {'Dave': 'dave@google.com', 'Steve': 'steve@gmail.com','Rob': 'rob@gmail.com', 'Wes': np.nan} data = Series(data) print(data) Dave dave@google.com Rob rob@gmail.com Steve steve@gmail.com Wes NaN dtype: object data.isnull() Dave False Rob False Steve False Wes True dtype: bool data.str.contains('gmail') Dave False Rob True Steve True Wes NaN dtype: object data.str.findall(pattern, flags=re.IGNORECASE) Dave [(dave, google, com)] Rob [(rob, gmail, com)] Steve [(steve, gmail, com)] Wes NaN dtype: object matches = data.str.match(pattern, flags=re.IGNORECASE) print(matches) Dave True Rob True Steve True Wes NaN dtype: object matches.str.get(1) Dave NaN Rob NaN Steve NaN Wes NaN dtype: float64 matches.str[0] Dave NaN Rob NaN Steve NaN Wes NaN dtype: float64 data.str[:5] Dave dave@ Rob rob@g Steve steve Wes NaN dtype: object

示例:USDA食品數據庫

''' {"id": 21441,"description": "KENTUCKY FRIED CHICKEN, Fried Chicken, EXTRA CRISPY, Wing, meat and skin with breading","tags": ["KFC"],"manufacturer": "Kentucky Fried Chicken","group": "Fast Foods","portions": [{"amount": 1,"unit": "wing, with skin","grams": 68.0},...],"nutrients": [{"value": 20.8,"units": "g","description": "Protein","group": "Composition"},...] } '''import jsondb = json.load(open('data/foods-2011-10-03.json')) len(db) 6636 db[0].keys() dict_keys(['id', 'description', 'tags', 'manufacturer', 'group', 'portions', 'nutrients']) db[0]['nutrients'][0] {'description': 'Protein','group': 'Composition','units': 'g','value': 25.18} nutrients = DataFrame(db[0]['nutrients']) nutrients[:7] descriptiongroupunitsvalue0123456
ProteinCompositiong25.18
Total lipid (fat)Compositiong29.20
Carbohydrate, by differenceCompositiong3.06
AshOtherg3.28
EnergyEnergykcal376.00
WaterCompositiong39.28
EnergyEnergykJ1573.00
info_keys = ['description', 'group', 'id', 'manufacturer'] info = DataFrame(db, columns=info_keys)info[:5] descriptiongroupidmanufacturer01234
Cheese, carawayDairy and Egg Products1008
Cheese, cheddarDairy and Egg Products1009
Cheese, edamDairy and Egg Products1018
Cheese, fetaDairy and Egg Products1019
Cheese, mozzarella, part skim milkDairy and Egg Products1028
pd.value_counts(info.group)[:10] Vegetables and Vegetable Products 812 Beef Products 618 Baked Products 496 Breakfast Cereals 403 Fast Foods 365 Legumes and Legume Products 365 Lamb, Veal, and Game Products 345 Sweets 341 Fruits and Fruit Juices 328 Pork Products 328 Name: group, dtype: int64 nutrients = []for rec in db:fnuts = DataFrame(rec['nutrients'])fnuts['id'] = rec['id']nutrients.append(fnuts)nutrients = pd.concat(nutrients, ignore_index=True)print(nutrients) description group units value \ 0 Protein Composition g 25.180 1 Total lipid (fat) Composition g 29.200 2 Carbohydrate, by difference Composition g 3.060 3 Ash Other g 3.280 4 Energy Energy kcal 376.000 5 Water Composition g 39.280 6 Energy Energy kJ 1573.000 7 Fiber, total dietary Composition g 0.000 8 Calcium, Ca Elements mg 673.000 9 Iron, Fe Elements mg 0.640 10 Magnesium, Mg Elements mg 22.000 11 Phosphorus, P Elements mg 490.000 12 Potassium, K Elements mg 93.000 13 Sodium, Na Elements mg 690.000 14 Zinc, Zn Elements mg 2.940 15 Copper, Cu Elements mg 0.024 16 Manganese, Mn Elements mg 0.021 17 Selenium, Se Elements mcg 14.500 18 Vitamin A, IU Vitamins IU 1054.000 19 Retinol Vitamins mcg 262.000 20 Vitamin A, RAE Vitamins mcg_RAE 271.000 21 Vitamin C, total ascorbic acid Vitamins mg 0.000 22 Thiamin Vitamins mg 0.031 23 Riboflavin Vitamins mg 0.450 24 Niacin Vitamins mg 0.180 25 Pantothenic acid Vitamins mg 0.190 26 Vitamin B-6 Vitamins mg 0.074 27 Folate, total Vitamins mcg 18.000 28 Vitamin B-12 Vitamins mcg 0.270 29 Folic acid Vitamins mcg 0.000 ... ... ... ... ... 389325 Selenium, Se Elements mcg 1.100 389326 Vitamin A, IU Vitamins IU 5.000 389327 Retinol Vitamins mcg 0.000 389328 Vitamin A, RAE Vitamins mcg_RAE 0.000 389329 Carotene, beta Vitamins mcg 2.000 389330 Carotene, alpha Vitamins mcg 2.000 389331 Vitamin E (alpha-tocopherol) Vitamins mg 0.250 389332 Vitamin D Vitamins IU 0.000 389333 Vitamin D (D2 + D3) Vitamins mcg 0.000 389334 Cryptoxanthin, beta Vitamins mcg 0.000 389335 Lycopene Vitamins mcg 0.000 389336 Lutein + zeaxanthin Vitamins mcg 20.000 389337 Vitamin C, total ascorbic acid Vitamins mg 21.900 389338 Thiamin Vitamins mg 0.020 389339 Riboflavin Vitamins mg 0.060 389340 Niacin Vitamins mg 0.540 389341 Vitamin B-6 Vitamins mg 0.260 389342 Folate, total Vitamins mcg 17.000 389343 Vitamin B-12 Vitamins mcg 0.000 389344 Choline, total Vitamins mg 4.100 389345 Vitamin K (phylloquinone) Vitamins mcg 0.500 389346 Folic acid Vitamins mcg 0.000 389347 Folate, food Vitamins mcg 17.000 389348 Folate, DFE Vitamins mcg_DFE 17.000 389349 Vitamin E, added Vitamins mg 0.000 389350 Vitamin B-12, added Vitamins mcg 0.000 389351 Cholesterol Other mg 0.000 389352 Fatty acids, total saturated Other g 0.072 389353 Fatty acids, total monounsaturated Other g 0.028 389354 Fatty acids, total polyunsaturated Other g 0.041 id 0 1008 1 1008 2 1008 3 1008 4 1008 5 1008 6 1008 7 1008 8 1008 9 1008 10 1008 11 1008 12 1008 13 1008 14 1008 15 1008 16 1008 17 1008 18 1008 19 1008 20 1008 21 1008 22 1008 23 1008 24 1008 25 1008 26 1008 27 1008 28 1008 29 1008 ... ... 389325 43546 389326 43546 389327 43546 389328 43546 389329 43546 389330 43546 389331 43546 389332 43546 389333 43546 389334 43546 389335 43546 389336 43546 389337 43546 389338 43546 389339 43546 389340 43546 389341 43546 389342 43546 389343 43546 389344 43546 389345 43546 389346 43546 389347 43546 389348 43546 389349 43546 389350 43546 389351 43546 389352 43546 389353 43546 389354 43546 [389355 rows x 5 columns] # 數據中有重復的數據 nutrients.duplicated().sum() 14179 # 去重 nutrients = nutrients.drop_duplicates() col_mapping = {'description': 'food','group': 'fgroup'} info = info.rename(columns=col_mapping, copy=False) print(info) food \ 0 Cheese, caraway 1 Cheese, cheddar 2 Cheese, edam 3 Cheese, feta 4 Cheese, mozzarella, part skim milk 5 Cheese, mozzarella, part skim milk, low moisture 6 Cheese, romano 7 Cheese, roquefort 8 Cheese spread, pasteurized process, american, ... 9 Cream, fluid, half and half 10 Sour dressing, non-butterfat, cultured, filled... 11 Milk, filled, fluid, with blend of hydrogenate... 12 Cream substitute, liquid, with lauric acid oil... 13 Cream substitute, powdered 14 Milk, producer, fluid, 3.7% milkfat 15 Milk, reduced fat, fluid, 2% milkfat, with add... 16 Milk, reduced fat, fluid, 2% milkfat, with add... 17 Milk, reduced fat, fluid, 2% milkfat, protein ... 18 Milk, lowfat, fluid, 1% milkfat, with added vi... 19 Milk, lowfat, fluid, 1% milkfat, with added no... 20 Milk, lowfat, fluid, 1% milkfat, protein forti... 21 Milk, nonfat, fluid, with added vitamin A and ... 22 Milk, nonfat, fluid, with added nonfat milk so... 23 Milk, nonfat, fluid, protein fortified, with a... 24 Milk, buttermilk, fluid, cultured, lowfat 25 Milk, low sodium, fluid 26 Milk, dry, whole, with added vitamin D 27 Milk, dry, nonfat, regular, without added vita... 28 Milk, dry, nonfat, instant, with added vitamin... 29 Milk, dry, nonfat, calcium reduced ... ... 6606 Beef, tenderloin, steak, separable lean only, ... 6607 Beef, top sirloin, steak, separable lean only,... 6608 Beef, short loin, top loin, steak, separable l... 6609 Beef, chuck, arm pot roast, separable lean onl... 6610 Beef, brisket, flat half, separable lean only,... 6611 Beef, chuck, arm pot roast, separable lean onl... 6612 Beef, brisket, flat half, separable lean only,... 6613 Beef, round, eye of round, roast, separable le... 6614 Beef, round, top round, steak, separable lean ... 6615 Beef, round, bottom round, roast, separable le... 6616 Beef, rib, small end (ribs 10-12), separable l... 6617 CAMPBELL Soup Company, CAMPBELL'S Red and Whit... 6618 CAMPBELL Soup Company, CAMPBELL's Red and Whit... 6619 CAMPBELL Soup Company, CAMPBELL'S SELECT Soups... 6620 CAMPBELL Soup Company, CAMPBELL'S SOUP AT HAND... 6621 CAMPBELL Soup Company, CAMPBELL'S SOUP AT HAND... 6622 CAMPBELL Soup Company, CAMPBELL'S SELECT Gold ... 6623 CAMPBELL Soup Company, CAMPBELL'S SELECT Gold ... 6624 CAMPBELL Soup Company, CAMPBELL'S SELECT Gold ... 6625 CAMPBELL Soup Company, CAMPBELL'S Red and Whit... 6626 CAMPBELL Soup Company, V8 Vegetable Juice, Ess... 6627 CAMPBELL Soup Company, V8 Vegetable Juice, Spi... 6628 CAMPBELL Soup Company, PACE, Jalapenos Nacho S... 6629 CAMPBELL Soup Company, V8 60% Vegetable Juice,... 6630 CAMPBELL Soup Company, V8 Vegetable Juice, Low... 6631 Bologna, beef, low fat 6632 Turkey and pork sausage, fresh, bulk, patty or... 6633 Babyfood, juice, pear 6634 Babyfood, dessert, banana yogurt, strained 6635 Babyfood, banana no tapioca, strained fgroup id manufacturer 0 Dairy and Egg Products 1008 1 Dairy and Egg Products 1009 2 Dairy and Egg Products 1018 3 Dairy and Egg Products 1019 4 Dairy and Egg Products 1028 5 Dairy and Egg Products 1029 6 Dairy and Egg Products 1038 7 Dairy and Egg Products 1039 8 Dairy and Egg Products 1048 9 Dairy and Egg Products 1049 10 Dairy and Egg Products 1058 11 Dairy and Egg Products 1059 12 Dairy and Egg Products 1068 13 Dairy and Egg Products 1069 14 Dairy and Egg Products 1078 15 Dairy and Egg Products 1079 None 16 Dairy and Egg Products 1080 17 Dairy and Egg Products 1081 18 Dairy and Egg Products 1082 19 Dairy and Egg Products 1083 20 Dairy and Egg Products 1084 21 Dairy and Egg Products 1085 22 Dairy and Egg Products 1086 23 Dairy and Egg Products 1087 24 Dairy and Egg Products 1088 25 Dairy and Egg Products 1089 26 Dairy and Egg Products 1090 27 Dairy and Egg Products 1091 28 Dairy and Egg Products 1092 29 Dairy and Egg Products 1093 ... ... ... ... 6606 Beef Products 23628 6607 Beef Products 23629 6608 Beef Products 23630 6609 Beef Products 23631 6610 Beef Products 23632 6611 Beef Products 23633 6612 Beef Products 23634 6613 Beef Products 23635 6614 Beef Products 23636 6615 Beef Products 23637 6616 Beef Products 23638 6617 Soups, Sauces, and Gravies 27015 Campbell Soup Co. 6618 Soups, Sauces, and Gravies 27016 Campbell Soup Co. 6619 Soups, Sauces, and Gravies 27021 Campbell Soup Co. 6620 Soups, Sauces, and Gravies 27022 Campbell Soup Co. 6621 Soups, Sauces, and Gravies 27023 Campbell Soup Co. 6622 Soups, Sauces, and Gravies 27024 Campbell Soup Co. 6623 Soups, Sauces, and Gravies 27025 Campbell Soup Co. 6624 Soups, Sauces, and Gravies 27026 Campbell Soup Co. 6625 Soups, Sauces, and Gravies 27032 Campbell Soup Co. 6626 Vegetables and Vegetable Products 31010 Campbell Soup Co. 6627 Vegetables and Vegetable Products 31013 Campbell Soup Co. 6628 Vegetables and Vegetable Products 31014 Campbell Soup Co. 6629 Vegetables and Vegetable Products 31016 Campbell Soup Co. 6630 Vegetables and Vegetable Products 31017 Campbell Soup Co. 6631 Sausages and Luncheon Meats 42161 6632 Sausages and Luncheon Meats 42173 6633 Baby Foods 43408 None 6634 Baby Foods 43539 None 6635 Baby Foods 43546 None [6636 rows x 4 columns] col_mapping = {'description': 'nutrient','group': 'nutgroup'} nutrients = nutrients.rename(columns=col_mapping, copy=False) print(nutrients) nutrient nutgroup units value \ 0 Protein Composition g 25.180 1 Total lipid (fat) Composition g 29.200 2 Carbohydrate, by difference Composition g 3.060 3 Ash Other g 3.280 4 Energy Energy kcal 376.000 5 Water Composition g 39.280 6 Energy Energy kJ 1573.000 7 Fiber, total dietary Composition g 0.000 8 Calcium, Ca Elements mg 673.000 9 Iron, Fe Elements mg 0.640 10 Magnesium, Mg Elements mg 22.000 11 Phosphorus, P Elements mg 490.000 12 Potassium, K Elements mg 93.000 13 Sodium, Na Elements mg 690.000 14 Zinc, Zn Elements mg 2.940 15 Copper, Cu Elements mg 0.024 16 Manganese, Mn Elements mg 0.021 17 Selenium, Se Elements mcg 14.500 18 Vitamin A, IU Vitamins IU 1054.000 19 Retinol Vitamins mcg 262.000 20 Vitamin A, RAE Vitamins mcg_RAE 271.000 21 Vitamin C, total ascorbic acid Vitamins mg 0.000 22 Thiamin Vitamins mg 0.031 23 Riboflavin Vitamins mg 0.450 24 Niacin Vitamins mg 0.180 25 Pantothenic acid Vitamins mg 0.190 26 Vitamin B-6 Vitamins mg 0.074 27 Folate, total Vitamins mcg 18.000 28 Vitamin B-12 Vitamins mcg 0.270 29 Folic acid Vitamins mcg 0.000 ... ... ... ... ... 389325 Selenium, Se Elements mcg 1.100 389326 Vitamin A, IU Vitamins IU 5.000 389327 Retinol Vitamins mcg 0.000 389328 Vitamin A, RAE Vitamins mcg_RAE 0.000 389329 Carotene, beta Vitamins mcg 2.000 389330 Carotene, alpha Vitamins mcg 2.000 389331 Vitamin E (alpha-tocopherol) Vitamins mg 0.250 389332 Vitamin D Vitamins IU 0.000 389333 Vitamin D (D2 + D3) Vitamins mcg 0.000 389334 Cryptoxanthin, beta Vitamins mcg 0.000 389335 Lycopene Vitamins mcg 0.000 389336 Lutein + zeaxanthin Vitamins mcg 20.000 389337 Vitamin C, total ascorbic acid Vitamins mg 21.900 389338 Thiamin Vitamins mg 0.020 389339 Riboflavin Vitamins mg 0.060 389340 Niacin Vitamins mg 0.540 389341 Vitamin B-6 Vitamins mg 0.260 389342 Folate, total Vitamins mcg 17.000 389343 Vitamin B-12 Vitamins mcg 0.000 389344 Choline, total Vitamins mg 4.100 389345 Vitamin K (phylloquinone) Vitamins mcg 0.500 389346 Folic acid Vitamins mcg 0.000 389347 Folate, food Vitamins mcg 17.000 389348 Folate, DFE Vitamins mcg_DFE 17.000 389349 Vitamin E, added Vitamins mg 0.000 389350 Vitamin B-12, added Vitamins mcg 0.000 389351 Cholesterol Other mg 0.000 389352 Fatty acids, total saturated Other g 0.072 389353 Fatty acids, total monounsaturated Other g 0.028 389354 Fatty acids, total polyunsaturated Other g 0.041 id 0 1008 1 1008 2 1008 3 1008 4 1008 5 1008 6 1008 7 1008 8 1008 9 1008 10 1008 11 1008 12 1008 13 1008 14 1008 15 1008 16 1008 17 1008 18 1008 19 1008 20 1008 21 1008 22 1008 23 1008 24 1008 25 1008 26 1008 27 1008 28 1008 29 1008 ... ... 389325 43546 389326 43546 389327 43546 389328 43546 389329 43546 389330 43546 389331 43546 389332 43546 389333 43546 389334 43546 389335 43546 389336 43546 389337 43546 389338 43546 389339 43546 389340 43546 389341 43546 389342 43546 389343 43546 389344 43546 389345 43546 389346 43546 389347 43546 389348 43546 389349 43546 389350 43546 389351 43546 389352 43546 389353 43546 389354 43546 [375176 rows x 5 columns] ndata = pd.merge(nutrients, info, on='id', how='outer') print(ndata) nutrient nutgroup units value \ 0 Protein Composition g 25.180 1 Total lipid (fat) Composition g 29.200 2 Carbohydrate, by difference Composition g 3.060 3 Ash Other g 3.280 4 Energy Energy kcal 376.000 5 Water Composition g 39.280 6 Energy Energy kJ 1573.000 7 Fiber, total dietary Composition g 0.000 8 Calcium, Ca Elements mg 673.000 9 Iron, Fe Elements mg 0.640 10 Magnesium, Mg Elements mg 22.000 11 Phosphorus, P Elements mg 490.000 12 Potassium, K Elements mg 93.000 13 Sodium, Na Elements mg 690.000 14 Zinc, Zn Elements mg 2.940 15 Copper, Cu Elements mg 0.024 16 Manganese, Mn Elements mg 0.021 17 Selenium, Se Elements mcg 14.500 18 Vitamin A, IU Vitamins IU 1054.000 19 Retinol Vitamins mcg 262.000 20 Vitamin A, RAE Vitamins mcg_RAE 271.000 21 Vitamin C, total ascorbic acid Vitamins mg 0.000 22 Thiamin Vitamins mg 0.031 23 Riboflavin Vitamins mg 0.450 24 Niacin Vitamins mg 0.180 25 Pantothenic acid Vitamins mg 0.190 26 Vitamin B-6 Vitamins mg 0.074 27 Folate, total Vitamins mcg 18.000 28 Vitamin B-12 Vitamins mcg 0.270 29 Folic acid Vitamins mcg 0.000 ... ... ... ... ... 375146 Selenium, Se Elements mcg 1.100 375147 Vitamin A, IU Vitamins IU 5.000 375148 Retinol Vitamins mcg 0.000 375149 Vitamin A, RAE Vitamins mcg_RAE 0.000 375150 Carotene, beta Vitamins mcg 2.000 375151 Carotene, alpha Vitamins mcg 2.000 375152 Vitamin E (alpha-tocopherol) Vitamins mg 0.250 375153 Vitamin D Vitamins IU 0.000 375154 Vitamin D (D2 + D3) Vitamins mcg 0.000 375155 Cryptoxanthin, beta Vitamins mcg 0.000 375156 Lycopene Vitamins mcg 0.000 375157 Lutein + zeaxanthin Vitamins mcg 20.000 375158 Vitamin C, total ascorbic acid Vitamins mg 21.900 375159 Thiamin Vitamins mg 0.020 375160 Riboflavin Vitamins mg 0.060 375161 Niacin Vitamins mg 0.540 375162 Vitamin B-6 Vitamins mg 0.260 375163 Folate, total Vitamins mcg 17.000 375164 Vitamin B-12 Vitamins mcg 0.000 375165 Choline, total Vitamins mg 4.100 375166 Vitamin K (phylloquinone) Vitamins mcg 0.500 375167 Folic acid Vitamins mcg 0.000 375168 Folate, food Vitamins mcg 17.000 375169 Folate, DFE Vitamins mcg_DFE 17.000 375170 Vitamin E, added Vitamins mg 0.000 375171 Vitamin B-12, added Vitamins mcg 0.000 375172 Cholesterol Other mg 0.000 375173 Fatty acids, total saturated Other g 0.072 375174 Fatty acids, total monounsaturated Other g 0.028 375175 Fatty acids, total polyunsaturated Other g 0.041 id food fgroup \ 0 1008 Cheese, caraway Dairy and Egg Products 1 1008 Cheese, caraway Dairy and Egg Products 2 1008 Cheese, caraway Dairy and Egg Products 3 1008 Cheese, caraway Dairy and Egg Products 4 1008 Cheese, caraway Dairy and Egg Products 5 1008 Cheese, caraway Dairy and Egg Products 6 1008 Cheese, caraway Dairy and Egg Products 7 1008 Cheese, caraway Dairy and Egg Products 8 1008 Cheese, caraway Dairy and Egg Products 9 1008 Cheese, caraway Dairy and Egg Products 10 1008 Cheese, caraway Dairy and Egg Products 11 1008 Cheese, caraway Dairy and Egg Products 12 1008 Cheese, caraway Dairy and Egg Products 13 1008 Cheese, caraway Dairy and Egg Products 14 1008 Cheese, caraway Dairy and Egg Products 15 1008 Cheese, caraway Dairy and Egg Products 16 1008 Cheese, caraway Dairy and Egg Products 17 1008 Cheese, caraway Dairy and Egg Products 18 1008 Cheese, caraway Dairy and Egg Products 19 1008 Cheese, caraway Dairy and Egg Products 20 1008 Cheese, caraway Dairy and Egg Products 21 1008 Cheese, caraway Dairy and Egg Products 22 1008 Cheese, caraway Dairy and Egg Products 23 1008 Cheese, caraway Dairy and Egg Products 24 1008 Cheese, caraway Dairy and Egg Products 25 1008 Cheese, caraway Dairy and Egg Products 26 1008 Cheese, caraway Dairy and Egg Products 27 1008 Cheese, caraway Dairy and Egg Products 28 1008 Cheese, caraway Dairy and Egg Products 29 1008 Cheese, caraway Dairy and Egg Products ... ... ... ... 375146 43546 Babyfood, banana no tapioca, strained Baby Foods 375147 43546 Babyfood, banana no tapioca, strained Baby Foods 375148 43546 Babyfood, banana no tapioca, strained Baby Foods 375149 43546 Babyfood, banana no tapioca, strained Baby Foods 375150 43546 Babyfood, banana no tapioca, strained Baby Foods 375151 43546 Babyfood, banana no tapioca, strained Baby Foods 375152 43546 Babyfood, banana no tapioca, strained Baby Foods 375153 43546 Babyfood, banana no tapioca, strained Baby Foods 375154 43546 Babyfood, banana no tapioca, strained Baby Foods 375155 43546 Babyfood, banana no tapioca, strained Baby Foods 375156 43546 Babyfood, banana no tapioca, strained Baby Foods 375157 43546 Babyfood, banana no tapioca, strained Baby Foods 375158 43546 Babyfood, banana no tapioca, strained Baby Foods 375159 43546 Babyfood, banana no tapioca, strained Baby Foods 375160 43546 Babyfood, banana no tapioca, strained Baby Foods 375161 43546 Babyfood, banana no tapioca, strained Baby Foods 375162 43546 Babyfood, banana no tapioca, strained Baby Foods 375163 43546 Babyfood, banana no tapioca, strained Baby Foods 375164 43546 Babyfood, banana no tapioca, strained Baby Foods 375165 43546 Babyfood, banana no tapioca, strained Baby Foods 375166 43546 Babyfood, banana no tapioca, strained Baby Foods 375167 43546 Babyfood, banana no tapioca, strained Baby Foods 375168 43546 Babyfood, banana no tapioca, strained Baby Foods 375169 43546 Babyfood, banana no tapioca, strained Baby Foods 375170 43546 Babyfood, banana no tapioca, strained Baby Foods 375171 43546 Babyfood, banana no tapioca, strained Baby Foods 375172 43546 Babyfood, banana no tapioca, strained Baby Foods 375173 43546 Babyfood, banana no tapioca, strained Baby Foods 375174 43546 Babyfood, banana no tapioca, strained Baby Foods 375175 43546 Babyfood, banana no tapioca, strained Baby Foods manufacturer 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 ... ... 375146 None 375147 None 375148 None 375149 None 375150 None 375151 None 375152 None 375153 None 375154 None 375155 None 375156 None 375157 None 375158 None 375159 None 375160 None 375161 None 375162 None 375163 None 375164 None 375165 None 375166 None 375167 None 375168 None 375169 None 375170 None 375171 None 375172 None 375173 None 375174 None 375175 None [375176 rows x 8 columns] ndata.loc[30000] nutrient Glycine nutgroup Amino Acids units g value 0.04 id 6158 food Soup, tomato bisque, canned, condensed fgroup Soups, Sauces, and Gravies manufacturer Name: 30000, dtype: object result = ndata.groupby(['nutrient', 'fgroup'])['value'].quantile(0.5) print(result) nutrient fgroup Adjusted Protein Sweets 12.9000Vegetables and Vegetable Products 2.1800 Alanine Baby Foods 0.0850Baked Products 0.2480Beef Products 1.5500Beverages 0.0030Breakfast Cereals 0.3110Cereal Grains and Pasta 0.3730Dairy and Egg Products 0.2710Ethnic Foods 1.2900Fast Foods 0.5140Fats and Oils 0.0000Finfish and Shellfish Products 1.2180Fruits and Fruit Juices 0.0270Lamb, Veal, and Game Products 1.4080Legumes and Legume Products 0.4100Meals, Entrees, and Sidedishes 0.3270Nut and Seed Products 0.7345Pork Products 1.3070Poultry Products 1.3940Restaurant Foods 0.4650Sausages and Luncheon Meats 0.9420Snacks 0.4335Soups, Sauces, and Gravies 0.0650Spices and Herbs 0.5550Sweets 0.1020Vegetables and Vegetable Products 0.0840 Alcohol, ethyl Baby Foods 0.0000Baked Products 0.0000Beef Products 0.0000... Water Snacks 3.5200Soups, Sauces, and Gravies 85.9000Spices and Herbs 43.6700Sweets 9.0500Vegetables and Vegetable Products 89.1950 Zinc, Zn Baby Foods 0.5900Baked Products 0.6600Beef Products 5.3900Beverages 0.0400Breakfast Cereals 2.8850Cereal Grains and Pasta 1.0900Dairy and Egg Products 1.3900Ethnic Foods 1.0450Fast Foods 1.2500Fats and Oils 0.0200Finfish and Shellfish Products 0.6700Fruits and Fruit Juices 0.1000Lamb, Veal, and Game Products 3.9400Legumes and Legume Products 1.1400Meals, Entrees, and Sidedishes 0.6300Nut and Seed Products 3.2900Pork Products 2.3200Poultry Products 2.5000Restaurant Foods 0.8000Sausages and Luncheon Meats 2.1300Snacks 1.4700Soups, Sauces, and Gravies 0.2000Spices and Herbs 2.7500Sweets 0.3600Vegetables and Vegetable Products 0.3300 Name: value, Length: 2246, dtype: float64 # result['Zinc, Zn'].order().plot(kind='barh') #AttributeError: 'Series' object has no attribute 'order' result['Zinc, Zn'].sort_values().plot(kind='barh') <matplotlib.axes._subplots.AxesSubplot at 0x2156ce3c748> by_nutrient = ndata.groupby(['nutgroup', 'nutrient'])get_maximum = lambda x: x.xs(x.value.idxmax()) get_minimum = lambda x: x.xs(x.value.idxmin())max_foods = by_nutrient.apply(get_maximum)[['value', 'food']] print(max_foods) value \ nutgroup nutrient Amino Acids Alanine 8.009 Arginine 7.436 Aspartic acid 10.203 Cystine 1.307 Glutamic acid 17.452 Glycine 19.049 Histidine 2.999 Hydroxyproline 0.803 Isoleucine 4.300 Leucine 7.200 Lysine 6.690 Methionine 1.859 Phenylalanine 4.600 Proline 12.295 Serine 4.600 Threonine 3.300 Tryptophan 1.600 Tyrosine 3.300 Valine 4.500 Composition Adjusted Protein 12.900 Carbohydrate, by difference 100.000 Fiber, total dietary 79.000 Protein 88.320 Sugars, total 99.800 Total lipid (fat) 100.000 Water 100.000 Elements Calcium, Ca 7364.000 Copper, Cu 15.050 Fluoride, F 584.000 Iron, Fe 87.470 ... ... Vitamins Cryptoxanthin, beta 6186.000 Dihydrophylloquinone 103.800 Folate, DFE 2630.000 Folate, food 2340.000 Folate, total 2340.000 Folic acid 1538.000 Lutein + zeaxanthin 39550.000 Lycopene 46260.000 Menaquinone-4 33.200 Niacin 97.000 Pantothenic acid 35.000 Retinol 30000.000 Riboflavin 14.300 Thiamin 20.000 Tocopherol, beta 6.490 Tocopherol, delta 30.880 Tocopherol, gamma 100.880 Vitamin A, IU 100000.000 Vitamin A, RAE 30000.000 Vitamin B-12 98.890 Vitamin B-12, added 24.000 Vitamin B-6 12.000 Vitamin C, total ascorbic acid 2400.000 Vitamin D 10000.000 Vitamin D (D2 + D3) 250.000 Vitamin D2 (ergocalciferol) 28.100 Vitamin D3 (cholecalciferol) 27.400 Vitamin E (alpha-tocopherol) 149.400 Vitamin E, added 46.550 Vitamin K (phylloquinone) 1714.500 food nutgroup nutrient Amino Acids Alanine Gelatins, dry powder, unsweetened Arginine Seeds, sesame flour, low-fat Aspartic acid Soy protein isolate Cystine Seeds, cottonseed flour, low fat (glandless) Glutamic acid Soy protein isolate Glycine Gelatins, dry powder, unsweetened Histidine Whale, beluga, meat, dried (Alaska Native) Hydroxyproline KENTUCKY FRIED CHICKEN, Fried Chicken, ORIGINA... Isoleucine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Leucine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Lysine Seal, bearded (Oogruk), meat, dried (Alaska Na... Methionine Fish, cod, Atlantic, dried and salted Phenylalanine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Proline Gelatins, dry powder, unsweetened Serine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Threonine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Tryptophan Sea lion, Steller, meat with fat (Alaska Native) Tyrosine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Valine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Composition Adjusted Protein Baking chocolate, unsweetened, squares Carbohydrate, by difference Sweeteners, tabletop, fructose, dry, powder Fiber, total dietary Corn bran, crude Protein Soy protein isolate, potassium type, crude pro... Sugars, total Sugars, granulated Total lipid (fat) Oil, wheat germ Water Water, bottled, POLAND SPRING Elements Calcium, Ca Leavening agents, baking powder, double-acting... Copper, Cu Veal, variety meats and by-products, liver, co... Fluoride, F Tea, instant, sweetened with sugar, lemon-flav... Iron, Fe Salad dressing, russian dressing, low calorie ... ... Vitamins Cryptoxanthin, beta Spices, paprika Dihydrophylloquinone Margarine, 80% fat, stick, includes regular an... Folate, DFE Cereals ready-to-eat, QUAKER, CAP'N CRUNCH Folate, food Leavening agents, yeast, baker's, active dry Folate, total Leavening agents, yeast, baker's, active dry Folic acid Cereals ready-to-eat, QUAKER, CAP'N CRUNCH Lutein + zeaxanthin Kale, raw Lycopene Tomato powder Menaquinone-4 Chicken, broilers or fryers, drumstick, meat a... Niacin Yeast extract spread Pantothenic acid Cereals ready-to-eat, KELLOGG, KELLOGG'S Compl... Retinol Fish oil, cod liver Riboflavin Yeast extract spread Thiamin MORNINGSTAR FARMS Hot and Spicy Veggie Sausage... Tocopherol, beta Yellow pond lily, Wocas, dried seeds (Pacific ... Tocopherol, delta Oil, cooking and salad, ENOVA, 80% diglycerides Tocopherol, gamma Oil, cooking and salad, ENOVA, 80% diglycerides Vitamin A, IU Fish oil, cod liver Vitamin A, RAE Fish oil, cod liver Vitamin B-12 Mollusks, clam, mixed species, cooked, moist heat Vitamin B-12, added Cereals ready-to-eat, KELLOGG, KELLOGG'S ALL-B... Vitamin B-6 Cereals ready-to-eat, KELLOGG, KELLOGG'S ALL-B... Vitamin C, total ascorbic acid Orange-flavor drink, breakfast type, low calor... Vitamin D Fish oil, cod liver Vitamin D (D2 + D3) Fish oil, cod liver Vitamin D2 (ergocalciferol) Mushrooms, maitake, raw Vitamin D3 (cholecalciferol) Fish, halibut, Greenland, raw Vitamin E (alpha-tocopherol) Oil, wheat germ Vitamin E, added Cereals ready-to-eat, GENERAL MILLS, Multi-Gra... Vitamin K (phylloquinone) Spices, sage, ground [94 rows x 2 columns] max_foods.loc['Amino Acids']['food'] nutrient Alanine Gelatins, dry powder, unsweetened Arginine Seeds, sesame flour, low-fat Aspartic acid Soy protein isolate Cystine Seeds, cottonseed flour, low fat (glandless) Glutamic acid Soy protein isolate Glycine Gelatins, dry powder, unsweetened Histidine Whale, beluga, meat, dried (Alaska Native) Hydroxyproline KENTUCKY FRIED CHICKEN, Fried Chicken, ORIGINA... Isoleucine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Leucine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Lysine Seal, bearded (Oogruk), meat, dried (Alaska Na... Methionine Fish, cod, Atlantic, dried and salted Phenylalanine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Proline Gelatins, dry powder, unsweetened Serine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Threonine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Tryptophan Sea lion, Steller, meat with fat (Alaska Native) Tyrosine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Valine Soy protein isolate, PROTEIN TECHNOLOGIES INTE... Name: food, dtype: object

參考資料:煉數成金Python數據分析課程

總結

以上是生活随笔為你收集整理的Python数据分析_第06课:数据清洗与初步分析_笔记的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。

国产一区福利在线 | www.狠狠色 | 国产精品毛片网 | 中文字幕国语官网在线视频 | 在线观看免费 | 国产精品毛片一区二区在线 | 国产视频资源在线观看 | 国产香蕉久久精品综合网 | 97av精品 | 日韩特级黄色片 | 天天天天干 | av动态图片 | 在线看不卡av| 91亚洲精品在线观看 | 午夜少妇一区二区三区 | 中文电影网| 在线国产专区 | 国产夫妻性生活自拍 | 国产色道| 中国一级片免费看 | 美女黄频视频大全 | 国产亚洲精品无 | 黄色一级性片 | 日韩欧美不卡 | 国产又粗又猛又色又黄视频 | 国产成人精品一区二区在线 | 久久激情五月婷婷 | 九热精品| 91成人网在线观看 | 在线中文字母电影观看 | 美女网站黄在线观看 | 久久国产精品免费观看 | 亚洲一区网 | 伊人六月 | 久久综合婷婷综合 | 丁香色综合 | 亚洲成 人精品 | 国产伦理久久精品久久久久_ | 欧美日韩中文在线视频 | 91麻豆精品国产91久久久使用方法 | 欧美看片 | 欧美日韩免费在线观看视频 | 中文字幕免费在线 | 成人在线观看资源 | 欧美精品视 | 国产不卡在线看 | 伊人天堂久久 | 日本久久精品 | 国产黄在线观看 | 五月婷婷导航 | 2022久久国产露脸精品国产 | www.啪啪.com | 在线观看视频亚洲 | 成人在线视频你懂的 | 亚洲精品福利在线 | av高清一区二区三区 | 精品综合久久 | 亚洲在线免费视频 | 亚洲一二三区精品 | 操高跟美女 | 久久69精品| 五月婷婷狠狠 | 亚洲精品视频在线观看免费视频 | 黄色一级在线免费观看 | 人人干,人人爽 | 国产成人精品午夜在线播放 | 日韩一级精品 | 91黄色影视 | 国内精品久久久久影院一蜜桃 | 中文字幕资源网在线观看 | 在线免费av网站 | 干综合网 | 日韩在线色 | 综合色综合色 | 国内精品小视频 | 一区二区三区精品在线视频 | 国产成人亚洲精品自产在线 | 韩国精品一区二区三区六区色诱 | 激情综合五月天 | 成人av影视观看 | 一区二区视频在线免费观看 | 在线播放 亚洲 | 国产精品理论在线观看 | www.com.黄| 美女一二三区 | 男女视频91 | 国产精品一区二区av影院萌芽 | 高清av在线免费观看 | 久久视频| 国产精品私拍 | www色片 | 日韩黄色av网站 | www免费黄色| 精品免费久久久久 | 久久xx视频 | 免费观看国产精品视频 | 精品视频999| 欧美va天堂va视频va在线 | 天天操天天操天天操 | 色先锋资源网 | 91丨九色丨高潮丰满 | 狠狠操狠狠 | 一区二区伦理电影 | 亚洲精品国产精品国产 | 蜜臀av网站 | 久久蜜臀一区二区三区av | 久保带人 | 91精品国产99久久久久久久 | 欧美日韩视频在线观看一区二区 | 少妇高潮冒白浆 | 亚洲一区二区精品3399 | 国产精品精品国产 | 国产又粗又猛又爽又黄的视频先 | 久久久精品网 | 日韩在线免费观看视频 | 亚洲第一成网站 | 伊人五月天.com | 五月天天天操 | 少妇bbb好爽 | 日韩精品中文字幕久久臀 | 国产精品视频全国免费观看 | 久久亚洲二区 | 亚洲成人家庭影院 | 日本黄色免费播放 | 亚洲高清视频在线播放 | 99 精品 在线 | 免费在线观看中文字幕 | 少妇bbbb | 黄色软件视频网站 | 国产一级一片免费播放放a 一区二区三区国产欧美 | 国产午夜在线观看视频 | 天天插夜夜操 | 精品久久一区二区三区 | 久久国产精品久久久久 | a级国产乱理论片在线观看 伊人宗合网 | 视频一区二区国产 | 男女拍拍免费视频 | 最新日韩视频在线观看 | 久久免费视频在线观看30 | 欧美日韩免费看 | 免费看麻豆| 国产97超碰| 久草线 | 国产啊v在线观看 | 国产一区二区精品久久91 | 久草视频一区 | 999久久a精品合区久久久 | 欧美最猛性xxxxx免费 | 黄色a级片在线观看 | 天天色综合久久 | 五月天网站在线 | 日韩电影一区二区在线观看 | 又黄又爽又无遮挡的视频 | 欧美精品成人在线 | 日韩激情片在线观看 | 在线日韩视频 | 999视频在线播放 | 久久综合久久鬼 | 欧美日韩免费一区二区三区 | 97网| 亚洲一区二区三区四区在线视频 | 国产精品久久久久久久久久直播 | 久久伊人爱 | 久久亚洲二区 | av三区在线 | 不卡的av在线播放 | 欧美日韩视频精品 | 四虎影视国产精品免费久久 | 草久中文字幕 | 欧美精品一区二区蜜臀亚洲 | 在线观看视频一区二区三区 | 蜜臀av性久久久久av蜜臀妖精 | 91av在线视频免费观看 | 久久免费的精品国产v∧ | 美女视频久久黄 | 乱男乱女www7788| 狠色在线 | 亚洲精品在线资源 | 麻豆极品| 免费看片在线观看 | 超碰人人干人人 | 精品99久久久久久 | 久久污视频 | 人人要人人澡人人爽人人dvd | 日本爱爱免费视频 | 欧美精品久久人人躁人人爽 | 国产在线精品国自产拍影院 | 欧美精品一二三 | 久久视频在线视频 | 日韩精品一区二区三区高清免费 | 久草在线资源免费 | 亚洲成成品网站 | 日韩精品久久一区二区 | 天躁狠狠躁 | 日本午夜在线观看 | 最新动作电影 | 久久久久久久久久久久久久免费看 | 午夜资源站 | 国产精品久久久av久久久 | 免费看的黄色网 | 一级片免费观看视频 | 黄色av电影在线观看 | 成人欧美一区二区三区黑人麻豆 | 成人污视频在线观看 | 久久蜜桃av| 丝袜美腿一区 | 18+视频网站链接 | 久久免费视频在线观看 | 在线国产视频 | 成人黄色中文字幕 | 午夜精品久久久久久久99水蜜桃 | 亚洲精品午夜久久久久久久 | 91亚洲精 | 久草成人在线 | 色婷婷成人网 | 亚洲欧美色婷婷 | 在线中文字幕电影 | 不卡的av在线播放 | 国产三级香港三韩国三级 | 丁香婷婷在线观看 | 最新av在线免费观看 | aav在线 | 日韩系列在线观看 | 日本中文字幕网址 | 丁香六月激情 | 婷婷久久国产 | 三上悠亚在线免费 | 久草在线一免费新视频 | 天天爽夜夜爽人人爽一区二区 | 亚洲国产日韩一区 | 午夜影院一区 | 香蕉视频免费在线播放 | 一级黄色大片在线观看 | 欧美日韩视频在线一区 | 天天操偷偷干 | 麻豆 91 在线| 日日干夜夜爱 | 日本久久久亚洲精品 | 国产精品久久久久久久久费观看 | 国内成人综合 | 国产亚洲精品免费 | avwww在线| 国产在线精品一区二区三区 | 中文字幕在线国产精品 | 日操操| 欧美综合在线视频 | 99视频一区 | 日韩久久网站 | 色网站国产精品 | 成人久久精品视频 | 精品国产乱码久久久久久天美 | 午夜国产福利在线 | 黄色一级大片在线免费看产 | 日韩美女黄色片 | 91香蕉国产在线观看软件 | 又污又黄网站 | 在线观看黄色免费视频 | 国产小视频网站 | 日本资源中文字幕在线 | 91一区二区三区在线观看 | 久99热| av观看免费在线 | 婷婷狠狠操 | 最近日本中文字幕a | www.av小说 | 丝袜制服综合网 | 久久免费福利 | 久久久综合色 | 精品黄色在线 | 日日操狠狠干 | 狠狠五月婷婷 | 国产免费av一区二区三区 | 日韩婷婷 | 深爱婷婷 | av一区在线播放 | 91人人网 | 欧美日韩精品影院 | 久久免费视频8 | 一二三区高清 | 香蕉视频免费在线播放 | 色wwww| 在线亚洲高清视频 | 麻豆 free xxxx movies hd | 成人黄色国产 | 十八岁以下禁止观看的1000个网站 | 日韩免费视频播放 | 国产精品美女久久久久久免费 | 亚洲精品大全 | 国产xxxxx在线观看 | 青青草在久久免费久久免费 | 国产精品久久久99 | 国产精品久久久久aaaa | 黄色精品在线看 | 成人在线免费视频观看 | 五月婷在线观看 | 中文字幕二区 | 日韩一级黄色av | 岛国精品一区二区 | 中文字幕永久在线 | 97视频成人| 美女网站免费福利视频 | 在线免费av电影 | 免费看片成年人 | 伊人国产女 | 欧美一级片在线观看视频 | 久草在线电影网 | 亚洲综合小说电影qvod | 日本中文字幕影院 | 欧美 日韩 国产 成人 在线 | 五月天丁香亚洲 | 亚洲欧美日本A∨在线观看 青青河边草观看完整版高清 | 区一区二区三在线观看 | 亚洲精品午夜国产va久久成人 | 欧美精品午夜 | 久久视频二区 | 日韩欧美在线综合网 | 国产区精品在线 | 国产99视频在线观看 | 五月婷婷综合久久 | 日本 在线 视频 中文 有码 | 亚洲小视频在线观看 | 亚洲精品婷婷 | 黄色一级动作片 | 蜜臀精品久久久久久蜜臀 | 国产免费久久久久 | 亚洲理论在线观看电影 | 久久久久久伊人 | 国产精品视频免费在线观看 | 五月婷婷激情五月 | 狠狠色伊人亚洲综合网站野外 | 亚洲精品tv久久久久久久久久 | 亚洲精选在线 | 亚洲国产精品久久久久 | 国产精品久久久久久久久久久久午夜 | 日韩城人在线 | 国产精品高清在线观看 | 婷婷综合av | 国产精品男女视频 | 色综合小说 | 欧美 另类 交 | 波多野结衣小视频 | 九九热视频在线 | 久久免费视频99 | 看黄色.com| 久久艹国产视频 | 五月天久久久 | 西西人体4444www高清视频 | 日本字幕网 | 97超碰在线久草超碰在线观看 | 亚洲极色 | 人人超碰在线 | 亚洲 欧洲av | 91色网址| 天天操天天操 | 国产尤物视频在线 | 久久99久久99精品中文字幕 | 久久国产午夜精品理论片最新版本 | 91av免费在线观看 | 一区二区三区免费看 | 久久久久国产精品厨房 | 美女网站视频色 | 亚洲精品在线播放视频 | 五月激情天 | 天堂入口网站 | www国产亚洲精品久久网站 | 日韩大片免费在线观看 | 精品自拍网 | 91免费的视频在线播放 | 久久有精品 | 国产精品美女www爽爽爽视频 | 中文字幕在线精品 | 97视频人人免费看 | 久久精品视频免费播放 | 中文字幕综合在线 | 久久精品国产成人 | 精品99视频 | 人人爽人人香蕉 | 国产一级特黄毛片在线毛片 | www黄色com | av电影一区二区 | 国产黄色理论片 | 色国产视频 | 91激情视频在线播放 | 欧美日韩精品影院 | 免费a级毛片在线看 | 蜜臀av夜夜澡人人爽人人桃色 | 九九在线播放 | 激情丁香综合 | 国产无遮挡猛进猛出免费软件 | 九色在线视频 | 免费看的黄色录像 | 久久久在线 | 丰满少妇麻豆av | 亚洲精品视频网 | 爱爱av网 | 中文字幕一区二区三区四区视频 | 97视频在线观看成人 | 精品国产乱码久久久久久1区2匹 | 久久综合久久综合久久 | av短片在线观看 | 欧美久久成人 | 国产精品区二区三区日本 | 日日爽夜夜爽 | 懂色av一区二区在线播放 | 日韩免费大片 | 国产一线二线三线性视频 | 成人福利在线观看 | 久久乱码卡一卡2卡三卡四 五月婷婷久 | 成全免费观看视频 | 亚洲色图美腿丝袜 | 亚洲国产精品女人久久久 | 91精品黄色 | 奇米网444| 国产精品亚洲精品 | 精久久久久| 一级性av | 欧美黑吊大战白妞欧美 | 中文字幕日本在线 | 视频在线精品 | 黄色在线网站噜噜噜 | 在线91色 | 91在线免费观看国产 | 日韩免费大片 | 欧美色综合天天久久综合精品 | 日韩国产高清在线 | 免费精品 | 国产又粗又猛又黄又爽视频 | 香蕉视频在线免费 | 日本精品一区二区三区在线播放视频 | 色婷婷国产 | 日韩免费在线观看视频 | 国产黑丝一区二区 | 水蜜桃亚洲一二三四在线 | 国产九九九视频 | 久久这里只有精品23 | 精品自拍av | 91精品少妇偷拍99 | 五月激情电影 | 中文字幕精品久久 | 日本三级久久久 | 日本高清中文字幕有码在线 | 最新中文字幕在线资源 | 91免费日韩| 欧美在线91 | 国产精品免费在线播放 | 成人国产一区二区 | 免费人做人爱www的视 | 69国产盗摄一区二区三区五区 | 丁香色婷 | 成人黄色电影视频 | 日韩在线观看精品 | 亚洲精品乱码久久久久久蜜桃欧美 | 免费高清在线观看电视网站 | 四虎国产精品成人免费4hu | 亚在线播放中文视频 | 久久精品在线免费观看 | 91在线国内视频 | 国产第一页福利影院 | 在线观看日韩免费视频 | 日韩在线高清 | 99色视频在线 | 免费视频一二三 | 99久久久久久 | 国产精品久久久久久模特 | 日韩激情av在线 | 激情文学综合丁香 | 免费人成在线观看 | 亚洲 欧美 精品 | 91尤物国产尤物福利在线播放 | 在线播放日韩av | 91在线视频一区 | 免费观看www7722午夜电影 | 奇米网在线观看 | 亚洲永久精品视频 | 成人a视频 | 8x成人在线 | 亚洲特级片 | 一区国产精品 | 四虎影院在线观看av | 欧洲亚洲激情 | 免费在线观看黄网站 | 亚洲欧洲成人精品av97 | 日本不卡123 | 在线观看日韩精品 | 狠狠色丁婷婷日日 | 在线免费视频一区 | 亚洲免费av电影 | 欧美一区二区三区免费看 | 在线观看日韩精品 | 国产在线资源 | 中文字幕欧美日韩va免费视频 | 免费在线中文字幕 | 九七在线视频 | 黄色精品一区 | 黄色三级免费片 | www国产亚洲精品久久麻豆 | 蜜臀av免费一区二区三区 | 免费看一级特黄a大片 | 97天天综合网 | 九九久久免费 | 日本女人b | 国产成人精品三级 | 久草在线欧美 | 国内外激情视频 | 99色资源 | 日本精品一 | 欧美成年黄网站色视频 | 成人av一二三区 | 久久久久久久久久久久久9999 | 免费看黄在线观看 | 亚洲天堂视频在线 | 97视频人人澡人人爽 | 热久久最新地址 | 最近中文字幕免费观看 | 中文字幕中文字幕在线中文字幕三区 | 日韩成人在线免费观看 | 色综合久久88色综合天天 | 免费试看一区 | 中国一级片在线观看 | 色欧美日韩 | 国产精品v欧美精品v日韩 | 亚洲特级毛片 | 久久精品这里精品 | 91在线www| 国产a视频免费观看 | 久久优 | 久久高清视频免费 | 精品国产乱码一区二 | 在线草| 成人a大片 | 亚洲va男人天堂 | 亚洲清纯国产 | 国产精品美女久久久久久久久 | av中文字幕在线免费观看 | 亚洲国产成人高清精品 | 国产黄色大片 | 国产亚洲精品女人久久久久久 | 中文在线a√在线 | 国产夫妻性生活自拍 | 天堂在线视频中文网 | 欧美激情h | 国产日韩欧美在线观看视频 | 在线影视 一区 二区 三区 | 国产一二区视频 | 成人毛片久久 | 久久这里只有精品首页 | 美女久久久久久久久久 | 久久视频这里只有精品 | 操高跟美女 | 婷婷久久综合网 | 久久久久国产免费免费 | 麻豆精品国产传媒 | 国产精品a久久久久 | 国产精品一区二区久久国产 | 免费av看片 | 麻豆国产精品va在线观看不卡 | 最近中文字幕高清字幕免费mv | 97在线看 | 免费a v观看 | 国产福利在线 | 日韩伦理片hd | 午夜精品一区二区三区免费 | 天天色天天射天天综合网 | 99久久www | 亚洲成人黄色av | 亚洲综合欧美精品电影 | 又粗又长又大又爽又黄少妇毛片 | 黄色小说视频在线 | 久久精品视频免费 | 一级黄色在线视频 | 欧美日韩国产二区 | 九九热在线观看 | 久久这里只有精品视频首页 | 久久人人爽av | 国产福利网站 | 婷婷久久一区二区三区 | 在线免费观看国产黄色 | 国产又粗又猛又黄又爽视频 | 91日韩免费 | 手机av在线网站 | 99久久夜色精品国产亚洲 | 日黄网站 | 在线观看亚洲国产精品 | 免费看污在线观看 | 国产女人18毛片水真多18精品 | 久久尤物电影视频在线观看 | 天天玩天天操天天射 | 又黄又刺激的网站 | 亚洲狠狠婷婷综合久久久 | 在线观看日本高清mv视频 | 免费在线观看亚洲视频 | 日本公妇在线观看 | 日韩久久网站 | 亚洲国产激情 | 一区二区电影在线观看 | 久久成人高清 | 色www精品视频在线观看 | 欧美 日韩 性 | 久久99久久99精品免观看粉嫩 | 西西444www| 午夜精品一区二区三区在线 | 欧美日韩aa| 色视频在线看 | 亚洲久草网 | 成人性生交大片免费观看网站 | 精品在线免费观看 | 日日碰狠狠添天天爽超碰97久久 | 999在线精品 | wwwwww国产 | 国产精品久久久久久久久久免费看 | 久久久精品视频网站 | 麻豆精品在线视频 | 成人亚洲精品久久久久 | 亚洲色图美腿丝袜 | 日韩一区二区三区免费视频 | 亚洲人成网站精品片在线观看 | 四虎天堂| 久久99国产精品久久99 | 亚洲一二区视频 | 久久理论电影 | 缴情综合网五月天 | 成人久久影院 | 日韩三级中文字幕 | 97在线免费观看 | 综合伊人久久 | 精品久久久久久亚洲 | 日本精品久久久一区二区三区 | 97国产超碰 | 国产亚洲免费的视频看 | 91丨九色丨首页 | 999久久久免费精品国产 | 五月婷网站 | 久久久久久久久久久综合 | 在线看中文字幕 | 天天干天天干天天干 | 欧美在线视频日韩 | 国产在线91在线电影 | 国产麻豆果冻传媒在线观看 | 亚州精品天堂中文字幕 | 天天射天天射天天射 | 国产日韩精品一区二区三区在线 | 亚洲五月婷婷 | 日韩午夜电影 | 韩国av一区二区三区 | 中文字幕观看视频 | 69绿帽绿奴3pvideos | 在线观看中文字幕第一页 | 婷婷丁香久久五月婷婷 | 国产九色在线播放九色 | 亚洲精品国产免费 | 精品国产伦一区二区三区免费 | 国产伦精品一区二区三区… | 韩国一区二区三区在线观看 | 一区二区视频在线观看免费 | 亚洲成人午夜在线 | 色播六月天 | 国产精品18久久久久久久 | 97夜夜澡人人双人人人喊 | 久久av不卡 | 欧美一级视频免费看 | 国产小视频你懂的 | 黄色a视频免费 | 久久国产区 | 欧美性大战久久久久 | 国产激情久久久 | 婷婷午夜天 | 欧美激情亚洲综合 | 一区免费视频 | 久久蜜臀av | 国产视频 亚洲精品 | 国产精品99久久久久久小说 | 一区二区三区电影大全 | 国产中文在线观看 | 九九久久免费视频 | 国产你懂的在线 | 国产精品成人aaaaa网站 | 一本一道波多野毛片中文在线 | 欧美色精品天天在线观看视频 | 四虎影视精品成人 | 亚洲精品综合一二三区在线观看 | 草久在线 | 欧美日韩国产综合网 | 人人爽人人爽人人片 | 午夜 在线 | 欧美性生活免费 | 国产精品99在线观看 | 亚洲精品免费视频 | 在线观看亚洲电影 | 在线观看午夜av | 国内精品久久久久 | 丁香婷婷综合色啪 | 成人免费视频免费观看 | 国产精品成人a免费观看 | 九色福利视频 | 中文字幕在线视频网站 | 欧美精品中文在线免费观看 | h网站免费在线观看 | 在线三级av| 欧洲视频一区 | 91网址在线观看 | 国产美腿白丝袜足在线av | 青草视频在线 | 人人干网 | 狠狠操夜夜操 | 久久久久日本精品一区二区三区 | 久草精品国产 | 婷婷深爱网 | 亚洲免费在线观看视频 | 激情一区二区三区欧美 | 日日夜操 | 久久久久综合视频 | 99久久婷婷国产综合精品 | 99久视频 | 国产美女精彩久久 | 日韩理论片中文字幕 | 去看片| 一区二区三区精品久久久 | 永久免费精品视频 | 免费在线观看视频a | 天天操欧美 | 亚洲伊人第一页 | 欧美精品免费一区二区 | 成人一级在线观看 | 国产视频二| 亚洲免费观看在线视频 | 综合中文字幕 | 丁香花中文在线免费观看 | 在线视频免费观看 | 九九久| 狠狠操电影网 | 中文av字幕在线观看 | 亚洲影视九九影院在线观看 | 久久成人麻豆午夜电影 | 麻豆国产电影 | 国产黄色免费看 | 亚州精品天堂中文字幕 | 激情丁香综合五月 | 免费视频一区二区 | 粉嫩av一区二区三区四区 | 色91在线 | 日本中文字幕网站 | a久久久久久 | 麻豆视频国产精品 | 99欧美视频 | 亚洲精品中文在线 | 国产伦理一区 | 欧美一级视频在线观看 | 欧美成人在线免费 | 波多野结衣在线中文字幕 | 毛片网免费 | 五月激情丁香婷婷 | 国产视频一区在线 | 亚洲精品国产拍在线 | 中文字幕第一页在线播放 | 亚洲精品久久久蜜桃直播 | 四虎国产精品免费 | 国产精品久久久亚洲 | 亚洲成人一二三 | 丁香婷婷激情国产高清秒播 | 欧美日韩免费一区二区三区 | 免费一级毛毛片 | 激情视频综合网 | 丁香久久 | 日韩av电影免费观看 | 狠狠干狠狠艹 | 最近更新中文字幕 | 中文字幕在线免费 | 色久网| 人人爽久久久噜噜噜电影 | 日日夜夜亚洲 | 中文字幕一区二区三 | 成人理论电影 | 视频在线观看91 | 在线视频 一区二区 | 免费观看国产成人 | 91精品在线播放 | 久久不卡视频 | 一区二区三区视频网站 | 在线观看色网 | www最近高清中文国语在线观看 | 国产精品白虎 | 久久涩视频 | 精品久久久久一区二区国产 | 性色av免费观看 | 国产小视频在线观看免费 | 91精品成人久久 | 一区二区男女 | 国产精品无 | 免费观看久久 | 一级免费观看 | 久久99精品久久久久久清纯直播 | 国产精品日韩在线 | 精品久久免费看 | 亚洲一区二区精品视频 | 成年人免费看的视频 | 西西4444www大胆无视频 | 亚洲激情 | 在线韩国电影免费观影完整版 | 97精品欧美91久久久久久 | 欧美一二三四在线 | av电影在线观看 | 高清日韩一区二区 | 91片黄在线观看 | 日韩成人免费观看 | 成人a v视频 | se视频网址 | 丁香婷婷自拍 | 久久最新| 久久99亚洲网美利坚合众国 | 在线观看免费91 | www在线免费观看 | 永久免费视频国产 | 久久久高清一区二区三区 | 天天色天| 中文字幕视频三区 | 玖玖综合网 | 日日夜夜精品免费 | 91污污 | 国产一级片播放 | 超级碰碰碰免费视频 | 狠狠色香婷婷久久亚洲精品 | 婷婷国产在线观看 | 久久久久女人精品毛片九一 | 欧洲亚洲国产视频 | 黄a在线看| 久久视频国产精品免费视频在线 | 免费观看v片在线观看 | 丁香激情综合国产 | 国产午夜三级一区二区三桃花影视 | 国产精品精品视频 | 亚洲精品视频免费看 | 国产成人免费观看久久久 | 在线观看免费日韩 | 久久久久久久久毛片精品 | 久久人人97超碰国产公开结果 | 国产在线精品福利 | 成人在线免费小视频 | 精品一二三四在线 | 亚洲国产99 | 国产精品第十页 | 久久久国产一区二区三区 | 91精品小视频 | av福利电影 | 欧美精品在线观看免费 | 国产精品久久久久一区二区国产 | 亚洲a资源| 色噜噜在线观看 | 日韩精品一区二区免费视频 | 蜜桃传媒一区二区 | 91中文字幕在线观看 | 日韩电影在线观看一区二区三区 | 96亚洲精品久久久蜜桃 | av官网| 99国产视频在线 | 综合在线观看 | 日韩av二区 | 日韩91av| 亚洲精品白浆高清久久久久久 | 成人a免费视频 | 欧美国产精品一区二区 | 丁香婷婷久久久综合精品国产 | 久久综合中文字幕 | 天天色天天干天天色 | 狠狠色丁香久久综合网 | 国产成人福利在线观看 | 日韩视频在线观看视频 | 日韩欧美电影在线 | 免费在线电影网址大全 | 最近中文字幕免费大全 | 欧美性生活免费看 | 国产 av 日韩 | 成人午夜网| 国产99久久久精品 | 欧美日韩中文在线观看 | 特级aaa毛片 | 中文字幕 91 | 5月丁香婷婷综合 | 在线v片免费观看视频 | 手机色在线| 丁香九月婷婷综合 | 天天综合色天天综合 | 伊人天堂久久 | 亚洲美女视频在线观看 | 亚洲激情婷婷 | 日批在线看 | 久草在线观看视频免费 | 久久综合久久88 | 四虎在线免费 | 国产一级二级在线播放 | 制服丝袜天堂 | 91少妇精拍在线播放 | 99久久www免费 | 色a在线观看 | 四虎免费在线观看 | 亚洲精品永久免费视频 | 国产免费视频一区二区裸体 | 欧美日韩中文字幕综合视频 | 国产一区二区精品在线 | 在线免费色 | av在线播放快速免费阴 | 色婷婷综合在线 | 精品久久影院 | 国产亚洲午夜高清国产拍精品 | 狠狠操综合网 | 缴情综合网五月天 | 91成品人影院 | 91av免费在线观看 | 蜜桃av观看| 亚洲精品videossex少妇 | 9797在线看片亚洲精品 | 蜜臀久久99精品久久久无需会员 | 亚洲精品黄色在线观看 | www.日日日.com| 国产婷婷视频在线 | 97超碰资源 | 日韩国产欧美视频 | 992tv在线| 日韩在线视频网址 | 91av播放| 中文字幕在线精品 | 日日天天av | 国产又粗又长的视频 | 中文字幕在线有码 | 日韩欧美在线综合网 | 久久影视中文字幕 | www五月天com | 2023年中文无字幕文字 | 国产精品一区二区三区免费看 | 天天亚洲综合 | 一级黄色大片在线观看 | www.狠狠操 | 免费黄在线观看 | 青青河边草观看完整版高清 | 免费a v视频| 日韩精品免费一线在线观看 | 天天综合网久久综合网 | 久久一区二区三区日韩 | 五月天婷婷视频 | 在线成人免费电影 | 2023av| 深爱激情av | 一区二区三区日韩在线观看 | 91在线一区二区 | 成人影音在线 | 亚洲美女免费精品视频在线观看 | 毛片美女网站 | www.大网伊人 | 亚洲色图美腿丝袜 | 国产精品久久久久久一区二区 | 99精品在线免费观看 | 玖玖精品在线 | 亚洲精品乱码久久久久v最新版 | 一区二区 不卡 | 国产福利91精品张津瑜 | 精品免费| 亚洲精品字幕在线 | 九九热久久久 | 五月婷婷网站 | 91麻豆文化传媒在线观看 | 国产中的精品av小宝探花 | 91精品视频免费在线观看 | 免费亚洲视频在线观看 | 久久免费中文视频 | 日韩精品免费专区 | 亚洲精品福利视频 | 国产色女人 | 久久99精品国产一区二区三区 | 91九色网站 | 国产福利不卡视频 | 成人免费视频在线观看 | 精品一区欧美 | 国产黑丝袜在线 | 国产偷在线 | 99在线精品观看 | 麻豆久久一区二区 | 国产精品一区二区62 | 在线观看视频在线 | 玖玖玖精品 | 日韩欧美网址 | 99热99| 国产九九九视频 | 日日操操操 | а中文在线天堂 | 久久视频网 | 欧美一级片免费在线观看 | 亚洲人成影院在线 | 波多野结衣一区二区三区中文字幕 | 福利视频精品 | 欧美精品你懂的 | 97成人在线| 成人免费观看网址 | 日本高清中文字幕有码在线 | 在线看欧美 | 高清一区二区 | 日韩在线观看视频中文字幕 | 91免费高清 | 天天爱天天操天天射 | 国产午夜精品久久 |